Transformer

Open-source projects categorized as Transformer

Top 23 Transformer Open-Source Projects

  • transformers

    🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

  • Project mention: How to count tokens in frontend for Popular LLM Models: GPT, Claude, and Llama | dev.to | 2024-05-21

    Thanks to transformers.js, we can run the tokenizer and model locally in the browser. Transformers.js is designed to be functionally equivalent to Hugging Face's transformers python library, meaning you can run the same pretrained models using a very similar API.

  • nn

    🧑‍🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • whisper.cpp

    Port of OpenAI's Whisper model in C/C++

  • Project mention: Show HN: I created automatic subtitling app to boost short videos | news.ycombinator.com | 2024-04-09

    whisper.cpp [1] has a karaoke example that uses ffmpeg's drawtext filter to display rudimentary karaoke-like captions. It also supports diarisation. Perhaps it could be a starting point to create a better script that does what you need.

    --

    1: https://github.com/ggerganov/whisper.cpp/blob/master/README....

  • mmdetection

    OpenMMLab Detection Toolbox and Benchmark

  • vllm

    A high-throughput and memory-efficient inference and serving engine for LLMs

  • Project mention: AI leaderboards are no longer useful. It's time to switch to Pareto curves | news.ycombinator.com | 2024-04-30

    I guess the root cause of my claim is that OpenAI won't tell us whether or not GPT-3.5 is an MoE model, and I assumed it wasn't. Since GPT-3.5 is clearly nondeterministic at temp=0, I believed the nondeterminism was due to FPU stuff, and this effect was amplified with GPT-4's MoE. But if GPT-3.5 is also MoE then that's just wrong.

    What makes this especially tricky is that small models are truly 100% deterministic at temp=0 because the relative likelihoods are too coarse for FPU issues to be a factor. I had thought 3.5 was big enough that some of its token probabilities were too fine-grained for the FPU. But that's probably wrong.

    On the other hand, it's not just GPT, there are currently floating-point difficulties in vllm which significantly affect the determinism of any model run on it: https://github.com/vllm-project/vllm/issues/966 Note that a suggested fix is upcasting to float32. So it's possible that GPT-3.5 is using an especially low-precision float and introducing nondeterminism by saving money on compute costs.

    Sadly I do not have the money[1] to actually run a test to falsify any of this. It seems like this would be a good little research project.

    [1] Or the time, or the motivation :) But this stuff is expensive.

  • CVPR2024-Papers-with-Code

    CVPR 2024 论文和开源项目合集

  • best-of-ml-python

    🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
  • nlp-tutorial

    Natural Language Processing Tutorial for Deep Learning Researchers

  • dio

    A powerful HTTP client for Dart and Flutter, which supports global settings, Interceptors, FormData, aborting and canceling a request, files uploading and downloading, requests timeout, custom adapters, etc.

  • RWKV-LM

    RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.

  • Project mention: Do LLMs need a context window? | news.ycombinator.com | 2023-12-25

    https://github.com/BlinkDL/RWKV-LM#rwkv-discord-httpsdiscord... lists a number of implementations of various versions of RWKV.

    https://github.com/BlinkDL/RWKV-LM#rwkv-parallelizable-rnn-w... :

    > RWKV: Parallelizable RNN with Transformer-level LLM Performance (pronounced as "RwaKuv", from 4 major params: R W K V)

    > RWKV is an RNN with Transformer-level LLM performance, which can also be directly trained like a GPT transformer (parallelizable). And it's 100% attention-free. You only need the hidden state at position t to compute the state at position t+1. You can use the "GPT" mode to quickly compute the hidden state for the "RNN" mode.

    > So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding (using the final hidden state).

    > "Our latest version is RWKV-6,*

  • LaTeX-OCR

    pix2tex: Using a ViT to convert images of equations into LaTeX code.

  • Project mention: Detexify LaTeX Handwriting Symbol Recognition | news.ycombinator.com | 2023-11-14
  • PaddleSpeech

    Easy-to-use Speech Toolkit including Self-Supervised Learning model, SOTA/Streaming ASR with punctuation, Streaming TTS with text frontend, Speaker Verification System, End-to-End Speech Translation and Keyword Spotting. Won NAACL2022 Best Demo Award.

  • Project mention: Open Source Libraries | /r/AudioAI | 2023-10-02

    PaddlePaddle/PaddleSpeech

  • faster-whisper

    Faster Whisper transcription with CTranslate2

  • Project mention: Creando Subtítulos Automáticos para Vídeos con Python, Faster-Whisper, FFmpeg, Streamlit, Pillow | dev.to | 2024-04-29

    Faster-whisper (https://github.com/SYSTRAN/faster-whisper)

  • petals

    🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading

  • Project mention: Chameleon: Meta's New Multi-Modal LLM | news.ycombinator.com | 2024-05-21

    Things like [petals](https://github.com/bigscience-workshop/petals) exist, distributed computing over willing participants. Right now corporate cash is being rammed into the space so why not snap it up while you can, but the moment it dries up projects like petals will see more of the love they deserve.

    I envision a future where crypto-style booms happen over tokens useful for purchasing priority computational time, which is earned by providing said computational time. This way researchers can daisy-chain their independent smaller rigs together into something with gargantuan capabilities.

  • PaddleSeg

    Easy-to-use image segmentation library with awesome pre-trained model zoo, supporting wide-range of practical tasks in Semantic Segmentation, Interactive Segmentation, Panoptic Segmentation, Image Matting, 3D Segmentation, etc.

  • LMFlow

    An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

  • Project mention: Your weekly machine learning digest | /r/learnmachinelearning | 2023-07-03
  • trax

    Trax — Deep Learning with Clear Code and Speed

  • Project mention: Maxtext: A simple, performant and scalable Jax LLM | news.ycombinator.com | 2024-04-23

    Is t5x an encoder/decoder architecture?

    Some more general options.

    The Flax ecosystem

    https://github.com/google/flax?tab=readme-ov-file

    or dm-haiku

    https://github.com/google-deepmind/dm-haiku

    were some of the best developed communities in the Jax AI field

    Perhaps the “trax” repo? https://github.com/google/trax

    Some HF examples https://github.com/huggingface/transformers/tree/main/exampl...

    Sadly it seems much of the work is proprietary these days, but one example could be Grok-1, if you customize the details. https://github.com/xai-org/grok-1/blob/main/run.py

  • text-generation-inference

    Large Language Model Text Generation Inference

  • Project mention: FLaNK AI-April 22, 2024 | dev.to | 2024-04-22
  • jukebox

    Code for the paper "Jukebox: A Generative Model for Music"

  • Project mention: Open Source Libraries | /r/AudioAI | 2023-10-02

    openai/jukebox: Music Generation

  • mmsegmentation

    OpenMMLab Semantic Segmentation Toolbox and Benchmark.

  • GPT2-Chinese

    Chinese version of GPT2 training code, using BERT tokenizer.

  • bertviz

    BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)

  • Project mention: StreamingLLM: tiny tweak to KV LRU improves long conversations | news.ycombinator.com | 2024-02-13

    This seems only to work cause large GPTs have redundant, undercomplex attentions. See this issue in BertViz about attention in Llama: https://github.com/jessevig/bertviz/issues/128

  • BERT-pytorch

    Google AI 2018 BERT pytorch implementation

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The open source projects on this list are ordered by number of github stars. The number of mentions indicates repo mentiontions in the last 12 Months or since we started tracking (Dec 2020).

Transformer related posts

  • Fish Speech TTS: clone OpenAI TTS in 30 minutes

    1 project | news.ycombinator.com | 22 May 2024
  • Maxtext: A simple, performant and scalable Jax LLM

    10 projects | news.ycombinator.com | 23 Apr 2024
  • Mistral AI Launches New 8x22B Moe Model

    4 projects | news.ycombinator.com | 9 Apr 2024
  • Show HN: I created automatic subtitling app to boost short videos

    1 project | news.ycombinator.com | 9 Apr 2024
  • LLMs on your local Computer (Part 1)

    7 projects | dev.to | 11 Mar 2024
  • Voxos.ai – An Open-Source Desktop Voice Assistant

    7 projects | news.ycombinator.com | 19 Jan 2024
  • RAG Using Structured Data: Overview and Important Questions

    5 projects | news.ycombinator.com | 10 Jan 2024
  • A note from our sponsor - InfluxDB
    www.influxdata.com | 24 May 2024
    Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →

Index

What are some of the best open-source Transformer projects? This list will help you:

Project Stars
1 transformers 126,516
2 nn 49,142
3 whisper.cpp 31,817
4 mmdetection 28,036
5 vllm 19,672
6 CVPR2024-Papers-with-Code 16,450
7 best-of-ml-python 15,672
8 nlp-tutorial 13,774
9 dio 12,250
10 RWKV-LM 11,773
11 LaTeX-OCR 11,015
12 PaddleSpeech 10,271
13 faster-whisper 9,278
14 petals 8,744
15 PaddleSeg 8,313
16 LMFlow 8,058
17 trax 7,964
18 text-generation-inference 8,053
19 jukebox 7,612
20 mmsegmentation 7,492
21 GPT2-Chinese 7,376
22 bertviz 6,449
23 BERT-pytorch 6,032

Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com