Python mixture-of-experts

Open-source Python projects categorized as mixture-of-experts

Top 9 Python mixture-of-expert Projects

  • DeepSpeed

    DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

  • Project mention: Can we discuss MLOps, Deployment, Optimizations, and Speed? | /r/LocalLLaMA | 2023-12-06

    DeepSpeed can handle parallelism concerns, and even offload data/model to RAM, or even NVMe (!?) . I'm surprised I don't see this project used more.

  • LLaMA-Factory

    Unify Efficient Fine-Tuning of 100+ LLMs

  • Project mention: FLaNK-AIM Weekly 06 May 2024 | dev.to | 2024-05-06
  • Scout Monitoring

    Free Django app performance insights with Scout Monitoring. Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.

    Scout Monitoring logo
  • mixtral-offloading

    Run Mixtral-8x7B models in Colab or consumer desktops

  • Project mention: DBRX: A New Open LLM | news.ycombinator.com | 2024-03-27

    Waiting for Mixed Quantization with MQQ and MoE Offloading [1]. With that I was able to run Mistral 8x7B on my 10 GB VRAM rtx3080... This should work for DBRX and should shave off a ton of VRAM requirement.

    1. https://github.com/dvmazur/mixtral-offloading?tab=readme-ov-...

  • hivemind

    Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.

  • Project mention: You can now train a 70B language model at home | news.ycombinator.com | 2024-03-07

    https://github.com/learning-at-home/hivemind is also relevant

  • mixture-of-experts

    PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538

  • Project mention: [Rumor] Potential GPT-4 architecture description | /r/LocalLLaMA | 2023-06-20
  • tutel

    Tutel MoE: An Optimized Mixture-of-Experts Implementation

  • mixture-of-experts

    A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models (by lucidrains)

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • mergoo

    A library for easily merging multiple LLM experts, and efficiently train the merged LLM.

  • Project mention: A Library to build MoE from HF models | news.ycombinator.com | 2024-04-08

    https://github.com/Leeroo-AI/mergoo

  • st-moe-pytorch

    Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch

  • Project mention: will the point meet in 2024? | /r/LocalLLaMA | 2023-12-05
NOTE: The open source projects on this list are ordered by number of github stars. The number of mentions indicates repo mentiontions in the last 12 Months or since we started tracking (Dec 2020).

Python mixture-of-experts related posts

  • Would anyone be interested in contributing to some group projects?

    4 projects | /r/learnmachinelearning | 24 Aug 2023
  • [Rumor] Potential GPT-4 architecture description

    2 projects | /r/LocalLLaMA | 20 Jun 2023
  • Hive mind:Train deep learning models on thousands of volunteers across the world

    1 project | news.ycombinator.com | 20 Jun 2023
  • Could a model not be trained by a decentralized network? Like Seti @ home or kinda-sorta like bitcoin. Petals accomplishes this somewhat, but if raw computer power is the only barrier to open-source I'd be happy to try organizing decentalized computing efforts

    2 projects | /r/LocalLLaMA | 17 Jun 2023
  • Orca (built on llama13b) looks like the new sheriff in town

    2 projects | /r/LocalLLaMA | 6 Jun 2023
  • Do you think that AI research will slow down to a halt because of regulation?

    1 project | /r/singularity | 21 May 2023
  • [D] Google "We Have No Moat, And Neither Does OpenAI": Leaked Internal Google Document Claims Open Source AI Will Outcompete Google and OpenAI

    1 project | /r/MachineLearning | 4 May 2023
  • A note from our sponsor - InfluxDB
    www.influxdata.com | 1 Jun 2024
    Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →

Index

What are some of the best open-source mixture-of-expert projects in Python? This list will help you:

Project Stars
1 DeepSpeed 33,122
2 LLaMA-Factory 22,989
3 mixtral-offloading 2,261
4 hivemind 1,847
5 mixture-of-experts 869
6 tutel 672
7 mixture-of-experts 555
8 mergoo 319
9 st-moe-pytorch 237

Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com