mistral-src VS chameleon-llm

Compare mistral-src vs chameleon-llm and see what are their differences.

mistral-src

Reference implementation of Mistral AI 7B v0.1 model. (by mistralai)

chameleon-llm

Codes for "Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models". (by lupantech)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
mistral-src chameleon-llm
9 3
8,732 1,020
4.1% -
7.3 6.2
about 2 months ago 5 months ago
Jupyter Notebook Jupyter Notebook
Apache License 2.0 Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

mistral-src

Posts with mentions or reviews of mistral-src. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-01-01.
  • Mistral 7B vs. Mixtral 8x7B
    1 project | dev.to | 26 Mar 2024
    A French startup, Mistral AI has released two impressive large language models (LLMs) - Mistral 7B and Mixtral 8x7B. These models push the boundaries of performance and introduce a better architectural innovation aimed at optimizing inference speed and computational efficiency.
  • How to have your own ChatGPT on your machine (and make him discussed with himself)
    1 project | dev.to | 24 Jan 2024
    However, some models are publicly available. It’s the case for Mistral, a fast, and efficient French model which seems to outperform GPT4 on some tasks. And it is under Apache 2.0 license 😊.
  • How to Serve LLM Completions in Production
    1 project | dev.to | 18 Jan 2024
    I recommend starting either with llama2 or Mistral. You need to download the pretrained weights and convert them into GGUF format before they can be used with llama.cpp.
  • Stuff we figured out about AI in 2023
    5 projects | news.ycombinator.com | 1 Jan 2024
    > Instead, it turns out a few hundred lines of Python is genuinely enough to train a basic version!

    actually its not just a basic version. Llama 1/2's model.py is 500 lines: https://github.com/facebookresearch/llama/blob/main/llama/mo...

    Mistral (is rumored to have) forked llama and is 369 lines: https://github.com/mistralai/mistral-src/blob/main/mistral/m...

    and both of these are SOTA open source models.

  • How Open is Generative AI? Part 2
    8 projects | dev.to | 19 Dec 2023
    MistralAI, a French startup, developed a 7.3 billion parameter LLM named Mistral for various applications. Committed to open-sourcing its technology under Apache 2.0, the training dataset details for Mistral remain undisclosed. The Mistral Instruct model was fine-tuned using publicly available instruction datasets from the Hugging Face repository, though specifics about the licenses and potential constraints are not detailed. Recently, MistralAI released Mixtral 8x7B, a model based on the sparse mixture of experts (SMoE) architecture, consisting of several specialized models (likely eight, as suggested by its name) activated as needed.
  • Mistral website was just updated
    3 projects | /r/LocalLLaMA | 11 Dec 2023
  • Mistral AI – open-source models
    1 project | news.ycombinator.com | 8 Dec 2023
  • Mistral 8x7B 32k model [magnet]
    6 projects | news.ycombinator.com | 8 Dec 2023
  • Ask HN: Why the LLaMA code base is so short
    2 projects | news.ycombinator.com | 22 Nov 2023
    I was getting into LLM and I pick up some projects. I tried to dive into the code to see what is secret sauce.

    But the code is so short to the point there is nothing to really read.

    https://github.com/facebookresearch/llama

    I then proceed to check https://github.com/mistralai/mistral-src and suprsingly it's same.

    What is exactly those codebases? It feels like just download the models.

chameleon-llm

Posts with mentions or reviews of chameleon-llm. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-05-08.
  • Giving GPT “Infinite” Knowledge
    3 projects | news.ycombinator.com | 8 May 2023
    > Do you know any active research in this area? I briefly considered playing with this, but my back-of-the-envelope semi-educated feeling for now is that it won't scale.

    I am aware of a couple of potentially promising research directions. One formally academic called Chameleon [0], and one that's more like a grassroots organic effort that aims to build an actually functional Auto-GPT-like, called Agent-LLM [1]. I have read the Chameleon paper, and I must say I'm quite impressed with their architecture. It added a few bits and pieces that most of the early GPT-based agents didn't have, and I have a strong intuition that these will contribute to these things actually working.

    Auto-GPT is another, relatively famous piece of work in this area. However, at least as of v0.2.2, I found it relatively underwhelming. For any online knowledge retrieval+synthesis and retrieval+usage tasks, it seemed to get stuck, but it did sort-of-kind-of OK on plain online knowledge retrieval. After having a look at the Auto-GPT source code, my intuition (yes, I know - "fuzzy feelings without a solid basis" - but I believe that this is simply due to not having an AI background to explain this with crystal-clear wording) is that the poor performance of the current version of Auto-GPT is insufficient skill in prompt-chain architecture and the surprisingly low quality and at times buggy code.

    I think Auto-GPT has some potential. I think the implementation lets down the concept, but that's just a question of refactoring the prompts and the overall code - which it seems like the upstream Github repo has been quite busy with, so I might give it another go in a couple of weeks to see how far it's moved forward.

    > Specifically, as task complexity grows, the amount of results to combine will quickly exceed the context window size of the "combiner" GPT-4. Sure, you can stuff another layer on top, turning it into a tree/DAG, but eventually, I think the partial result itself will be larger than 8k, or even 32k tokens - and I feel this "eventually" will be hit rather quickly. But maybe my feelings are wrong and there is some mileage in this approach.

    Auto-GPT uses an approach based on summarisation and something I'd term 'micro-agents'. For example, when Auto-GPT is searching for an answer to a particular question online, for each search result it finds, it spins up a sub-chain that gets asked a question 'What does this page say about X?' or 'Based on the contents of this page, how can you do Y?'. Ultimately, intelligence is about lossy compression, and this is a starkly exposed when it comes to LLMs because you have no choice but to lose some information.

    > I think the partial result itself will be larger than 8k, or even 32k tokens - and I feel this "eventually" will be hit rather quickly. But maybe my feelings are wrong and there is some mileage in this approach.

    The solution to that would be to synthesize output section by section, or even as an "output stream" that can be captured and/or edited outside the LLM in whole or in chunks. IMO, I do think there's some mileage to be exploited in a recursive "store, summarise, synthesise" approach, but the problem will be that of signal loss. Every time you pass a subtask to a sub-agent, or summarise the outcome of that sub-agent into your current knowledge base, some noise is introduced. It might be that the signal to noise ratio will dissipate as higher and higher order LLM chains are used - analogously to how terrible it was to use electricity or radio waves before any amplification technology became available.

    One possible avenue to explore to crack down on decreasing SNR (based on my own original research, but I can also see some people disclosing online that they are exploring the same path), is to have a second LLM in the loop, double-checking the result of the first one. This has some limitations, but I have successfully used this approach to verify that, for example, the LLM does not outright refuse to carry out a task. This is currently cost-prohibitive to do in a way that would make me personally satisfied and confident enough in the output to make it run full-auto, but I expect that increasing ability to run AI locally will make people more willing to experiment with massive layering of cooperating LLM chains that check each others' work, cooperate, and/or even repeat work using different prompts to pick the best output a la redundant avionics computers.

    [0]: https://github.com/lupantech/chameleon-llm

  • chameleon-llm: Codes for "Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models".
    1 project | /r/aipromptprogramming | 25 Apr 2023
    1 project | /r/coolgithubprojects | 24 Apr 2023

What are some alternatives?

When comparing mistral-src and chameleon-llm you can also consider the following projects:

ReAct - [ICLR 2023] ReAct: Synergizing Reasoning and Acting in Language Models

ragas - Evaluation framework for your Retrieval Augmented Generation (RAG) pipelines

lida - Automatic Generation of Visualizations and Infographics using Large Language Models

Azure-Cognitive-Search-Azure-OpenAI-Accelerator - Virtual Assistant - GPT Smart Search Engine - Bot Framework + Azure OpenAI + Azure AI Search + Azure SQL + Bing API + Azure Document Intelligence + LangChain + CosmosDB

vault-ai - OP Vault ChatGPT: Give ChatGPT long-term memory using the OP Stack (OpenAI + Pinecone Vector Database). Upload your own custom knowledge base files (PDF, txt, epub, etc) using a simple React frontend.

vllm - A high-throughput and memory-efficient inference and serving engine for LLMs

AgentBench - A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)

llama - Inference code for Llama models

AGiXT - AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.

text-generation-webui-colab - A colab gradio web UI for running Large Language Models

LLM-Training-Puzzles - What would you do with 1000 H100s...