SaaSHub helps you find the best software and product alternatives Learn more →
Top 10 C++ llama Projects
-
LocalAI
:robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
cortex
Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM). Powers 👋 Jan (by janhq)
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
-
llama-server-chat-terminal-client
Lightweight chat terminal-interface for llama.cpp server compilable for windows and linux.
The first example will build an Embeddings database backed by llama.cpp vectorization.
Project mention: LocalAI: Self-hosted OpenAI alternative reaches 2.14.0 | news.ycombinator.com | 2024-05-03
Jan incorporates a lightweight, built-in inference server called Nitro. Nitro supports both llama.cpp and NVIDIA's TensorRT-LLM engines. This means many open LLMs in the GGUF format are supported. Jan's Model Hub is designed for easy installation of pre-configured models but it also allows you to install virtually any model from Hugging Face or even your own.
Python sits on the C-glue segment of programming languages (where Perl, PHP, Ruby and Node are also notable members). Being a glue language means having APIs to a lot of external toolchains written in not only C/C++ but many other compiled languages, APIs and system resources. Conda, virtualenv, etc. are godsend modules for making it all work, or even better, to freeze things once they all work, without resourcing to Docker, VMs or shell scripts. It's meant for application and DevOps people who need to slap together, ie, ML, Numpy, Elasticsearch, AWS APIs and REST endpoints and Get $hit Done.
It's annoying to see them "glueys" compared to the binary compiled segment where the heavy lifting is done. Python and others exist to latch on and assimilate. Resistance is futile:
https://pypi.org/project/pyllamacpp/
https://www.npmjs.com/package/llama-node
https://packagist.org/packages/kambo/llama-cpp-php
https://github.com/yoshoku/llama_cpp.rb
Project mention: Show HN: Collider – the platform for local LLM debug and inference at warp speed | news.ycombinator.com | 2023-11-30
C++ llama related posts
-
RAG with llama.cpp and external API services
-
Deploying llama.cpp on AWS (with Troubleshooting)
-
Mistral Fine-Tune
-
Distributed LLM Inference with Llama.cpp
-
Ask HN: Which LLMs can run locally on most consumer computers
-
The prospects for 128 bit processors (John R. Mashey, 1995)
-
New exponent functions that make SiLU and SoftMax 2x faster, at full acc
-
A note from our sponsor - SaaSHub
www.saashub.com | 1 Jun 2024
Index
What are some of the best open-source llama projects in C++? This list will help you:
Project | Stars | |
---|---|---|
1 | llama.cpp | 59,389 |
2 | LocalAI | 20,764 |
3 | PowerInfer | 7,024 |
4 | cortex | 1,661 |
5 | LlamaGPTJ-chat | 211 |
6 | llama_cpp.rb | 149 |
7 | booster | 126 |
8 | pyllamacpp | 59 |
9 | llama-server-chat-terminal-client | 10 |
10 | llama-chat | 7 |
Sponsored