markasoftware / llama-cpuLinks
Fork of Facebooks LLaMa model to run on CPU
☆772Updated 2 years ago
Alternatives and similar repositories for llama-cpu
Users that are interested in llama-cpu are comparing it to the libraries listed below
Sorting:
- Quantized inference code for LLaMA models☆1,049Updated 2 years ago
- C++ implementation for BLOOM☆810Updated 2 years ago
- Simple UI for LLM Model Finetuning☆2,061Updated last year
- Instruct-tune LLaMA on consumer hardware☆362Updated 2 years ago
- Llama 2 Everywhere (L2E)☆1,519Updated 6 months ago
- MiniLLM is a minimal system for running modern LLMs on consumer-grade GPUs☆918Updated 2 years ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆568Updated last year
- Finetune llama2-70b and codellama on MacBook Air without quantization☆448Updated last year
- Chat with Meta's LLaMA models at home made easy☆837Updated 2 years ago
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆323Updated 2 years ago
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆49Updated 2 years ago
- Inference code for LLaMA models☆189Updated 2 years ago
- A school for camelids☆1,208Updated 2 years ago
- ☆1,481Updated 2 years ago
- Structured and typehinted GPT responses in Python☆742Updated last year
- ☆405Updated 2 years ago
- Inference code and configs for the ReplitLM model family☆985Updated last year
- Locally run an Assistant-Tuned Chat-Style LLM☆499Updated 2 years ago
- Run LLaMA (and Stanford-Alpaca) inference on Apple Silicon GPUs.☆589Updated 2 years ago
- Basaran is an open-source alternative to the OpenAI text completion API. It provides a compatible streaming API for your Hugging Face Tra…☆1,298Updated last year
- Explore large language models in 512MB of RAM☆1,197Updated last week
- ☆534Updated last year
- fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backe…☆410Updated 2 years ago
- ☆1,028Updated last year
- INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model☆1,537Updated 4 months ago
- SoTA Transformers with C-backend for fast inference on your CPU.☆309Updated last year
- AI-controlled Linux Containers☆668Updated 2 years ago
- JS tokenizer for LLaMA 1 and 2☆355Updated last year
- Lightweight inference library for ONNX files, written in C++. It can run Stable Diffusion XL 1.0 on a RPI Zero 2 (or in 298MB of RAM) but…☆1,972Updated last week
- Official supported Python bindings for llama.cpp + gpt4all☆1,018Updated 2 years ago