randaller / llama-cpuLinks
Inference on CPU code for LLaMA models
☆137Updated 2 years ago
Alternatives and similar repositories for llama-cpu
Users that are interested in llama-cpu are comparing it to the libraries listed below
Sorting:
- Inference code for facebook LLaMA models with Wrapyfi support☆129Updated 2 years ago
- LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.☆128Updated 2 years ago
- Automated prompting and scoring framework to evaluate LLMs using updated human knowledge prompts☆110Updated 2 years ago
- Python bindings for llama.cpp☆198Updated 2 years ago
- Merge Transformers language models by use of gradient parameters.☆206Updated last year
- Falcon LLM ggml framework with CPU and GPU support☆246Updated last year
- 4 bits quantization of LLaMa using GPTQ☆130Updated 2 years ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆146Updated last year
- Inference code for LLaMA models☆42Updated 2 years ago
- Embeddings focused small version of Llama NLP model☆103Updated 2 years ago
- Chat with Meta's LLaMA models at home made easy☆837Updated 2 years ago
- 💬 Chatbot web app + HTTP and Websocket endpoints for LLM inference with the Petals client☆314Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆162Updated last year
- Train llama with lora on one 4090 and merge weight of lora to work as stanford alpaca.☆51Updated 2 years ago
- C++ implementation for 💫StarCoder☆456Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated last year
- Extend the original llama.cpp repo to support redpajama model.☆118Updated 11 months ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated 2 years ago
- Host the GPTQ model using AutoGPTQ as an API that is compatible with text generation UI API.☆91Updated 2 years ago
- An unsupervised model merging algorithm for Transformers-based language models.☆106Updated last year
- Instruct-tuning LLaMA on consumer hardware☆66Updated 2 years ago
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆87Updated last week
- 4 bits quantization of SantaCoder using GPTQ☆51Updated 2 years ago
- llama.cpp to PyTorch Converter☆34Updated last year
- 1.58-bit LLaMa model☆81Updated last year
- ☆458Updated last year
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆126Updated last year
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆37Updated 2 years ago
- ☆534Updated last year
- Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.☆71Updated 2 years ago