leloykun / llama2.cppLinks
Inference Llama 2 in one file of pure C++
☆87Updated 2 years ago
Alternatives and similar repositories for llama2.cpp
Users that are interested in llama2.cpp are comparing it to the libraries listed below
Sorting:
- LLM training in simple, raw C/CUDA☆112Updated last year
- Inference of Mamba and Mamba2 models in pure C☆196Updated 2 weeks ago
- Python bindings for ggml☆147Updated last year
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆74Updated last year
- RWKV in nanoGPT style☆197Updated last year
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆169Updated 5 months ago
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆280Updated 2 years ago
- An innovative library for efficient LLM inference via low-bit quantization☆352Updated last year
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆93Updated this week
- llama.cpp to PyTorch Converter☆36Updated last year
- A C++ port of karpathy/llm.c features a tiny torch library while maintaining overall simplicity.☆42Updated last year
- Experiments with BitNet inference on CPU☆55Updated last year
- ☆71Updated 10 months ago
- ☆79Updated last year
- ☆120Updated last year
- ☆172Updated this week
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆155Updated last year
- llama3.cuda is a pure C/CUDA implementation for Llama 3 model.☆350Updated 9 months ago
- Easy and Efficient Quantization for Transformers☆204Updated last week
- High-Performance FP32 GEMM on CUDA devices☆117Updated last year
- Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs☆110Updated 2 years ago
- A collection of all available inference solutions for the LLMs☆94Updated 11 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆267Updated 2 months ago
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆306Updated last year
- A torchless, c++ rwkv implementation using 8bit quantization, written in cuda/hip/vulkan for maximum compatibility and minimum dependenci…☆313Updated 2 years ago
- nanogpt turned into a chat model☆81Updated 2 years ago
- 1.58-bit LLaMa model☆82Updated last year
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆131Updated 4 months ago
- Data preparation code for Amber 7B LLM☆94Updated last year
- ring-attention experiments☆165Updated last year