okuvshynov / llama_duoLinks
asynchronous/distributed speculative evaluation for llama3
☆39Updated last year
Alternatives and similar repositories for llama_duo
Users that are interested in llama_duo are comparing it to the libraries listed below
Sorting:
- Inference of Mamba models in pure C☆191Updated last year
- A minimalistic C++ Jinja templating engine for LLM chat templates☆180Updated last week
- A faithful clone of Karpathy's llama2.c (one file inference, zero dependency) but fully functional with LLaMA 3 8B base and instruct mode…☆135Updated last year
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆52Updated 6 months ago
- Lightweight Llama 3 8B Inference Engine in CUDA C☆49Updated 5 months ago
- LLM training in simple, raw C/CUDA☆104Updated last year
- High-Performance SGEMM on CUDA devices☆101Updated 7 months ago
- The Finite Field Assembly Programming Language☆36Updated 4 months ago
- Custom PTX Instruction Benchmark☆127Updated 6 months ago
- iterate quickly with llama.cpp hot reloading. use the llama.cpp bindings with bun.sh☆51Updated last year
- Samples of good AI generated CUDA kernels☆90Updated 3 months ago
- minimal C implementation of speculative decoding based on llama2.c☆25Updated last year
- GGUF implementation in C as a library and a tools CLI program☆290Updated 3 weeks ago
- Thin wrapper around GGML to make life easier☆40Updated 2 months ago
- Python bindings for ggml☆146Updated last year
- GGML implementation of BERT model with Python bindings and quantization.☆56Updated last year
- Fast and vectorizable algorithms for searching in a vector of sorted floating point numbers☆149Updated 9 months ago
- Experiments with BitNet inference on CPU☆54Updated last year
- C API for MLX☆131Updated 2 weeks ago
- Inference RWKV v7 in pure C.☆38Updated 3 weeks ago
- ☆217Updated 7 months ago
- Simple high-throughput inference library☆127Updated 4 months ago
- Repository for the QUIK project, enabling the use of 4bit kernels for generative inference - EMNLP 2024☆182Updated last year
- Learning about CUDA by writing PTX code.☆135Updated last year
- WebGPU LLM inference tuned by hand☆151Updated 2 years ago
- Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs☆110Updated last year
- Editor with LLM generation tree exploration☆75Updated 7 months ago
- Tiny Dream - An embedded, Header Only, Stable Diffusion C++ implementation☆265Updated last year
- ☆53Updated last year
- tiny code to access tenstorrent blackhole☆59Updated 3 months ago