catid / bitnet_cpuLinks
Experiments with BitNet inference on CPU
☆54Updated last year
Alternatives and similar repositories for bitnet_cpu
Users that are interested in bitnet_cpu are comparing it to the libraries listed below
Sorting:
- GGML implementation of BERT model with Python bindings and quantization.☆55Updated last year
- RWKV-7: Surpassing GPT☆92Updated 7 months ago
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆72Updated 5 months ago
- Inference of Mamba models in pure C☆188Updated last year
- Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs☆110Updated last year
- Thin wrapper around GGML to make life easier☆36Updated 3 weeks ago
- A fast RWKV Tokenizer written in Rust☆46Updated this week
- Course Project for COMP4471 on RWKV☆17Updated last year
- QuIP quantization☆54Updated last year
- ☆49Updated 11 months ago
- Train your own small bitnet model☆74Updated 8 months ago
- An open source replication of the stawberry method that leverages Monte Carlo Search with PPO and or DPO☆29Updated last week
- ☆49Updated last year
- Lightweight toolkit package to train and fine-tune 1.58bit Language models☆81Updated last month
- RWKV in nanoGPT style☆191Updated last year
- Video+code lecture on building nanoGPT from scratch☆69Updated last year
- Port of Facebook's LLaMA model in C/C++☆22Updated last year
- Simple high-throughput inference library☆120Updated 2 months ago
- ☆56Updated 6 months ago
- Code for paper: "QuIP: 2-Bit Quantization of Large Language Models With Guarantees" adapted for Llama models☆38Updated last year
- Python bindings for ggml☆142Updated 10 months ago
- Testing LLM reasoning abilities with family relationship quizzes.☆62Updated 5 months ago
- GGUF parser in Python☆28Updated 11 months ago
- Collection of autoregressive model implementation☆85Updated 2 months ago
- Implementation of the Mamba SSM with hf_integration.☆56Updated 10 months ago
- ☆49Updated this week
- Tokun to can tokens☆18Updated 3 weeks ago
- tinygrad port of the RWKV large language model.☆45Updated 4 months ago
- Fast approximate inference on a single GPU with sparsity aware offloading☆38Updated last year
- Latent Large Language Models☆18Updated 10 months ago