kolinko / effort
An implementation of bucketMul LLM inference
☆214Updated 5 months ago
Alternatives and similar repositories for effort:
Users that are interested in effort are comparing it to the libraries listed below
- Mistral7B playing DOOM☆122Updated 5 months ago
- ☆235Updated 8 months ago
- Visualize the intermediate output of Mistral 7B☆322Updated 10 months ago
- Visualizing the internal board state of a GPT trained on chess PGN strings, and performing interventions on its internal board state and …☆195Updated 3 weeks ago
- Fast parallel LLM inference for MLX☆149Updated 5 months ago
- GGUF implementation in C as a library and a tools CLI program☆246Updated 5 months ago
- Stop messing around with finicky sampling parameters and just use DRµGS!☆319Updated 6 months ago
- WebGPU LLM inference tuned by hand☆147Updated last year
- A complete end-to-end pipeline for LLM interpretability with sparse autoencoders (SAEs) using Llama 3.2, written in pure PyTorch and full…☆603Updated 2 weeks ago
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆250Updated last year
- ☆113Updated last month
- ☆162Updated 6 months ago
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated 11 months ago
- ☆240Updated last month
- Inference of Mamba models in pure C☆181Updated 9 months ago
- Bayesian Optimization as a Coverage Tool for Evaluating LLMs. Accurate evaluation (benchmarking) that's 10 times faster with just a few l…☆271Updated 2 weeks ago
- a curated list of data for reasoning ai☆115Updated 4 months ago
- Run GGML models with Kubernetes.☆173Updated last year
- scalable and robust tree-based speculative decoding algorithm☆322Updated 4 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated 2 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆156Updated 7 months ago
- look how they massacred my boy☆62Updated 2 months ago
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆648Updated last week
- Tiny inference-only implementation of LLaMA☆91Updated 8 months ago
- run paligemma in real time☆128Updated 6 months ago
- Open weights language model from Google DeepMind, based on Griffin.☆612Updated 5 months ago
- Implement recursion using English as the programming language and an LLM as the runtime.☆129Updated last year
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆262Updated last year
- Finetune llama2-70b and codellama on MacBook Air without quantization☆448Updated 8 months ago
- Full finetuning of large language models without large memory requirements☆93Updated 11 months ago