chu-tianxiang / llama-cpp-torch
llama.cpp to PyTorch Converter
☆33Updated 11 months ago
Alternatives and similar repositories for llama-cpp-torch:
Users that are interested in llama-cpp-torch are comparing it to the libraries listed below
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆100Updated last year
- RWKV-7: Surpassing GPT☆82Updated 4 months ago
- A toolkit for fine-tuning, inferencing, and evaluating GreenBitAI's LLMs.☆79Updated 2 weeks ago
- ☆46Updated 8 months ago
- QuIP quantization☆52Updated last year
- Python bindings for ggml☆140Updated 6 months ago
- tinygrad port of the RWKV large language model.☆44Updated 3 weeks ago
- Experiments with BitNet inference on CPU☆53Updated last year
- QLoRA with Enhanced Multi GPU Support☆36Updated last year
- ☆73Updated last year
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆87Updated this week
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆155Updated 5 months ago
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated 10 months ago
- ☆74Updated last year
- Evaluating LLMs with Dynamic Data☆78Updated last month
- This is our own implementation of 'Layer Selective Rank Reduction'☆233Updated 10 months ago
- experiments with inference on llama☆104Updated 9 months ago
- ☆49Updated last year
- Inference of Mamba models in pure C☆187Updated last year
- ☆66Updated 10 months ago
- ☆53Updated 10 months ago
- Data preparation code for CrystalCoder 7B LLM☆44Updated 10 months ago
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆70Updated last month
- Data preparation code for Amber 7B LLM☆86Updated 10 months ago
- Gpu benchmark☆57Updated 2 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆117Updated last year
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆40Updated last year
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- A super simple web interface to perform blind tests on LLM outputs.☆28Updated last year