marella / ctransformers
Python bindings for the Transformer models implemented in C/C++ using GGML library.
☆1,842Updated last year
Alternatives and similar repositories for ctransformers:
Users that are interested in ctransformers are comparing it to the libraries listed below
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆2,823Updated last year
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.☆4,702Updated last month
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆3,970Updated last week
- 4 bits quantization of LLaMA using GPTQ☆3,036Updated 7 months ago
- AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:☆1,946Updated last month
- Customizable implementation of the self-instruct paper.☆1,038Updated 11 months ago
- S-LoRA: Serving Thousands of Concurrent LoRA Adapters☆1,790Updated last year
- Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".☆2,031Updated 10 months ago
- [MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration☆2,743Updated last week
- Alpaca dataset from Stanford, cleaned and curated☆1,537Updated last year
- Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs☆2,362Updated last week
- Fine-tune mistral-7B on 3090s, a100s, h100s☆705Updated last year
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆2,456Updated 6 months ago
- INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model☆1,472Updated this week
- RayLLM - LLMs on Ray☆1,257Updated 8 months ago
- Accessible large language models via k-bit quantization for PyTorch.☆6,697Updated this week
- Tune any FALCON in 4-bit☆466Updated last year
- [ACL 2023] One Embedder, Any Task: Instruction-Finetuned Text Embeddings☆1,915Updated last month
- MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.☆1,965Updated last week
- YaRN: Efficient Context Window Extension of Large Language Models☆1,421Updated 10 months ago
- Python bindings for llama.cpp☆8,647Updated 3 weeks ago
- Fast inference engine for Transformer models☆3,602Updated last week
- Inference Llama 2 in one file of pure 🔥☆2,107Updated 9 months ago
- Large-scale LLM inference engine☆1,295Updated this week
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆687Updated 10 months ago
- C++ implementation for BLOOM☆810Updated last year
- Public repo for the NeurIPS 2023 paper "Unlimiformer: Long-Range Transformers with Unlimited Length Input"☆1,060Updated 11 months ago
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions☆818Updated last year
- A blazing fast inference solution for text embeddings models☆3,175Updated 3 weeks ago
- ggml implementation of BERT☆480Updated 11 months ago