cmp-nct / ggllm.cpp
Falcon LLM ggml framework with CPU and GPU support
β244Updated 9 months ago
Related projects β
Alternatives and complementary repositories for ggllm.cpp
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRAβ124Updated last year
- C++ implementation for π«StarCoderβ445Updated last year
- Automated prompting and scoring framework to evaluate LLMs using updated human knowledge promptsβ111Updated last year
- β534Updated 11 months ago
- A fast batching API to serve LLM modelsβ172Updated 6 months ago
- TheBloke's Dockerfilesβ299Updated 8 months ago
- Extend the original llama.cpp repo to support redpajama model.β117Updated 2 months ago
- Tune any FALCON in 4-bitβ468Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hubβ155Updated last year
- Web UI for ExLlamaV2β438Updated last month
- β168Updated last year
- This is our own implementation of 'Layer Selective Rank Reduction'β231Updated 5 months ago
- SoTA Transformers with C-backend for fast inference on your CPU.β312Updated 11 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.β66Updated last year
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytesβ¦β143Updated last year
- GPT-2 small trained on phi-like dataβ65Updated 8 months ago
- Command-line script for inferencing from models such as falcon-7b-instructβ75Updated last year
- ggml implementation of BERTβ464Updated 8 months ago
- LLM-based code completion engineβ173Updated last year
- A prompt/context management systemβ165Updated last year
- The code we currently use to fine-tune models.β108Updated 6 months ago
- An unsupervised model merging algorithm for Transformers-based language models.β99Updated 6 months ago
- Simple, hackable and fast implementation for training/finetuning medium-sized LLaMA-based modelsβ152Updated this week
- Python bindings for ggmlβ132Updated 2 months ago
- Merge Transformers language models by use of gradient parameters.β202Updated 3 months ago
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.β242Updated last week
- LLaMa retrieval plugin script using OpenAI's retrieval pluginβ324Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMsβ77Updated 6 months ago
- Landmark Attention: Random-Access Infinite Context Length for Transformersβ415Updated 10 months ago