AlexBuz / llama-zipLinks
LLM-powered lossless compression tool
☆285Updated 9 months ago
Alternatives and similar repositories for llama-zip
Users that are interested in llama-zip are comparing it to the libraries listed below
Sorting:
- A fast batching API to serve LLM models☆181Updated last year
- Experimental adventure game with AI-generated content☆111Updated last month
- This is our own implementation of 'Layer Selective Rank Reduction'☆238Updated last year
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆72Updated 8 months ago
- ☆90Updated 5 months ago
- ☆291Updated 2 months ago
- Web UI for ExLlamaV2☆495Updated 4 months ago
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆767Updated last week
- Stop messing around with finicky sampling parameters and just use DRµGS!☆349Updated last year
- 1.58-bit LLaMa model☆81Updated last year
- Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.☆377Updated last year
- Python bindings for ggml☆141Updated 9 months ago
- Fast parallel LLM inference for MLX☆189Updated 10 months ago
- A multimodal, function calling powered LLM webui.☆214Updated 8 months ago
- ☆157Updated 10 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆153Updated 7 months ago
- Guaranteed Structured Output from any Language Model via Hierarchical State Machines☆134Updated this week
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆153Updated last year
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆139Updated 3 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆222Updated last year
- Blue-text Bot AI. Uses Ollama + AppleScript☆50Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆231Updated 7 months ago
- An implementation of bucketMul LLM inference☆217Updated 11 months ago
- Mistral7B playing DOOM☆131Updated 10 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆199Updated 10 months ago
- automatically quant GGUF models☆181Updated this week
- Let's create synthetic textbooks together :)☆75Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆171Updated last year
- LLM-based code completion engine☆188Updated 4 months ago
- Train your own small bitnet model☆71Updated 7 months ago