Codys12 / airllmLinks
AirLLM 70B inference with single 4GB GPU
☆14Updated 5 months ago
Alternatives and similar repositories for airllm
Users that are interested in airllm are comparing it to the libraries listed below
Sorting:
- ☆24Updated 10 months ago
- run ollama & gguf easily with a single command☆52Updated last year
- The hearth of The Pulsar App, fast, secure and shared inference with modern UI☆59Updated last year
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆85Updated this week
- Glyphs, acting as collaboratively defined symbols linking related concepts, add a layer of multidimensional semantic richness to user-AI …☆54Updated 10 months ago
- "a towel is about the most massively useful thing an interstellar AI hitchhiker can have"☆48Updated last year
- Yet another frontend for LLM, written using .NET and WinUI 3☆10Updated 3 months ago
- Tcurtsni: Reverse Instruction Chat, ever wonder what your LLM wants to ask you?☆23Updated last year
- Yet Another (LLM) Web UI, made with Gemini☆12Updated 11 months ago
- fast state-of-the-art speech models and a runtime that runs anywhere 💥☆57Updated 6 months ago
- ☆17Updated 11 months ago
- Create text chunks which end at natural stopping points without using a tokenizer☆26Updated 2 weeks ago
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆47Updated last month
- Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vecto…☆43Updated last year
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Updated last year
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆99Updated 5 months ago
- Serving LLMs in the HF-Transformers format via a PyFlask API☆72Updated last year
- ☆68Updated last year
- ☆108Updated 3 months ago
- Unleash the full potential of exascale LLMs on consumer-class GPUs, proven by extensive benchmarks, with no long-term adjustments and min…☆26Updated last year
- an auto-sleeping and -waking framework around llama.cpp☆12Updated 10 months ago
- Easily convert HuggingFace models to GGUF-format for llama.cpp☆23Updated last year
- Smart proxy for LLM APIs that enables model-specific parameter control, automatic mode switching (like Qwen3's /think and /no_think), and…☆51Updated 6 months ago
- convert a saved pytorch model to gguf and generate as much corresponding ggml c code as possible☆15Updated last year
- Easy to use, High Performant Knowledge Distillation for LLMs☆97Updated 7 months ago
- Local LLM inference & management server with built-in OpenAI API☆31Updated last year
- ☆23Updated last year
- A sleek, customizable interface for managing LLMs with responsive design and easy agent personalization.☆17Updated last year
- Lightweight continuous batching OpenAI compatibility using HuggingFace Transformers include T5 and Whisper.☆29Updated 8 months ago
- Light WebUI for lm.rs☆24Updated last year