moritztng / fltrLinks
Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.
β381Updated last year
Alternatives and similar repositories for fltr
Users that are interested in fltr are comparing it to the libraries listed below
Sorting:
- Stateful load balancer custom-tailored for llama.cpp ππ¦β792Updated this week
- Stop messing around with finicky sampling parameters and just use DRΒ΅GS!β349Updated last year
- From anywhere you can type, query and stream the output of an LLM or any other scriptβ497Updated last year
- LLM-powered lossless compression toolβ283Updated 11 months ago
- Replace OpenAI with Llama.cpp Automagically.β320Updated last year
- An implementation of bucketMul LLM inferenceβ220Updated last year
- An AI assistant beyond the chat box.β328Updated last year
- Clipboard Conqueror is a novel copy and paste copilot alternative designed to bring your very own LLM AI assistant to any text field.β413Updated 6 months ago
- All-in-one desktop app for running LLMs locally.β456Updated this week
- Web UI for ExLlamaV2β503Updated 5 months ago
- Mistral7B playing DOOMβ132Updated last year
- A fast batching API to serve LLM modelsβ183Updated last year
- Visualize the intermediate output of Mistral 7Bβ366Updated 5 months ago
- Minimal LLM inference in Rustβ1,003Updated 8 months ago
- β308Updated 3 months ago
- Guaranteed Structured Output from any Language Model via Hierarchical State Machinesβ140Updated last month
- Blue-text Bot AI. Uses Ollama + AppleScriptβ50Updated last year
- function calling-based LLM agentsβ287Updated 10 months ago
- Open source alternative to Perplexity AI with ability to run locallyβ213Updated 9 months ago
- WebGPU LLM inference tuned by handβ151Updated 2 years ago
- A multimodal, function calling powered LLM webui.β214Updated 9 months ago
- β204Updated last month
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibesβ207Updated 5 months ago
- git-like rag pipelineβ234Updated last week
- β163Updated last year
- Run any ML model from any programming language.β422Updated last year
- Live-bending a foundation modelβs output at neural network level.β263Updated 3 months ago
- Finetune llama2-70b and codellama on MacBook Air without quantizationβ447Updated last year
- LLM Frontend in a single html fileβ517Updated 6 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.β156Updated last year