moritztng / fltr
Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.
β376Updated last year
Alternatives and similar repositories for fltr:
Users that are interested in fltr are comparing it to the libraries listed below
- From anywhere you can type, query and stream the output of an LLM or any other scriptβ494Updated last year
- Stateful load balancer custom-tailored for llama.cpp ππ¦β753Updated last week
- An AI assistant beyond the chat box.β328Updated last year
- Replace OpenAI with Llama.cpp Automagically.β318Updated 11 months ago
- A fast batching API to serve LLM modelsβ182Updated last year
- Stop messing around with finicky sampling parameters and just use DRΒ΅GS!β349Updated 11 months ago
- A multimodal, function calling powered LLM webui.β214Updated 7 months ago
- Mistral7B playing DOOMβ131Updated 9 months ago
- Minimal LLM inference in Rustβ983Updated 6 months ago
- Web UI for ExLlamaV2β493Updated 3 months ago
- Finetune llama2-70b and codellama on MacBook Air without quantizationβ448Updated last year
- All-in-one desktop app for running LLMs locally.β445Updated last week
- Clipboard Conqueror is a novel copy and paste copilot alternative designed to bring your very own LLM AI assistant to any text field.β400Updated 4 months ago
- Visualize the intermediate output of Mistral 7Bβ360Updated 3 months ago
- β288Updated last month
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibesβ196Updated 2 months ago
- LLM-based code completion engineβ185Updated 3 months ago
- Experimental LLM Inference UX to aid in creative writingβ116Updated 4 months ago
- LLM-powered lossless compression toolβ280Updated 8 months ago
- β163Updated 11 months ago
- C++ implementation for π«StarCoderβ454Updated last year
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM β¦β558Updated 2 months ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.β264Updated this week
- An implementation of bucketMul LLM inferenceβ217Updated 10 months ago
- LLaVA server (llama.cpp).β180Updated last year
- High-level, optionally asynchronous Rust bindings to llama.cppβ220Updated 11 months ago
- Run any ML model from any programming language.β423Updated last year
- OpenAI compatible API for serving LLAMA-2 modelβ218Updated last year
- GGUF implementation in C as a library and a tools CLI programβ270Updated 4 months ago
- Efficient visual programming for AI language modelsβ359Updated 8 months ago