moritztng / fltrLinks
Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.
β381Updated last year
Alternatives and similar repositories for fltr
Users that are interested in fltr are comparing it to the libraries listed below
Sorting:
- Stateful load balancer custom-tailored for llama.cpp ππ¦β782Updated this week
- A fast batching API to serve LLM modelsβ183Updated last year
- LLM-powered lossless compression toolβ285Updated 10 months ago
- Stop messing around with finicky sampling parameters and just use DRΒ΅GS!β349Updated last year
- Visualize the intermediate output of Mistral 7Bβ367Updated 5 months ago
- Replace OpenAI with Llama.cpp Automagically.β318Updated last year
- Guaranteed Structured Output from any Language Model via Hierarchical State Machinesβ136Updated 3 weeks ago
- From anywhere you can type, query and stream the output of an LLM or any other scriptβ497Updated last year
- Minimal LLM inference in Rustβ995Updated 8 months ago
- An implementation of bucketMul LLM inferenceβ217Updated 11 months ago
- Web UI for ExLlamaV2β502Updated 4 months ago
- An AI assistant beyond the chat box.β328Updated last year
- A multimodal, function calling powered LLM webui.β214Updated 9 months ago
- β301Updated 2 months ago
- Clipboard Conqueror is a novel copy and paste copilot alternative designed to bring your very own LLM AI assistant to any text field.β408Updated 5 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.β154Updated last year
- Mistral7B playing DOOMβ132Updated 11 months ago
- A cross-platform browser ML framework.β702Updated 7 months ago
- Super-simple, fully Rust powered "memory" (doc store + semantic search) for LLM projects, semantic search, etc.β62Updated last year
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)β567Updated last year
- The Fast Vector Similarity Library is designed to provide efficient computation of various similarity measures between vectors.β393Updated 4 months ago
- β130Updated last year
- Official Rust Implementation of Model2Vecβ118Updated 3 weeks ago
- Finetune llama2-70b and codellama on MacBook Air without quantizationβ447Updated last year
- This is our own implementation of 'Layer Selective Rank Reduction'β239Updated last year
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datasβ¦β179Updated last week
- function calling-based LLM agentsβ288Updated 9 months ago
- Falcon LLM ggml framework with CPU and GPU supportβ246Updated last year
- Accelerate your Hugging Face Transformers 7.6-9x. Native to Hugging Face and PyTorch.β683Updated 10 months ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.β273Updated last week