multiplexerai / Namespace-RAGLinks
☆13Updated last year
Alternatives and similar repositories for Namespace-RAG
Users that are interested in Namespace-RAG are comparing it to the libraries listed below
Sorting:
- ☆25Updated last year
- ☆38Updated last year
- GRDN.AI app for garden optimization☆70Updated last year
- A simple experiment on letting two local LLM have a conversation about anything!☆110Updated last year
- Gradio based tool to run opensource LLM models directly from Huggingface☆93Updated last year
- Experimental LLM Inference UX to aid in creative writing☆114Updated 7 months ago
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Updated last year
- After my server ui improvements were successfully merged, consider this repo a playground for experimenting, tinkering and hacking around…☆54Updated 10 months ago
- All the world is a play, we are but actors in it.☆50Updated this week
- GPT-4 Level Conversational QA Trained In a Few Hours☆62Updated 10 months ago
- Serving LLMs in the HF-Transformers format via a PyFlask API☆71Updated 10 months ago
- Minimal, clean code implementation of RAG with mlx using gguf model weights☆52Updated last year
- A stable, fast and easy-to-use inference library with a focus on a sync-to-async API☆45Updated 9 months ago
- LLaVA server (llama.cpp).☆180Updated last year
- ☆49Updated last year
- Embed anything.☆28Updated last year
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆37Updated 2 years ago
- ☆17Updated 6 months ago
- Local LLM inference & management server with built-in OpenAI API☆31Updated last year
- run ollama & gguf easily with a single command☆52Updated last year
- For inferring and serving local LLMs using the MLX framework☆104Updated last year
- Efficient computer use agent powered by Meta Llama 4 Maverick☆40Updated 2 months ago
- ☆101Updated last month
- ☆43Updated 3 months ago
- ☆115Updated 6 months ago
- A Python library to orchestrate LLMs in a neural network-inspired structure☆49Updated 9 months ago
- Very basic framework for composable parameterized large language model (Q)LoRA / (Q)Dora fine-tuning using mlx, mlx_lm, and OgbujiPT.☆42Updated 3 weeks ago
- Local LLaMAs/Models in VSCode☆53Updated 2 years ago
- A fast batching API to serve LLM models☆183Updated last year
- Scripts to create your own moe models using mlx☆90Updated last year