intentee / paddlerLinks
Open-source LLMOps platform for hosting and scaling AI in your own infrastructure ππ¦
β1,312Updated 3 weeks ago
Alternatives and similar repositories for paddler
Users that are interested in paddler are comparing it to the libraries listed below
Sorting:
- Minimal LLM inference in Rustβ1,014Updated 11 months ago
- A high-performance inference engine for AI modelsβ1,323Updated last week
- Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.β385Updated last year
- A cross-platform browser ML framework.β718Updated 10 months ago
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inferenceβ896Updated last month
- VS Code extension for LLM-assisted code/text completionβ973Updated 2 weeks ago
- Scalable, fast, and disk-friendly vector search in Postgres, the successor of pgvecto.rs.β1,131Updated last week
- SeekStorm - sub-millisecond full-text search library & multi-tenancy server in Rustβ1,741Updated 2 weeks ago
- Korvus is a search SDK that unifies the entire RAG pipeline in a single database query. Built on top of Postgres with bindings for Pythonβ¦β1,451Updated 8 months ago
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.β473Updated last week
- Super-fast Structured Outputsβ539Updated last week
- Model swapping for llama.cpp (or any local OpenAI API compatible server)β1,615Updated last week
- Embeddable library or single binary for indexing and searching 1B vectorsβ304Updated 3 months ago
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibesβ237Updated 2 months ago
- Fully neural approach for text chunkingβ372Updated 5 months ago
- Big & Small LLMs working togetherβ1,170Updated last week
- A multi-platform desktop application to evaluate and compare LLM models, written in Rust and React.β843Updated 5 months ago
- Run any ML model from any programming language.β424Updated last year
- Rust framework for LLM orchestrationβ203Updated last year
- A realtime serving engine for Data-Intensive Generative AI Applicationsβ1,054Updated last week
- Things you can do with the token embeddings of an LLMβ1,448Updated 6 months ago
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.β612Updated 11 months ago
- Fast, streaming indexing, query, and agentic LLM applications in Rustβ577Updated last week
- Official Rust Implementation of Model2Vecβ138Updated last week
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM β¦β592Updated 7 months ago
- Rust library for generating vector embeddings, reranking. Re-write of qdrant/fastembed.β617Updated 2 weeks ago
- Local voice chatbot for engaging conversations, powered by Ollama, Hugging Face Transformers, and Coqui TTS Toolkitβ785Updated last year
- git-like rag pipelineβ245Updated last week
- Highly Performant, Modular, Memory Safe and Production-ready Inference, Ingestion and Indexing built in Rust π¦β722Updated last week
- βΎοΈ Helix is a private GenAI stack for building AI agents with declarative pipelines, knowledge (RAG), API bindings, and first-class testiβ¦β521Updated this week