intentee / paddlerLinks
Open-source LLMOps platform for hosting and scaling AI in your own infrastructure ππ¦
β1,098Updated this week
Alternatives and similar repositories for paddler
Users that are interested in paddler are comparing it to the libraries listed below
Sorting:
- Minimal LLM inference in Rustβ1,012Updated 10 months ago
- Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.β384Updated last year
- A cross-platform browser ML framework.β712Updated 9 months ago
- A high-performance inference engine for AI modelsβ1,264Updated this week
- VS Code extension for LLM-assisted code/text completionβ917Updated this week
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibesβ229Updated 2 weeks ago
- Scalable, fast, and disk-friendly vector search in Postgres, the successor of pgvecto.rs.β1,031Updated this week
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.β420Updated last week
- Rust framework for LLM orchestrationβ204Updated last year
- A realtime serving engine for Data-Intensive Generative AI Applicationsβ1,049Updated this week
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inferenceβ843Updated 3 weeks ago
- Super-fast Structured Outputsβ432Updated last week
- SeekStorm - sub-millisecond full-text search library & multi-tenancy server in Rustβ1,721Updated 2 weeks ago
- Model swapping for llama.cpp (or any local OpenAPI compatible server)β1,333Updated last week
- Production-ready Inference, Ingestion and Indexing built in Rust π¦β694Updated this week
- Fast, streaming indexing, query, and agentic LLM applications in Rustβ540Updated this week
- git-like rag pipelineβ243Updated 3 weeks ago
- Felafax is building AI infra for non-NVIDIA GPUsβ567Updated 7 months ago
- Official Rust Implementation of Model2Vecβ124Updated last month
- Rust library for generating vector embeddings, reranking. Re-write of qdrant/fastembed.β573Updated last week
- βΎοΈ Helix is a private GenAI stack for building AI agents with declarative pipelines, knowledge (RAG), API bindings, and first-class testiβ¦β515Updated this week
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM β¦β587Updated 6 months ago
- Fully neural approach for text chunkingβ368Updated 3 months ago
- A fast llama2 decoder in pure Rust.β1,055Updated last year
- Faster structured generationβ242Updated 3 months ago
- Split text into semantic chunks, up to a desired chunk size. Supports calculating length by characters and tokens, and is callable from Rβ¦β465Updated this week
- β342Updated this week
- Korvus is a search SDK that unifies the entire RAG pipeline in a single database query. Built on top of Postgres with bindings for Pythonβ¦β1,392Updated 6 months ago
- LLama.cpp rust bindingsβ398Updated last year
- High-level, optionally asynchronous Rust bindings to llama.cppβ228Updated last year