devanmolsharma / cachelmLinks
A semantic caching layer for LLM apps. It’s meant to cut down on repeated API calls even when the user phrases things differently
☆14Updated 7 months ago
Alternatives and similar repositories for cachelm
Users that are interested in cachelm are comparing it to the libraries listed below
Sorting:
- Code for evaluating with Flow-Judge-v0.1 - an open-source, lightweight (3.8B) language model optimized for LLM system evaluations. Crafte…☆84Updated last year
- A stable, fast and easy-to-use inference library with a focus on a sync-to-async API☆47Updated last year
- ☆134Updated last month
- Practical and advanced guide to LLMOps. It provides a solid understanding of large language models’ general concepts, deployment techniqu…☆78Updated last year
- Agentic RAG to help you build a startup🚀☆55Updated 10 months ago
- ☆24Updated last year
- Convert URLs into LLM-friendly markdown chunks☆65Updated last year
- A Lightweight Library for AI Observability☆255Updated 11 months ago
- Embed anything.☆27Updated last year
- 🔌 Want one client library for all your embeddings? 💙 Choose Catsu! 🐱☆57Updated 3 weeks ago
- The one who calls upon functions - Function-Calling Language Model☆36Updated 2 years ago
- Generate python documentation using LLMs☆71Updated last year
- 🚀 Scale your RAG pipeline using Ragswift: A scalable centralized embeddings management platform☆38Updated 2 years ago
- ☆207Updated last year
- Synthetic Data for LLM Fine-Tuning☆120Updated 2 years ago
- chrome & firefox extension to chat with webpages: local llms☆131Updated last year
- A simple Python sandbox for helpful LLM data agents☆305Updated last year
- Train an adapter for any embedding model in under a minute☆130Updated 10 months ago
- ☆30Updated last year
- Serving LLMs in the HF-Transformers format via a PyFlask API☆72Updated last year
- Code action agent with local execution sandbox and first-class support for programmatic tool calling☆121Updated this week
- Example demonstrating how to use gpt-4o-mini for fine-tuning☆28Updated last year
- Function Calling Benchmark & Testing☆92Updated last year
- Declarative framework to build LLM-based applications☆130Updated last year
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆100Updated 7 months ago
- Simple examples using Argilla tools to build AI☆57Updated last year
- A fast batching API to serve LLM models☆189Updated last year
- ☆159Updated 9 months ago
- Dynamic Metadata based RAG Framework☆78Updated 2 months ago
- Dataset Crafting w/ RAG/Wikipedia ground truth and Efficient Fine-Tuning Using MLX and Unsloth. Includes configurable dataset annotation …☆193Updated last year