e-p-armstrong / augmentoolkitLinks
Create Custom LLMs
☆1,722Updated 2 weeks ago
Alternatives and similar repositories for augmentoolkit
Users that are interested in augmentoolkit are comparing it to the libraries listed below
Sorting:
- Large-scale LLM inference engine☆1,530Updated this week
- Optimizing inference proxy for LLMs☆2,812Updated this week
- ☆1,060Updated 11 months ago
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.☆607Updated 10 months ago
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆2,863Updated this week
- Software to implement GoT with a weviate vectorized database☆675Updated 5 months ago
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆586Updated 6 months ago
- What If Language Models Expertly Routed All Inference? WilmerAI allows prompts to be routed to specialized workflows based on the domain …☆764Updated this week
- Ingest files for retrieval augmented generation (RAG) with open-source Large Language Models (LLMs), all without 3rd parties or sensitive…☆688Updated last year
- Simple Python library/structure to ablate features in LLMs which are supported by TransformerLens☆504Updated last year
- Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs☆3,397Updated 3 months ago
- Synthetic data curation for post-training and structured data extraction☆1,488Updated last month
- Convenience scripts to finetune (chat-)LLaMa3 and other models for any language☆312Updated last year
- A lightweight, low-dependency, unified API to use all common reranking and cross-encoder models.☆1,523Updated 3 months ago
- Generic rag framework to apply the power of LLMs on any given dataset☆638Updated 2 months ago
- A framework for serving and evaluating LLM routers - save LLM costs without compromising quality☆4,227Updated last year
- Infinity is a high-throughput, low-latency serving engine for text-embeddings, reranking models, clip, clap and colpali☆2,398Updated this week
- Build datasets using natural language☆522Updated 3 months ago
- Automatically evaluate your LLMs in Google Colab☆655Updated last year
- An AI memory layer with short- and long-term storage, semantic clustering, and optional memory decay for context-aware applications.☆656Updated 7 months ago
- LLM Frontend in a single html file☆635Updated 7 months ago
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆335Updated 6 months ago
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,042Updated this week
- A toolkit to create optimal Production-readyRetrieval Augmented Generation(RAG) setup for your data☆1,465Updated 3 months ago
- WikiChat is an improved RAG. It stops the hallucination of large language models by retrieving data from a corpus.☆1,507Updated 4 months ago
- Autonomously train research-agent LLMs on custom data using reinforcement learning and self-verification.☆660Updated 5 months ago
- This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?☆1,261Updated last month
- function calling-based LLM agents☆287Updated 11 months ago
- Curated list of datasets and tools for post-training.☆3,404Updated last month
- Model swapping for llama.cpp (or any local OpenAI API compatible server)☆1,432Updated this week