stanford-oval / WikiChatLinks
WikiChat is an improved RAG. It stops the hallucination of large language models by retrieving data from a corpus.
☆1,465Updated 2 months ago
Alternatives and similar repositories for WikiChat
Users that are interested in WikiChat are comparing it to the libraries listed below
Sorting:
- [ICLR 2025] LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs☆1,673Updated this week
- ☆902Updated 9 months ago
- Optimizing inference proxy for LLMs☆2,574Updated this week
- High-performance retrieval engine for unstructured data☆1,419Updated last week
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆572Updated 4 months ago
- Framework for enhancing LLMs for RAG tasks using fine-tuning.☆742Updated last month
- A lightweight, low-dependency, unified API to use all common reranking and cross-encoder models.☆1,470Updated last month
- Software to implement GoT with a weviate vectorized database☆669Updated 3 months ago
- Efficient Retrieval Augmentation and Generation Framework☆1,578Updated 5 months ago
- [ACL 2023] One Embedder, Any Task: Instruction-Finetuned Text Embeddings☆1,973Updated 5 months ago
- A lightweight library for generating synthetic instruction tuning datasets for your data without GPT.☆780Updated 4 months ago
- Inference engine powering open source models on OpenRouter☆967Updated 5 months ago
- Empowering RAG with a memory-based data interface for all-purpose applications!☆1,831Updated 2 months ago
- Customizable implementation of the self-instruct paper.☆1,045Updated last year
- The Open Source Memory Layer For Autonomous Agents☆2,260Updated 8 months ago
- This includes the original implementation of SELF-RAG: Learning to Retrieve, Generate and Critique through self-reflection by Akari Asai,…☆2,113Updated last year
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆2,773Updated this week
- Llama-3 agents that can browse the web by following instructions and talking to you☆1,407Updated 6 months ago
- Easily use and train state of the art late-interaction retrieval methods (ColBERT) in any RAG pipeline. Designed for modularity and ease-…☆3,531Updated last month
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆382Updated last month
- Create Custom LLMs☆1,653Updated this week
- Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization☆1,316Updated 6 months ago
- Enforce the output format (JSON Schema, Regex etc) of a language model☆1,825Updated 4 months ago
- Python & JS/TS SDK for running AI-generated code/code interpreting in your AI app☆1,848Updated this week
- multi1: create o1-like reasoning chains with multiple AI providers (and locally). Supports LiteLLM as backend too for 100+ providers at o…☆347Updated 5 months ago
- LLaMA-Omni is a low-latency and high-quality end-to-end speech interaction model built upon Llama-3.1-8B-Instruct, aiming to achieve spee…☆2,942Updated last month
- g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains☆4,223Updated 5 months ago
- Knowledge Agents and Management in the Cloud☆4,031Updated this week
- A tool for generating function arguments and choosing what function to call with local LLMs☆428Updated last year
- Infinity is a high-throughput, low-latency serving engine for text-embeddings, reranking models, clip, clap and colpali☆2,262Updated this week