run-llama / finetune-embedding
Fine-Tuning Embedding for RAG with Synthetic Data
☆469Updated last year
Related projects ⓘ
Alternatives and complementary repositories for finetune-embedding
- Domain Adapted Language Modeling Toolkit - E2E RAG☆311Updated last week
- Automated Evaluation of RAG Systems☆484Updated 2 weeks ago
- ☆433Updated 10 months ago
- Open-Source Implementation of WizardLM to turn documents into Q:A pairs for LLM fine-tuning☆295Updated 3 weeks ago
- RankLLM is a Python toolkit for reproducible information retrieval research using rerankers, with a focus on listwise reranking.☆349Updated last week
- ☆299Updated 11 months ago
- ☆800Updated 3 weeks ago
- ☆744Updated 10 months ago
- Forward-Looking Active REtrieval-augmented generation (FLARE)☆588Updated last year
- A lightweight library for generating synthetic instruction tuning datasets for your data without GPT.☆700Updated 2 months ago
- Lite & Super-fast re-ranking for your search & retrieval pipelines. Supports SoTA Listwise and Pairwise reranking based on LLMs and cro…☆665Updated last month
- Easily embed, cluster and semantically label text datasets☆462Updated 7 months ago
- Generative Representational Instruction Tuning☆567Updated this week
- HyDE: Precise Zero-Shot Dense Retrieval without Relevance Labels☆450Updated last year
- Generate textbook-quality synthetic LLM pretraining data☆488Updated last year
- Code for explaining and evaluating late chunking (chunked pooling)☆246Updated last month
- In-Context Learning for eXtreme Multi-Label Classification (XMC) using only a handful of examples.☆385Updated 9 months ago
- ☆489Updated 3 months ago
- wandbot is a technical support bot for Weights & Biases' AI developer tools that can run in Discord, Slack, ChatGPT and Zendesk☆275Updated 2 weeks ago
- This repository shares end-to-end notebooks on how to use various Weaviate features and integrations!☆570Updated this week
- ☆451Updated 3 weeks ago
- A lightweight, low-dependency, unified API to use all common reranking and cross-encoder models.☆1,095Updated last week
- Automatically evaluate your LLMs in Google Colab☆559Updated 6 months ago
- Best practices for distilling large language models.☆397Updated 9 months ago
- 🦜💯 Flex those feathers!☆234Updated 3 weeks ago
- ☆240Updated last year
- The official implementation of RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval☆966Updated 2 months ago
- Efficient Retrieval Augmentation and Generation Framework☆1,341Updated last week
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆675Updated 7 months ago
- Framework for enhancing LLMs for RAG tasks using fine-tuning.☆504Updated this week