vibrantlabsai / nemesisLinks
Reward Model framework for LLM RLHF
β62Updated 2 years ago
Alternatives and similar repositories for nemesis
Users that are interested in nemesis are comparing it to the libraries listed below
Sorting:
- Open Implementations of LLM Analysesβ107Updated last year
- Lightweight demos for finetuning LLMs. Powered by π€ transformers and open-source datasets.β77Updated last year
- β85Updated 2 years ago
- Mixing Language Models with Self-Verification and Meta-Verificationβ111Updated last year
- Official repo for NAACL 2024 Findings paper "LeTI: Learning to Generate from Textual Interactions."β66Updated 2 years ago
- Codebase accompanying the Summary of a Haystack paper.β80Updated last year
- Small and Efficient Mathematical Reasoning LLMsβ73Updated last year
- Retrieval Augmented Generation Generalized Evaluation Datasetβ59Updated 5 months ago
- Meta-CoT: Generalizable Chain-of-Thought Prompting in Mixed-task Scenarios with Large Language Modelsβ100Updated 2 years ago
- Large-language Model Evaluation framework with Elo Leaderboard and A-B testingβ52Updated last year
- Supervised instruction finetuning for LLM with HF trainer and Deepspeedβ36Updated 2 years ago
- β162Updated last year
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo rankerβ126Updated 2 months ago
- Evaluating tool-augmented LLMs in conversation settingsβ88Updated last year
- β42Updated last year
- AuditNLG: Auditing Generative AI Language Modeling for Trustworthinessβ101Updated 11 months ago
- Code and data for "StructLM: Towards Building Generalist Models for Structured Knowledge Grounding" (COLM 2024)β75Updated last year
- Based on the tree of thoughts paperβ48Updated 2 years ago
- A set of utilities for running few-shot prompting experiments on large-language modelsβ126Updated 2 years ago
- Official repo of Respond-and-Respond: data, code, and evaluationβ103Updated last year
- Code accompanying "How I learned to start worrying about prompt formatting".β113Updated 7 months ago
- Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instruction collection encompassing various tasks. Vicuna is alβ¦β111Updated 2 years ago
- [NeurIPS 2023] This is the code for the paper `Large Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias`.β156Updated 2 years ago
- β44Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β51Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"β107Updated 2 years ago
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first appβ¦β169Updated last year
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 laβ¦β49Updated 2 years ago
- Code for NeurIPS LLM Efficiency Challengeβ59Updated last year
- Simple replication of [ColBERT-v1](https://arxiv.org/abs/2004.12832).β80Updated last year