vibrantlabsai / nemesisLinks
Reward Model framework for LLM RLHF
☆61Updated 2 years ago
Alternatives and similar repositories for nemesis
Users that are interested in nemesis are comparing it to the libraries listed below
Sorting:
- Open Implementations of LLM Analyses☆108Updated last year
- Supervised instruction finetuning for LLM with HF trainer and Deepspeed☆36Updated 2 years ago
- Small and Efficient Mathematical Reasoning LLMs☆73Updated last year
- Based on the tree of thoughts paper☆48Updated 2 years ago
- Codebase accompanying the Summary of a Haystack paper.☆79Updated last year
- Large-language Model Evaluation framework with Elo Leaderboard and A-B testing☆52Updated last year
- ☆85Updated 2 years ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆78Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆110Updated last year
- Meta-CoT: Generalizable Chain-of-Thought Prompting in Mixed-task Scenarios with Large Language Models☆100Updated 2 years ago
- A set of utilities for running few-shot prompting experiments on large-language models☆126Updated 2 years ago
- Official repo for NAACL 2024 Findings paper "LeTI: Learning to Generate from Textual Interactions."☆66Updated 2 years ago
- [SIGIR 2024 (Demo)] CoSearchAgent: A Lightweight Collborative Search Agent with Large Language Models☆28Updated last year
- Code and data for "StructLM: Towards Building Generalist Models for Structured Knowledge Grounding" (COLM 2024)☆75Updated last year
- Finding semantically meaningful and accurate prompts.☆48Updated 2 years ago
- ☆86Updated last year
- ☆129Updated last year
- PyTorch implementation for MRL☆20Updated last year
- AuditNLG: Auditing Generative AI Language Modeling for Trustworthiness☆101Updated 10 months ago
- Retrieval Augmented Generation Generalized Evaluation Dataset☆59Updated 5 months ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Updated last year
- A repository for transformer critique learning and generation☆89Updated 2 years ago
- Source codes for the paper "Bounding the Capabilities of Large Language Models in Open Text Generation with Prompt Constraints"☆27Updated 2 years ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆51Updated last year
- [NAACL 2024] Struc-Bench: Are Large Language Models Good at Generating Complex Structured Tabular Data? https://aclanthology.org/2024.naa…☆55Updated 4 months ago
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo ranker☆125Updated last month
- SCREWS: A Modular Framework for Reasoning with Revisions☆27Updated 2 years ago
- ☆43Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆107Updated 2 years ago
- Supercharge huggingface transformers with model parallelism.☆77Updated 5 months ago