facebookresearch / llm-cross-capabilities
Official implementation for "Law of the Weakest Link: Cross capabilities of Large Language Models"
☆42Updated 5 months ago
Alternatives and similar repositories for llm-cross-capabilities:
Users that are interested in llm-cross-capabilities are comparing it to the libraries listed below
- [NeurIPS 2024] Train LLMs with diverse system messages reflecting individualized preferences to generalize to unseen system messages☆44Updated 3 months ago
- [𝐄𝐌𝐍𝐋𝐏 𝐅𝐢𝐧𝐝𝐢𝐧𝐠𝐬 𝟐𝟎𝟐𝟒 & 𝐀𝐂𝐋 𝟐𝟎𝟐𝟒 𝐍𝐋𝐑𝐒𝐄 𝐎𝐫𝐚𝐥] 𝘌𝘯𝘩𝘢𝘯𝘤𝘪𝘯𝘨 𝘔𝘢𝘵𝘩𝘦𝘮𝘢𝘵𝘪𝘤𝘢𝘭 𝘙𝘦𝘢𝘴𝘰𝘯𝘪𝘯…☆48Updated 10 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆72Updated 9 months ago
- [NeurIPS 2024] OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AI☆96Updated last week
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆44Updated 2 months ago
- [EMNLP 2023, Findings] GRACE: Discriminator-Guided Chain-of-Thought Reasoning☆47Updated 5 months ago
- ☆22Updated 3 months ago
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆136Updated 4 months ago
- Large Language Models Can Self-Improve in Long-context Reasoning☆62Updated 3 months ago
- [NeurIPS 2024] A comprehensive benchmark for evaluating critique ability of LLMs☆39Updated 3 months ago
- Critique-out-Loud Reward Models☆53Updated 5 months ago
- We introduce ScaleQuest, a scalable, novel and cost-effective data synthesis method to unleash the reasoning capability of LLMs.☆60Updated 4 months ago
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆130Updated 5 months ago
- ☆59Updated 6 months ago
- FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions☆43Updated 8 months ago
- ☆34Updated 11 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆46Updated last year
- Benchmarking Benchmark Leakage in Large Language Models☆51Updated 9 months ago
- A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.☆84Updated last year
- BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval☆86Updated last month
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"☆58Updated last year
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆79Updated 7 months ago
- ☆29Updated 2 months ago
- Unofficial Implementation of Chain-of-Thought Reasoning Without Prompting☆27Updated last year
- ☆43Updated 4 months ago
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆80Updated 5 months ago