Spico197 / MoE-SFT
πΌ Official implementation of Dynamic Data Mixing Maximizes Instruction Tuning for Mixture-of-Experts
β38Updated 5 months ago
Alternatives and similar repositories for MoE-SFT:
Users that are interested in MoE-SFT are comparing it to the libraries listed below
- We introduce ScaleQuest, a scalable, novel and cost-effective data synthesis method to unleash the reasoning capability of LLMs.β60Updated 4 months ago
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.β54Updated 8 months ago
- [ICML'2024] Can AI Assistants Know What They Don't Know?β79Updated last year
- [ACL 2024 (Oral)] A Prospector of Long-Dependency Data for Large Language Modelsβ54Updated 7 months ago
- Official implementation of the paper "From Complex to Simple: Enhancing Multi-Constraint Complex Instruction Following Ability of Large Lβ¦β46Updated 8 months ago
- Code and data for "Living in the Moment: Can Large Language Models Grasp Co-Temporal Reasoning?" (ACL 2024)β32Updated 8 months ago
- Code and data for "ConflictBank: A Benchmark for Evaluating the Influence of Knowledge Conflicts in LLM" (NeurIPS 2024 Track Datasets andβ¦β36Updated 4 months ago
- L-CITEEVAL: DO LONG-CONTEXT MODELS TRULY LEVERAGE CONTEXT FOR RESPONDING?β23Updated 4 months ago
- [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".β73Updated 2 months ago
- Towards Systematic Measurement for Long Text Qualityβ33Updated 6 months ago
- BeHonest: Benchmarking Honesty in Large Language Modelsβ31Updated 7 months ago
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"β63Updated last year
- [ICLR'24 spotlight] Tool-Augmented Reward Modelingβ44Updated 2 months ago
- Intuitive Fine-Tuning: Towards Simplifying Alignment into a Single Processβ24Updated 7 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"β72Updated 9 months ago
- Source code for Truth-Aware Context Selection: Mitigating the Hallucinations of Large Language Models Being Misled by Untruthful Contextsβ17Updated 6 months ago
- β29Updated 2 months ago
- The implementation of paper "LLM Critics Help Catch Bugs in Mathematics: Towards a Better Mathematical Verifier with Natural Language Feeβ¦β38Updated 7 months ago
- M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoningβ55Updated 2 months ago
- [ICLR'25] Data and code for our paper "Why Does the Effective Context Length of LLMs Fall Short?"β70Updated 3 months ago
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"β58Updated last year
- β59Updated 6 months ago
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoningβ24Updated last year
- [EMNLP 2023] Plan, Verify and Switch: Integrated Reasoning with Diverse X-of-Thoughtsβ26Updated last year
- The official repository of the Omni-MATH benchmark.β74Updated 2 months ago
- The official code repository for PRMBench.β68Updated last month
- β17Updated 3 months ago
- The rule-based evaluation subset and code implementation of Omni-MATHβ17Updated 2 months ago
- Code for M4LE: A Multi-Ability Multi-Range Multi-Task Multi-Domain Long-Context Evaluation Benchmark for Large Language Modelsβ22Updated 7 months ago
- [ICLR 25 Oral] RM-Bench: Benchmarking Reward Models of Language Models with Subtlety and Styleβ26Updated last month