InsTag: A Tool for Data Analysis in LLM Supervised Fine-tuning
☆285Aug 20, 2023Updated 2 years ago
Alternatives and similar repositories for InsTag
Users that are interested in InsTag are comparing it to the libraries listed below
Sorting:
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]☆588Dec 9, 2024Updated last year
- [NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other mo…☆416Jun 25, 2025Updated 8 months ago
- ☆324Jul 25, 2024Updated last year
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data …☆829Mar 17, 2025Updated 11 months ago
- [ACL 2024] FollowBench: A Multi-level Fine-grained Constraints Following Benchmark for Large Language Models☆119Jun 12, 2025Updated 8 months ago
- [ACL'24] Superfiltering: Weak-to-Strong Data Filtering for Fast Instruction-Tuning☆189Jun 25, 2025Updated 8 months ago
- Codes and Data for Scaling Relationship on Learning Mathematical Reasoning with Large Language Models☆270Sep 12, 2024Updated last year
- Benchmarking Complex Instruction-Following with Multiple Constraints Composition (NeurIPS 2024 Datasets and Benchmarks Track)☆102Feb 20, 2025Updated last year
- [ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning☆512Oct 20, 2024Updated last year
- A self-ailgnment method for role-play. Benchmark for role-play. Resources for "Large Language Models are Superpositions of All Characters…☆211May 28, 2024Updated last year
- [ICML 2024] Selecting High-Quality Data for Training Language Models☆201Dec 8, 2025Updated 2 months ago
- Collection of training data management explorations for large language models☆336Aug 2, 2024Updated last year
- [NIPS2023] RRHF & Wombat☆809Sep 22, 2023Updated 2 years ago
- Recipes to train reward model for RLHF.☆1,515Apr 24, 2025Updated 10 months ago
- OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, …☆6,688Updated this week
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning☆367Sep 6, 2024Updated last year
- 万卷1.0多模态语料☆569Oct 20, 2023Updated 2 years ago
- Official implementation of the paper "From Complex to Simple: Enhancing Multi-Constraint Complex Instruction Following Ability of Large L…☆53Jun 24, 2024Updated last year
- ☆84Apr 18, 2024Updated last year
- A large-scale, fine-grained, diverse preference dataset (and models).☆363Dec 29, 2023Updated 2 years ago
- A flexible and efficient training framework for large-scale alignment tasks☆450Oct 23, 2025Updated 4 months ago
- 大模型多维度中文对齐评测基准 (ACL 2024)☆420Oct 25, 2025Updated 4 months ago
- Data and tools for generating and inspecting OLMo pre-training data.☆1,411Nov 5, 2025Updated 3 months ago
- RewardBench: the first evaluation tool for reward models.☆696Feb 16, 2026Updated last week
- This is the official implementation of TAGCOS: Task-agnostic Gradient Clustered Coreset Selection for Instruction Tuning Data☆13Jul 21, 2024Updated last year
- Official Repo for Open-Reasoner-Zero☆2,087Jun 2, 2025Updated 8 months ago
- 800,000 step-level correctness labels on LLM solutions to MATH problems☆2,091Jun 1, 2023Updated 2 years ago
- ☆565Nov 20, 2024Updated last year
- ☆1,104Jan 10, 2026Updated last month
- Open Academic Research on Improving LLaMA to SOTA LLM☆1,611Aug 30, 2023Updated 2 years ago
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.☆459Apr 18, 2024Updated last year
- An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)☆9,037Feb 21, 2026Updated last week
- ☆72Apr 2, 2024Updated last year
- AllenAI's post-training codebase☆3,592Updated this week
- ☆148Jul 1, 2024Updated last year
- [EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs☆260Dec 16, 2024Updated last year
- [NeurIPS'24] Official code for *🎯DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving*☆120Dec 10, 2024Updated last year
- ☆87Dec 29, 2023Updated 2 years ago
- Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"☆391Jan 19, 2025Updated last year