xiatingyu / SFT-DataSelection-at-scaleLinks
☆30Updated 4 months ago
Alternatives and similar repositories for SFT-DataSelection-at-scale
Users that are interested in SFT-DataSelection-at-scale are comparing it to the libraries listed below
Sorting:
- Model merging is a highly efficient approach for long-to-short reasoning.☆65Updated 3 weeks ago
- [ACL-25] We introduce ScaleQuest, a scalable, novel and cost-effective data synthesis method to unleash the reasoning capability of LLMs.☆63Updated 7 months ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆122Updated 7 months ago
- Official code implementation for the ACL 2025 paper: 'CoT-based Synthesizer: Enhancing LLM Performance through Answer Synthesis'☆27Updated last month
- A Sober Look at Language Model Reasoning☆74Updated last week
- [NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$☆45Updated 8 months ago
- [ICLR 2025] 🧬 RegMix: Data Mixture as Regression for Language Model Pre-training (Spotlight)☆146Updated 4 months ago
- ☆65Updated 2 months ago
- Repo for the EMNLP'24 Paper "Dual-Space Knowledge Distillation for Large Language Models". A general white-box KD framework for both same…☆54Updated 7 months ago
- Official code for SEAL: Steerable Reasoning Calibration of Large Language Models for Free☆27Updated 2 months ago
- [ACL'24] Superfiltering: Weak-to-Strong Data Filtering for Fast Instruction-Tuning☆158Updated 9 months ago
- ☆15Updated 6 months ago
- [ACL 2024] Long-Context Language Modeling with Parallel Encodings☆154Updated last year
- [ICML 2024] Selecting High-Quality Data for Training Language Models☆176Updated last year
- ☆101Updated 8 months ago
- Official github repo for AutoDetect, an automated weakness detection framework for LLMs.☆42Updated last year
- ☆116Updated 3 weeks ago
- ☆81Updated last year
- [ICML'2024] Can AI Assistants Know What They Don't Know?☆81Updated last year
- This is an official implementation of the Reward rAnked Fine-Tuning Algorithm (RAFT), also known as iterative best-of-n fine-tuning or re…☆32Updated 9 months ago
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆156Updated 3 months ago
- Code for ACL 2024 accepted paper titled "SAPT: A Shared Attention Framework for Parameter-Efficient Continual Learning of Large Language …☆35Updated 5 months ago
- [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".☆78Updated 5 months ago
- Implementation for the research paper "Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision".☆54Updated 6 months ago
- [NeurIPS'24] Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models☆61Updated 6 months ago
- [ACL 2024] Learning to Edit: Aligning LLMs with Knowledge Editing☆36Updated 10 months ago
- ☆46Updated 2 months ago
- ACL'2025: SoftCoT: Soft Chain-of-Thought for Efficient Reasoning with LLMs. and preprint: SoftCoT++: Test-Time Scaling with Soft Chain-of…☆28Updated 3 weeks ago
- [ICLR 25 Oral] RM-Bench: Benchmarking Reward Models of Language Models with Subtlety and Style☆48Updated last month
- Laser: Learn to Reason Efficiently with Adaptive Length-based Reward Shaping☆47Updated last month