yichengchen24 / MIGLinks
[ACL2025 Findings] Official code for MIG: Automatic Data Selection for Instruction Tuning by Maximizing Information Gain in Semantic Space
β25Updated 3 months ago
Alternatives and similar repositories for MIG
Users that are interested in MIG are comparing it to the libraries listed below
Sorting:
- [ICLR 2025] 𧬠RegMix: Data Mixture as Regression for Language Model Pre-training (Spotlight)β181Updated 10 months ago
- β109Updated 5 months ago
- [ICML 2024] Selecting High-Quality Data for Training Language Modelsβ197Updated 3 weeks ago
- Official code for paper "UniIR: Training and Benchmarking Universal Multimodal Information Retrievers" (ECCV 2024)β174Updated last year
- [MM 2025] CMM-Math: A Chinese Multimodal Math Dataset To Evaluate and Enhance the Mathematics Reasoning of Large Multimodal Modelsβ48Updated last year
- Touchstone: Evaluating Vision-Language Models by Language Modelsβ83Updated last year
- [ACL 2024 Oral] This is the code repo for our ACLβ24 paper "MARVEL: Unlocking the Multi-Modal Capability of Dense Retrieval via Visual Moβ¦β39Updated last year
- [ACL'24] Superfiltering: Weak-to-Strong Data Filtering for Fast Instruction-Tuningβ184Updated 6 months ago
- Official Repository of MMLONGBENCH-DOC: Benchmarking Long-context Document Understanding with Visualizationsβ113Updated 3 months ago
- Scaling Preference Data Curation via Human-AI Synergyβ133Updated 5 months ago
- Attaching human-like eyes to the large language model. The codes of IEEE TMM paper "LMEye: An Interactive Perception Network for Large Laβ¦β48Updated last year
- Code for Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Modelsβ92Updated last year
- Repo for the EMNLP'24 Paper "Dual-Space Knowledge Distillation for Large Language Models". A general white-box KD framework for both sameβ¦β61Updated 4 months ago
- [2024-ACL]: TextBind: Multi-turn Interleaved Multimodal Instruction-following in the Wildrounded Conversationβ47Updated 2 years ago
- Pre-trained, Scalable, High-performance Reward Models via Policy Discriminative Learning.β163Updated 3 months ago
- β58Updated 10 months ago
- [NeurIPS'24] Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Modelsβ64Updated last year
- β100Updated 2 years ago
- Our code for ICLR'25 paper "DataMan: Data Manager for Pre-training Large Language Models".β109Updated 4 months ago
- [NAACL 2024] MMC: Advancing Multimodal Chart Understanding with LLM Instruction Tuningβ96Updated 11 months ago
- Official implementation of our paper "Finetuned Multimodal Language Models are High-Quality Image-Text Data Filters".β69Updated 8 months ago
- Implementations of online merging optimizers proposed by Online Merging Optimizers for Boosting Rewards and Mitigating Tax in Alignmentβ80Updated last year
- [ICLR 2024] CLEX: Continuous Length Extrapolation for Large Language Modelsβ78Updated last year
- RM-R1: Unleashing the Reasoning Potential of Reward Modelsβ156Updated 6 months ago
- β48Updated last year
- β87Updated 2 years ago
- [NeurIPS 2025 Spotlight] A Token is Worth over 1,000 Tokens: Efficient Knowledge Distillation through Low-Rank Clone.β43Updated 2 months ago
- [ACL 2024] Long-Context Language Modeling with Parallel Encodingsβ167Updated last year
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".β138Updated last year
- Are Intermediate Layers and Labels Really Necessary? A General Language Model Distillation Method ; GKD: A General Knowledge Distillationβ¦β33Updated 2 years ago