liangyuwang / zo2
ZO2 (Zeroth-Order Offloading): Full Parameter Fine-Tuning 175B LLMs with 18GB GPU Memory
☆59Updated this week
Alternatives and similar repositories for zo2:
Users that are interested in zo2 are comparing it to the libraries listed below
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆96Updated last month
- A Framework for Decoupling and Assessing the Capabilities of VLMs☆41Updated 8 months ago
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆106Updated last week
- This project aims to collect and collate various datasets for multimodal large model training, including but not limited to pre-training …☆34Updated 5 months ago
- ☆73Updated last year
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆18Updated last month
- Multimodal Open-O1 (MO1) is designed to enhance the accuracy of inference models by utilizing a novel prompt-based approach. This tool wo…☆29Updated 6 months ago
- An benchmark for evaluating the capabilities of large vision-language models (LVLMs)☆45Updated last year
- ☆26Updated last month
- ☆68Updated 2 months ago
- Efficient Mixture of Experts for LLM Paper List☆47Updated 3 months ago
- [ICLR 2025] Benchmarking Agentic Workflow Generation☆62Updated last month
- [NeurIPS 2024] CharXiv: Charting Gaps in Realistic Chart Understanding in Multimodal LLMs☆99Updated 3 months ago
- The huggingface implementation of Fine-grained Late-interaction Multi-modal Retriever.☆83Updated 2 months ago
- [ACL 2024] Multi-modal preference alignment remedies regression of visual instruction tuning on language model☆37Updated 4 months ago
- From Hours to Minutes: Lossless Acceleration of Ultra Long Sequence Generation☆80Updated last week
- ☆17Updated last week
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆55Updated last month
- This is a repo for showcasing using MCTS with LLMs to solve gsm8k problems☆67Updated this week
- ☆83Updated 2 weeks ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆116Updated 4 months ago
- Official completion of “Training on the Benchmark Is Not All You Need”.☆30Updated 2 months ago
- MMR1: Advancing the Frontiers of Multimodal Reasoning☆145Updated last week
- SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning. COLM 2024 Accepted Paper☆30Updated 9 months ago
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆158Updated this week
- Exploring Efficient Fine-Grained Perception of Multimodal Large Language Models☆60Updated 4 months ago
- [ICLR 2025] MiniPLM: Knowledge Distillation for Pre-Training Language Models☆34Updated 4 months ago
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆62Updated 3 weeks ago
- ☆35Updated 3 weeks ago
- Open-Pandora: On-the-fly Control Video Generation☆32Updated 3 months ago