JerryYin777 / Jerry_CVLinks
☆10Updated last year
Alternatives and similar repositories for Jerry_CV
Users that are interested in Jerry_CV are comparing it to the libraries listed below
Sorting:
- ☆18Updated 10 months ago
- 学术双语简历模板,涵盖教育背景、论文发表、项目经历、竞赛经历和个人陈述等关键部分,可适用于申请研究生项目、学术职位或相关行业岗位。☆114Updated 2 months ago
- Recent Advances on MLLM's Reasoning Ability☆25Updated 4 months ago
- 🔥 【Meta Awesome List】: AI/ML Research Hub - Solving the "Chasing Hot Topics" Problem for AI Researchers. 🤖 Agent-driven intelligence au…☆42Updated this week
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆154Updated 2 months ago
- ☆115Updated last year
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆222Updated 9 months ago
- ☆148Updated 11 months ago
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆80Updated 2 months ago
- A curated list of awesome papers on dataset reduction, including dataset distillation (dataset condensation) and dataset pruning (coreset…☆58Updated 7 months ago
- Code release for VTW (AAAI 2025 Oral)☆49Updated last month
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆126Updated last month
- Awesome Low-Rank Adaptation☆43Updated last month
- [ICLR 2025] Mitigating Modality Prior-Induced Hallucinations in Multimodal Large Language Models via Deciphering Attention Causality☆38Updated 2 months ago
- [ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibrati…☆42Updated last year
- GPUSnatcher is a tool for GPU resource monitoring and snatching, designed to help users temporarily monitor and grab idle GPU resources.☆78Updated this week
- ☆49Updated 9 months ago
- ICML 2025 Oral: ABKD: Pursuing a Proper Allocation of the Probability Mass in Knowledge Distillation via α-β-Divergence☆35Updated 3 weeks ago
- An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT☆112Updated 5 months ago
- [ICLR 2025] The official pytorch implement of "Dynamic-LLaVA: Efficient Multimodal Large Language Models via Dynamic Vision-language Cont…☆51Updated 9 months ago
- ☆79Updated last year
- Awesome-Low-Rank-Adaptation☆115Updated 10 months ago
- 🔥 How to efficiently and effectively compress the CoTs or directly generate concise CoTs during inference while maintaining the reasonin…☆57Updated 3 months ago
- toy reproduction of Auxiliary-Loss-Free Load Balancing Strategy for Mixture-of-Experts☆21Updated last year
- Sharing my research toolchain☆85Updated last year
- CorDA: Context-Oriented Decomposition Adaptation of Large Language Models for task-aware parameter-efficient fine-tuning(NeurIPS 2024)☆49Updated 7 months ago
- EMPO, A Fully Unsupervised RLVR Method☆65Updated last week
- ☆41Updated last year
- Inference Code for Paper "Harder Tasks Need More Experts: Dynamic Routing in MoE Models"☆62Updated last year
- CLIP-MoE: Mixture of Experts for CLIP☆46Updated 10 months ago