InuyashaYang / AIDIYLinks
JoinAI是一个开源仓库,专注于算法工程能力的培养,包括工程和数学原理的整理
☆11Updated 7 months ago
Alternatives and similar repositories for AIDIY
Users that are interested in AIDIY are comparing it to the libraries listed below
Sorting:
- 本仓库是关于大模型面试中常见面试试题和面试经验的整理。这里收集了各类与大模型相关的面试题目,并提供详细的解答和分析。本仓库由上海交大交影社区维护☆109Updated last year
- 🔥 How to efficiently and effectively compress the CoTs or directly generate concise CoTs during inference while maintaining the reasonin…☆63Updated 6 months ago
- Sharing my research toolchain☆85Updated last year
- ☆230Updated last year
- 🌐 Permanent Hosting Site: http://ai-paper-finder.info/ 🌐 Hugging Face Hosting: https://huggingface.co/spaces/wenhanacademia/ai-paper-f…☆214Updated this week
- ☆13Updated 7 months ago
- 😎 A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, Agent, and Beyond☆317Updated last month
- [TMLR 2025] Efficient Reasoning Models: A Survey☆280Updated last month
- [TKDE'25] The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".☆457Updated 4 months ago
- "what, how, where, and how well? a survey on test-time scaling in large language models" repository☆77Updated last week
- ☆10Updated last year
- One-shot Entropy Minimization☆187Updated 5 months ago
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆161Updated 5 months ago
- ☆19Updated last year
- ☆60Updated 4 months ago
- Code for ACL 2025 Main paper "Data Whisperer: Efficient Data Selection for Task-Specific LLM Fine-Tuning via Few-Shot In-Context Learning…☆43Updated 3 months ago
- ☆26Updated last year
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆145Updated 4 months ago
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆82Updated 5 months ago
- A curated collection of resources focused on the Mechanistic Interpretability (MI) of Large Multimodal Models (LMMs). This repository agg…☆164Updated last month
- This repository provides a comprehensive library for parallel training and LoRA algorithm implementations, supporting multiple parallel s…☆52Updated this week
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆197Updated last year
- Course materials for MIT6.5940: TinyML and Efficient Deep Learning Computing☆60Updated 10 months ago
- 必要的计算机科学及软件开发知识☆37Updated this week
- A curated list of awesome papers on dataset reduction, including dataset distillation (dataset condensation) and dataset pruning (coreset…☆58Updated 10 months ago
- ☆56Updated 5 months ago
- PrefixKV: Adaptive Prefix KV Cache is What Vision Instruction-Following Models Need for Efficient Generation [NeurIPS 2025]☆15Updated last month
- toy reproduction of Auxiliary-Loss-Free Load Balancing Strategy for Mixture-of-Experts☆25Updated last year
- The official implementation of Natural Language Fine-Tuning☆52Updated 10 months ago
- ☆184Updated 6 months ago