XiaoMi / MobileBenchLinks
Mobile-Bench: An Evaluation Benchmark for LLM-based Mobile Agents
☆21Updated last year
Alternatives and similar repositories for MobileBench
Users that are interested in MobileBench are comparing it to the libraries listed below
Sorting:
- KV cache compression via sparse coding☆14Updated last month
- Official code for the paper "HEXA-MoE: Efficient and Heterogeneous-Aware MoE Acceleration with Zero Computation Redundancy"☆14Updated 9 months ago
- MobileVLM: A Vision-Language Model for Better Intra- and Inter-UI Understanding☆75Updated 9 months ago
- ☆85Updated 8 months ago
- D^2-MoE: Delta Decompression for MoE-based LLMs Compression☆71Updated 8 months ago
- Official implementation of ICML 2024 paper "ExCP: Extreme LLM Checkpoint Compression via Weight-Momentum Joint Shrinking".☆48Updated last year
- [ICCV 2025] StreamMind: Unlocking Full Frame Rate Streaming Video Dialogue through Event-Gated Cognition☆51Updated 5 months ago
- Implementation for the paper: CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM Inference☆30Updated 9 months ago
- ZO2 (Zeroth-Order Offloading): Full Parameter Fine-Tuning 175B LLMs with 18GB GPU Memory [COLM2025]☆196Updated 4 months ago
- ☆74Updated 5 months ago
- Unveiling Super Experts in Mixture-of-Experts Large Language Models☆32Updated 2 months ago
- Due to the huge vocaburary size (151,936) of Qwen models, the Embedding and LM Head weights are excessively heavy. Therefore, this projec…☆29Updated last year
- Official Implementation of APB (ACL 2025 main Oral)☆31Updated 9 months ago
- APRIL: Active Partial Rollouts in Reinforcement Learning to Tame Long-tail Generation. A system-level optimization for scalable LLM tra…☆44Updated last month
- ☆19Updated 11 months ago
- [TMLR] LLM-Powered GUI Agents in Phone Automation: Surveying Progress and Prospects☆129Updated 7 months ago
- [ICML'24] The official implementation of “Rethinking Optimization and Architecture for Tiny Language Models”☆125Updated 10 months ago
- ☆34Updated 4 months ago
- (ACL 2025 oral) SCOPE: Optimizing KV Cache Compression in Long-context Generation☆33Updated 6 months ago
- ☆11Updated 11 months ago
- ☆23Updated 10 months ago
- ZeroGUI: Automating Online GUI Learning at Zero Human Cost☆102Updated 4 months ago
- [ICML 2025 Oral] Mixture of Lookup Experts☆55Updated 6 months ago
- Efficient Mixture of Experts for LLM Paper List☆145Updated 2 months ago
- ☆21Updated 2 weeks ago
- ☆104Updated 2 months ago
- [NAACL'25 🏆 SAC Award] Official code for "Advancing MoE Efficiency: A Collaboration-Constrained Routing (C2R) Strategy for Better Expert…☆13Updated 10 months ago
- PLM: Efficient Peripheral Language Models Hardware-Co-Designed for Ubiquitous Computing☆20Updated 8 months ago
- [ICML24] Pruner-Zero: Evolving Symbolic Pruning Metric from scratch for LLMs☆96Updated last year
- Model compression toolkit engineered for enhanced usability, comprehensiveness, and efficiency.☆212Updated last week