Oliver-FutureAI / Awesome-MoELinks
Awesome list of Mixture-of-Experts (MoE)
☆20Updated 11 months ago
Alternatives and similar repositories for Awesome-MoE
Users that are interested in Awesome-MoE are comparing it to the libraries listed below
Sorting:
- [CVPR2024] Efficient Dataset Distillation via Minimax Diffusion☆90Updated last year
- ☆108Updated last year
- ☆10Updated 4 months ago
- A pytorch implementation of CVPR24 paper "D4M: Dataset Distillation via Disentangled Diffusion Model"☆31Updated 8 months ago
- [CVPR2024 highlight] Generalized Large-Scale Data Condensation via Various Backbone and Statistical Matching (G-VBSM)☆28Updated 7 months ago
- Diffusion-TTA improves pre-trained discriminative models such as image classifiers or segmentors using pre-trained generative models.☆73Updated last year
- source code for NeurIPS'23 paper "Dream the Impossible: Outlier Imagination with Diffusion Models"☆68Updated last month
- Official repo of M$^2$PT: Multimodal Prompt Tuning for Zero-shot Instruction Learning☆22Updated 2 months ago
- ☆16Updated 11 months ago
- Official implementation for 'Class-Balancing Diffusion Models'☆54Updated last year
- [NeurIPS'23] DropPos: Pre-Training Vision Transformers by Reconstructing Dropped Positions☆60Updated last year
- [CVPR2024 Highlight] Official implementation for Transferable Visual Prompting. The paper "Exploring the Transferability of Visual Prompt…☆43Updated 5 months ago
- Code for ICML 2024 paper (Oral) — Test-Time Model Adaptation with Only Forward Passes☆77Updated 9 months ago
- The official implementation of "2024NeurIPS Dynamic Tuning Towards Parameter and Inference Efficiency for ViT Adaptation"☆46Updated 5 months ago
- Idempotent Generative Network's unofficial pytorch implementation☆45Updated last year
- [CVPR 2025 (Oral)] Mitigating Hallucinations in Large Vision-Language Models via DPO: On-Policy Data Hold the Key☆57Updated 2 weeks ago
- (NeurIPS 2023 spotlight) Large-scale Dataset Distillation/Condensation, 50 IPC (Images Per Class) achieves the highest 60.8% on original …☆128Updated 6 months ago
- The official github repo for "Test-Time Training with Masked Autoencoders"☆83Updated last year
- Code for our ICML'24 on multimodal dataset distillation☆37Updated 7 months ago
- Data distillation benchmark☆64Updated this week
- [CVPR 2024] On the Diversity and Realism of Distilled Dataset: An Efficient Dataset Distillation Paradigm☆71Updated 3 months ago
- ☆56Updated 5 months ago
- [ICLR 2024] ViDA: Homeostatic Visual Domain Adapter for Continual Test Time Adaptation☆65Updated last year
- ☆20Updated 6 months ago
- ☆87Updated 2 years ago
- Official repository of "Back to Source: Diffusion-Driven Test-Time Adaptation"☆78Updated last year
- [NeurIPS 2022] Revisiting Realistic Test-Time Training: Sequential Inference and Adaptation by Anchored Clustering☆48Updated last year
- Everything to the Synthetic: Diffusion-driven Test-time Adaptation via Synthetic-Domain Alignment, arXiv 2024 / CVPR 2025☆28Updated 3 months ago
- Official implementation for paper "Knowledge Diffusion for Distillation", NeurIPS 2023☆86Updated last year
- ☆133Updated 11 months ago