Oliver-FutureAI / Awesome-MoE
Awesome list of Mixture-of-Experts (MoE)
☆16Updated 5 months ago
Related projects ⓘ
Alternatives and complementary repositories for Awesome-MoE
- Official implementation for 'Class-Balancing Diffusion Models'☆46Updated 5 months ago
- This is a collection of awesome papers I have read (carefully or roughly) in the fields of computer vision, machine learning, pattern rec…☆13Updated 3 months ago
- Official repository of "Back to Source: Diffusion-Driven Test-Time Adaptation"☆70Updated 11 months ago
- The official implementation of "2024NeurIPS Dynamic Tuning Towards Parameter and Inference Efficiency for ViT Adaptation"☆36Updated 3 weeks ago
- [CVPR2024] Efficient Dataset Distillation via Minimax Diffusion☆78Updated 7 months ago
- Code for ICML 2022 paper — Efficient Test-Time Model Adaptation without Forgetting☆106Updated last year
- The official github repo for "Test-Time Training with Masked Autoencoders"☆75Updated last year
- ☆41Updated last year
- [ICLR 2024] ViDA: Homeostatic Visual Domain Adapter for Continual Test Time Adaptation☆52Updated 6 months ago
- Continual Forgetting for Pre-trained Vision Models (CVPR 2024)☆35Updated last month
- A pytorch implementation of CVPR24 paper "D4M: Dataset Distillation via Disentangled Diffusion Model"☆23Updated 2 months ago
- ☆100Updated 7 months ago
- ICLR 2024, Towards Lossless Dataset Distillation via Difficulty-Aligned Trajectory Matching☆94Updated 5 months ago
- Code for ICML 2024 paper (Oral) — Test-Time Model Adaptation with Only Forward Passes☆46Updated 2 months ago
- A Unified Continual Learning Framework with General Parameter-Efficient Tuning, ICCV 2023 [PyTorch Code]☆69Updated last month
- Uni-OVSeg is a weakly supervised open-vocabulary segmentation framework that leverages unpaired mask-text pairs.☆50Updated 5 months ago
- ☆40Updated last year
- (NeurIPS 2023 spotlight) Large-scale Dataset Distillation/Condensation, 50 IPC (Images Per Class) achieves the highest 60.8% on original …☆119Updated this week
- This codebase is the official implementation of Test-Time Classifier Adjustment Module for Model-Agnostic Domain Generalization (NeurIPS2…☆93Updated 2 years ago
- AAAI 2024, M3D: Dataset Condensation by Minimizing Maximum Mean Discrepancy☆22Updated 8 months ago
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆68Updated this week
- [NeurIPS21] TTT++: When Does Self-supervised Test-time Training Fail or Thrive?☆58Updated 2 years ago
- ☆26Updated last year
- [NeurIPS 2023] Generalized Logit Adjustment☆34Updated 6 months ago
- Pytorch implementation of TDPM☆31Updated last year
- [ICCV 2023] DataDAM: Efficient Dataset Distillation with Attention Matching☆28Updated 4 months ago
- [NeurIPS 2022] Revisiting Realistic Test-Time Training: Sequential Inference and Adaptation by Anchored Clustering☆45Updated last year
- Consistent Prompting for Rehearsal-Free Continual Learning [CVPR2024]☆29Updated 4 months ago
- [CVPR 2024] On the Diversity and Realism of Distilled Dataset: An Efficient Dataset Distillation Paradigm☆52Updated 6 months ago
- Official PyTorch Code for "Is Synthetic Data From Diffusion Models Ready for Knowledge Distillation?" (https://arxiv.org/abs/2305.12954)☆45Updated 11 months ago