From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)
☆790Oct 30, 2024Updated last year
Alternatives and similar repositories for makeMoE
Users that are interested in makeMoE are comparing it to the libraries listed below
Sorting:
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Models☆1,663Mar 8, 2024Updated last year
- 【TMM 2025🔥】 Mixture-of-Experts for Large Vision-Language Models☆2,303Jul 15, 2025Updated 7 months ago
- ☆251Mar 20, 2024Updated last year
- Tools for merging pretrained large language models.☆6,826Updated this week
- An autoregressive character-level language model for making more things