junfanz1 / MoE-Mixture-of-Experts-in-PyTorchLinks

Implementations of a Mixture-of-Experts (MoE) architecture designed for research on large language models (LLMs) and scalable neural network designs. One implementation targets a **single-device/NPU environment** while the other is built for multi-device distributed computing. Both versions showcase the core principles.
36Updated 8 months ago

Alternatives and similar repositories for MoE-Mixture-of-Experts-in-PyTorch

Users that are interested in MoE-Mixture-of-Experts-in-PyTorch are comparing it to the libraries listed below

Sorting: