wrmedford / moe-scaling

Scaling Laws for Mixture of Experts Models
β˜†10Updated last month

Alternatives and similar repositories for moe-scaling:

Users that are interested in moe-scaling are comparing it to the libraries listed below