antonio-f / mixture-of-experts-from-scratchLinks
Mixture of Experts from scratch
☆12Updated last year
Alternatives and similar repositories for mixture-of-experts-from-scratch
Users that are interested in mixture-of-experts-from-scratch are comparing it to the libraries listed below
Sorting:
- Distributed training (multi-node) of a Transformer model☆91Updated last year
- GPU Kernels☆218Updated 8 months ago
- First-principle implementations of groundbreaking AI algorithms using a wide range of deep learning frameworks, accompanied by supporting…☆181Updated 5 months ago
- Notes and commented code for RLHF (PPO)☆121Updated last year
- ☆45Updated 7 months ago
- This repository contains an exhaustive coverage of a hands on approach to PyTorch along side powerful tools to accelerate model tuning an…☆219Updated last month
- ☆233Updated last year
- LLaMA 2 implemented from scratch in PyTorch☆365Updated 2 years ago
- Tutorial for how to build BERT from scratch☆101Updated last year
- ☆45Updated 8 months ago
- 100 days of building GPU kernels!☆561Updated 8 months ago
- An extension of the nanoGPT repository for training small MOE models.☆225Updated 10 months ago
- LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorch☆119Updated 2 years ago
- minimal GRPO implementation from scratch☆102Updated 10 months ago
- Advanced NLP, Spring 2025 https://cmu-l3.github.io/anlp-spring2025/☆70Updated 9 months ago
- ☆89Updated 9 months ago
- LoRA and DoRA from Scratch Implementations☆215Updated last year
- Notes on Direct Preference Optimization☆23Updated last year
- Building LLaMA 4 MoE from Scratch☆72Updated 9 months ago
- Research projects built on top of Transformers☆110Updated 10 months ago
- Implementation of BERT-based Language Models☆25Updated last year
- Recreating PyTorch from scratch (C/C++, CUDA, NCCL and Python, with multi-GPU support and automatic differentiation!)☆161Updated last month
- Implementations of a Mixture-of-Experts (MoE) architecture designed for research on large language models (LLMs) and scalable neural netw…☆41Updated 9 months ago
- Survey: A collection of AWESOME papers and resources on the latest research in Mixture of Experts.☆140Updated last year
- 🧠 A study guide to learn about Transformers☆12Updated 2 years ago
- Complete implementation of Llama2 with/without KV cache & inference 🚀☆49Updated last year
- A curated list of resources for learning and exploring Triton, OpenAI's programming language for writing efficient GPU code.☆453Updated 10 months ago
- making the official triton tutorials actually comprehensible☆93Updated 4 months ago
- ☆224Updated last month
- RL significantly the reasoning capability of Qwen2.5-1.5B-Instruct☆31Updated 10 months ago