AI-Hypercomputer / pathways-utilsLinks
Package of Pathways-on-Cloud utilities
☆23Updated last week
Alternatives and similar repositories for pathways-utils
Users that are interested in pathways-utils are comparing it to the libraries listed below
Sorting:
- ☆72Updated this week
- ☆16Updated 3 months ago
- A collection of lightweight interpretability scripts to understand how LLMs think☆89Updated last week
- ☆21Updated 11 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆68Updated this week
- 👷 Build compute kernels☆214Updated last week
- Train, tune, and infer Bamba model☆138Updated 8 months ago
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆141Updated 4 months ago
- AI-Driven Research Systems (ADRS)☆117Updated last month
- PyTorch centric eager mode debugger☆48Updated last year
- LM engine is a library for pretraining/finetuning LLMs☆113Updated this week
- Simple repository for training small reasoning models☆48Updated 11 months ago
- ☆60Updated this week
- OLMost every training recipe you need to perform data interventions with the OLMo family of models.☆64Updated this week
- ☆73Updated this week
- xpk (Accelerated Processing Kit, pronounced x-p-k,) is a software tool to help Cloud developers to orchestrate training jobs on accelerat…☆162Updated this week
- Train a SmolLM-style llm on fineweb-edu in JAX/Flax with an assortment of optimizers.☆18Updated 6 months ago
- Simple and efficient DeepSeek V3 SFT using pipeline parallel and expert parallel, with both FP8 and BF16 trainings☆114Updated 6 months ago
- Make triton easier☆50Updated last year
- Official repo of dataset-decomposition paper [NeurIPS 2024]☆20Updated last year
- a Jax quantization library☆87Updated last week
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆197Updated 8 months ago
- ML/DL Math and Method notes☆66Updated 2 years ago
- A lightweight, user-friendly data-plane for LLM training.☆38Updated 4 months ago
- MoE training for Me and You and maybe other people☆335Updated last month
- Memory optimized Mixture of Experts☆72Updated 6 months ago
- ☆237Updated last month
- A collection of reproducible inference engine benchmarks☆38Updated 9 months ago
- ☆16Updated last year
- ☆72Updated last month