AI-Hypercomputer / pathways-utilsLinks
Package of Pathways-on-Cloud utilities
☆23Updated this week
Alternatives and similar repositories for pathways-utils
Users that are interested in pathways-utils are comparing it to the libraries listed below
Sorting:
- ☆72Updated last week
- ☆16Updated 3 months ago
- AI-Driven Research Systems (ADRS)☆117Updated last month
- Write a fast kernel and run it on Discord. See how you compare against the best!☆68Updated this week
- Make triton easier☆50Updated last year
- PyTorch centric eager mode debugger☆48Updated last year
- xpk (Accelerated Processing Kit, pronounced x-p-k,) is a software tool to help Cloud developers to orchestrate training jobs on accelerat…☆162Updated this week
- Cray-LM unified training and inference stack.☆22Updated last year
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆141Updated 4 months ago
- A collection of lightweight interpretability scripts to understand how LLMs think☆89Updated 2 weeks ago
- Ship correct and fast LLM kernels to PyTorch☆140Updated 3 weeks ago
- ☆16Updated last year
- Intel Gaudi's Megatron DeepSpeed Large Language Models for training☆18Updated last year
- Recipes for reproducing training and serving benchmarks for large machine learning models using GPUs on Google Cloud.☆112Updated last week
- Some microbenchmarks and design docs before commencement☆12Updated 5 years ago
- LM engine is a library for pretraining/finetuning LLMs☆113Updated last week
- Official repo of dataset-decomposition paper [NeurIPS 2024]☆20Updated last year
- ☆21Updated 11 months ago
- PathwaysJob API is an OSS Kubernetes-native API, to deploy ML training and batch inference workloads, using Pathways on GKE.☆17Updated 3 months ago
- A collection of reproducible inference engine benchmarks☆38Updated 9 months ago
- 🚀 Collection of libraries used with fms-hf-tuning to accelerate fine-tuning and training of large models.☆13Updated last week
- Experiment of using Tangent to autodiff triton☆82Updated 2 years ago
- A lightweight, user-friendly data-plane for LLM training.☆38Updated 4 months ago
- Simple repository for training small reasoning models☆48Updated last year
- Two implementations of ZeRO-1 optimizer sharding in JAX☆14Updated 2 years ago
- Parallel framework for training and fine-tuning deep neural networks☆70Updated 2 months ago
- Docker image NVIDIA GH200 machines - optimized for vllm serving and hf trainer finetuning☆53Updated 11 months ago
- ☆47Updated 2 years ago
- ML/DL Math and Method notes☆66Updated 2 years ago
- JAX implementation of the Mistral 7b v0.2 model☆35Updated last year