AI-Hypercomputer / kitharaLinks
☆14Updated 4 months ago
Alternatives and similar repositories for kithara
Users that are interested in kithara are comparing it to the libraries listed below
Sorting:
- torchprime is a reference model implementation for PyTorch on TPU.☆39Updated last week
- ☆146Updated 2 weeks ago
- ☆15Updated 4 months ago
- ☆23Updated last week
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆73Updated 3 weeks ago
- xpk (Accelerated Processing Kit, pronounced x-p-k,) is a software tool to help Cloud developers to orchestrate training jobs on accelerat…☆143Updated last week
- ☆46Updated this week
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆379Updated 3 months ago
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆538Updated last month
- ☆189Updated last week
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆392Updated this week
- Recipes for reproducing training and serving benchmarks for large machine learning models using GPUs on Google Cloud.☆88Updated last week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆414Updated last week
- ☆16Updated 6 months ago
- ☆538Updated last year
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆268Updated 2 months ago
- Two implementations of ZeRO-1 optimizer sharding in JAX☆14Updated 2 years ago
- ☆331Updated 3 weeks ago
- ☆23Updated 3 weeks ago
- Testing framework for Deep Learning models (Tensorflow and PyTorch) on Google Cloud hardware accelerators (TPU and GPU)☆65Updated 3 months ago
- Google TPU optimizations for transformers models☆120Updated 8 months ago
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆161Updated 2 weeks ago
- Scalable and Performant Data Loading☆304Updated 2 weeks ago
- Experimenting with how best to do multi-host dataloading☆10Updated 2 years ago
- PyTorch Single Controller☆425Updated this week
- ☆122Updated last year
- jax-triton contains integrations between JAX and OpenAI Triton☆426Updated last month
- Minimal yet performant LLM examples in pure JAX☆181Updated 2 weeks ago
- Tokamax: A GPU and TPU kernel library.☆87Updated this week
- Load compute kernels from the Hub☆293Updated this week