AI-Hypercomputer / kitharaLinks
☆16Updated 5 months ago
Alternatives and similar repositories for kithara
Users that are interested in kithara are comparing it to the libraries listed below
Sorting:
- ☆15Updated 6 months ago
- torchprime is a reference model implementation for PyTorch on TPU.☆40Updated last month
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆77Updated 2 months ago
- ☆53Updated last week
- ☆24Updated this week
- ☆121Updated last year
- Official repo to On the Generalization Ability of Retrieval-Enhanced Transformers☆44Updated last year
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆110Updated 2 weeks ago
- some common Huggingface transformers in maximal update parametrization (µP)☆86Updated 3 years ago
- Various transformers for FSDP research☆38Updated 3 years ago
- ☆20Updated 2 years ago
- A library for unit scaling in PyTorch☆132Updated 4 months ago
- TPU inference for vLLM, with unified JAX and PyTorch support.☆161Updated this week
- Google TPU optimizations for transformers models☆122Updated 9 months ago
- A set of Python scripts that makes your experience on TPU better☆54Updated 2 months ago
- ☆145Updated last week
- Load compute kernels from the Hub☆326Updated last week
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆271Updated last week
- ☆190Updated 3 weeks ago
- A toolkit for scaling law research ⚖☆53Updated 9 months ago
- ☆66Updated 3 years ago
- Muon fsdp 2☆44Updated 3 months ago
- 🚀 Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.☆216Updated this week
- Fast, Modern, and Low Precision PyTorch Optimizers☆116Updated 2 months ago
- Pragmatic approach to parsing import profiles for CI's☆12Updated last year
- Official code release for "SuperBPE: Space Travel for Language Models"☆75Updated 2 weeks ago
- ☆149Updated 2 years ago
- ☆47Updated last year
- Efficient encoder-decoder architecture for small language models (≤1B parameters) with cross-architecture knowledge distillation and visi…☆32Updated 9 months ago
- A flexible and efficient implementation of Flash Attention 2.0 for JAX, supporting multiple backends (GPU/TPU/CPU) and platforms (Triton/…☆30Updated 8 months ago