AI-Hypercomputer / ray-tpuLinks
☆15Updated 4 months ago
Alternatives and similar repositories for ray-tpu
Users that are interested in ray-tpu are comparing it to the libraries listed below
Sorting:
- torchprime is a reference model implementation for PyTorch on TPU.☆36Updated this week
- Fast, Modern, and Low Precision PyTorch Optimizers☆109Updated last week
- ☆20Updated 2 years ago
- ☆21Updated 6 months ago
- ☆118Updated last year
- DPO, but faster 🚀☆44Updated 9 months ago
- Experimenting with how best to do multi-host dataloading☆10Updated 2 years ago
- ☆14Updated 3 months ago
- some common Huggingface transformers in maximal update parametrization (µP)☆82Updated 3 years ago
- Experimental playground for benchmarking language model (LM) architectures, layers, and tricks on smaller datasets. Designed for flexible…☆78Updated 2 weeks ago
- Various transformers for FSDP research☆38Updated 2 years ago
- Tiled Flash Linear Attention library for fast and efficient mLSTM Kernels.☆68Updated last month
- A set of Python scripts that makes your experience on TPU better☆54Updated last year
- Implementation of a Light Recurrent Unit in Pytorch☆48Updated 11 months ago
- ☆15Updated last year
- Triton Implementation of HyperAttention Algorithm☆48Updated last year
- An implementation of the Llama architecture, to instruct and delight☆21Updated 3 months ago
- A toolkit for scaling law research ⚖☆51Updated 7 months ago
- Griffin MQA + Hawk Linear RNN Hybrid☆89Updated last year
- JAX/Flax implementation of the Hyena Hierarchy☆34Updated 2 years ago
- ☆21Updated 10 months ago
- Train a SmolLM-style llm on fineweb-edu in JAX/Flax with an assortment of optimizers.☆19Updated last month
- A JAX-native LLM Post-Training Library☆143Updated this week
- Blazing fast data loading with HuggingFace Dataset and Ray Data☆16Updated last year
- Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount…☆53Updated last year
- Randomized Positional Encodings Boost Length Generalization of Transformers☆82Updated last year
- Implementation of GateLoop Transformer in Pytorch and Jax☆90Updated last year
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆107Updated 6 months ago
- A byte-level decoder architecture that matches the performance of tokenized Transformers.☆65Updated last year
- Machine Learning eXperiment Utilities☆47Updated last month