google / flaxformerLinks
☆364Updated last year
Alternatives and similar repositories for flaxformer
Users that are interested in flaxformer are comparing it to the libraries listed below
Sorting:
- ☆190Updated 2 weeks ago
- Train very large language models in Jax.☆210Updated 2 years ago
- Implementation of Flash Attention in Jax☆223Updated last year
- Task-based datasets, preprocessing, and evaluation for sequence models.☆590Updated 3 weeks ago
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆189Updated 3 years ago
- JAX Synergistic Memory Inspector☆183Updated last year
- ☆259Updated 6 months ago
- ☆66Updated 3 years ago
- Inference code for LLaMA models in JAX☆120Updated last year
- JAX implementation of the Llama 2 model☆216Updated last year
- Sequence modeling with Mega.☆301Updated 2 years ago
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆540Updated 3 weeks ago
- Amos optimizer with JEstimator lib.☆82Updated last year
- Implementation of a Transformer, but completely in Triton☆277Updated 3 years ago
- ☆166Updated 2 years ago
- Language Modeling with the H3 State Space Model☆519Updated 2 years ago
- Swarm training framework using Haiku + JAX + Ray for layer parallel transformer language models on unreliable, heterogeneous nodes☆242Updated 2 years ago
- jax-triton contains integrations between JAX and OpenAI Triton☆436Updated this week
- JMP is a Mixed Precision library for JAX.☆211Updated 10 months ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆685Updated last week
- Named tensors with first-class dimensions for PyTorch☆332Updated 2 years ago
- LoRA for arbitrary JAX models and functions☆143Updated last year
- Implementation of Mega, the Single-head Attention with Multi-headed EMA architecture that currently holds SOTA on Long Range Arena☆207Updated 2 years ago
- ☆62Updated 3 years ago
- Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT☆224Updated last year
- Implementation of https://srush.github.io/annotated-s4☆506Updated 5 months ago
- An interpreter for RASP as described in the ICML 2021 paper "Thinking Like Transformers"☆323Updated last year
- ☆250Updated 5 years ago
- Implementation of the conditionally routed attention in the CoLT5 architecture, in Pytorch☆230Updated last year
- A minimal PyTorch Lightning OpenAI GPT w DeepSpeed Training!☆113Updated 2 years ago