lucidrains / flash-attention-jaxView external linksLinks
Implementation of Flash Attention in Jax
☆225Mar 1, 2024Updated last year
Alternatives and similar repositories for flash-attention-jax
Users that are interested in flash-attention-jax are comparing it to the libraries listed below
Sorting:
- Implementation of fused cosine similarity attention in the same style as Flash Attention☆220Feb 13, 2023Updated 3 years ago
- jax-triton contains integrations between JAX and OpenAI Triton☆439Feb 9, 2026Updated last week
- LoRA for arbitrary JAX models and functions☆145Feb 26, 2024Updated last year
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆548Updated this week
- JAX implementation of the Llama 2 model☆216Feb 2, 2024Updated 2 years ago
- JMP is a Mixed Precision library for JAX.☆211Jan 30, 2025Updated last year
- ☆367Apr 12, 2024Updated last year
- Inference code for LLaMA models in JAX☆120May 21, 2024Updated last year
- ☆35Nov 22, 2024Updated last year
- Source-to-Source Debuggable Derivatives in Pure Python☆15Jan 23, 2024Updated 2 years ago
- JAX Synergistic Memory Inspector☆184Jul 16, 2024Updated last year
- A user-friendly tool chain that enables the seamless execution of ONNX models using JAX as the backend.☆131Updated this week
- A simple, performant and scalable Jax LLM!☆2,138Updated this week
- JAX-Toolbox☆382Updated this week
- ☆345Updated this week
- Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts☆123Oct 17, 2024Updated last year
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆694Jan 26, 2026Updated 3 weeks ago
- Train very large language models in Jax.☆210Oct 21, 2023Updated 2 years ago
- Implementation of a Transformer, but completely in Triton☆279Apr 5, 2022Updated 3 years ago
- Accelerate, Optimize performance with streamlined training and serving options with JAX.☆337Updated this week
- Elegant easy-to-use neural networks + scientific computing in JAX. https://docs.kidger.site/equinox/☆2,771Updated this week
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆93Jan 25, 2024Updated 2 years ago
- Minimal library to train LLMs on TPU in JAX with pjit().☆301Dec 20, 2023Updated 2 years ago
- Minimal yet performant LLM examples in pure JAX☆242Jan 14, 2026Updated last month
- Implementation of the Kalman Filtering Attention proposed in "Kalman Filtering Attention for User Behavior Modeling in CTR Prediction"☆59Oct 22, 2023Updated 2 years ago
- A simple library for scaling up JAX programs☆145Nov 4, 2025Updated 3 months ago
- ☆922Jan 29, 2026Updated 2 weeks ago
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆40Dec 2, 2023Updated 2 years ago
- JORA: JAX Tensor-Parallel LoRA Library (ACL 2024)☆36Apr 25, 2024Updated last year
- Optax is a gradient processing and optimization library for JAX.☆2,177Feb 9, 2026Updated last week
- Orbax provides common checkpointing and persistence utilities for JAX users☆481Updated this week
- Implementation of Diffusion Transformers and Rectified Flow in Jax☆27Jul 9, 2024Updated last year
- Implementing the Denoising Diffusion Probabilistic Model in Flax☆157Nov 1, 2022Updated 3 years ago
- Task-based datasets, preprocessing, and evaluation for sequence models.☆594Feb 3, 2026Updated 2 weeks ago
- Implementation of the transformer proposed in "Building Blocks for a Complex-Valued Transformer Architecture"☆88Oct 13, 2023Updated 2 years ago
- ☆51Jan 28, 2024Updated 2 years ago
- JAX/Flax implementation of the Hyena Hierarchy☆34Apr 27, 2023Updated 2 years ago
- Reference implementation of "Softmax Attention with Constant Cost per Token" (Heinsen, 2024)☆24Jun 6, 2024Updated last year
- Everything you want to know about Google Cloud TPU☆562Jul 16, 2024Updated last year