kyegomez / zeta
Build high-performance AI models with modular building blocks
☆476Updated this week
Alternatives and similar repositories for zeta:
Users that are interested in zeta are comparing it to the libraries listed below
- PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"☆158Updated last month
- Integrating Mamba/SSMs with Transformer for Enhanced Long Context and High-Quality Sequence Modeling☆187Updated last month
- A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest…☆447Updated this week
- A simple and efficient Mamba implementation in pure PyTorch and MLX.☆1,140Updated 3 months ago
- Collection of papers on state-space models☆577Updated this week
- Repo for "Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture"☆548Updated 2 months ago
- Annotated version of the Mamba paper☆475Updated last year
- Code repository for Black Mamba☆239Updated last year
- PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention…☆286Updated 10 months ago
- Implementation of MoE Mamba from the paper: "MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts" in Pytorch and Ze…☆97Updated last month
- Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch☆314Updated 8 months ago
- Official JAX implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States☆396Updated 6 months ago
- Awesome list of papers that extend Mamba to various applications.☆131Updated 2 months ago
- Reading list for research topics in state-space models☆263Updated last month
- Causal depthwise conv1d in CUDA, with a PyTorch interface☆396Updated 3 months ago
- Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"☆850Updated 2 weeks ago
- A easy, reliable, fluid template for python packages complete with docs, testing suites, readme's, github workflows, linting and much muc…☆162Updated last month
- The official implementation of Tensor ProducT ATTenTion Transformer (T6)☆316Updated 2 weeks ago
- [ICLR2025 Spotlight🔥] Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆530Updated 3 weeks ago
- ☆171Updated 2 months ago
- Implementation of Vision Mamba from the paper: "Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Mod…☆430Updated 3 weeks ago
- Helpful tools and examples for working with flex-attention☆667Updated 2 weeks ago
- Muon optimizer: +>30% sample efficiency with <3% wallclock overhead☆439Updated this week
- Implementation of MambaByte in "MambaByte: Token-free Selective State Space Model" in Pytorch and Zeta☆114Updated last month
- Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793☆388Updated 3 months ago
- xLSTM as Generic Vision Backbone☆464Updated 4 months ago
- Notes on the Mamba and the S4 model (Mamba: Linear-Time Sequence Modeling with Selective State Spaces)☆161Updated last year
- From scratch implementation of a vision language model in pure PyTorch☆197Updated 9 months ago
- Huggingface compatible implementation of RetNet (Retentive Networks, https://arxiv.org/pdf/2307.08621.pdf) including parallel, recurrent,…☆225Updated 11 months ago
- 🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton☆2,040Updated this week