Mamba-Chat: A chat LLM based on the state-space model architecture π
β942Mar 3, 2024Updated 2 years ago
Alternatives and similar repositories for mamba-chat
Users that are interested in mamba-chat are comparing it to the libraries listed below
Sorting:
- Mamba SSM architectureβ17,311Feb 18, 2026Updated 2 weeks ago
- Simple, minimal implementation of the Mamba SSM in one file of PyTorch.β2,920Mar 8, 2024Updated last year
- Implementation of the Mamba SSM with hf_integration.β55Aug 31, 2024Updated last year
- This is the code that went into our practical dive using mamba as information extractionβ57Dec 22, 2023Updated 2 years ago
- A simple and efficient Mamba implementation in pure PyTorch and MLX.β1,434Jan 26, 2026Updated last month
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modelingβ952Nov 16, 2025Updated 3 months ago
- Inference of Mamba and Mamba2 models in pure Cβ197Jan 22, 2026Updated last month
- Some preliminary explorations of Mamba's context scaling.β218Feb 8, 2024Updated 2 years ago
- The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.β8,896May 3, 2024Updated last year
- Code repository for Black Mambaβ263Feb 8, 2024Updated 2 years ago
- Annotated version of the Mamba paperβ497Feb 27, 2024Updated 2 years ago
- Repository for StripedHyena, a state-of-the-art beyond Transformer architectureβ413Mar 7, 2024Updated last year
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)β¦β14,393Feb 21, 2026Updated last week
- A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplestβ¦β463Feb 13, 2026Updated 3 weeks ago
- Tools for merging pretrained large language models.β6,826Updated this week
- [ICLR 2024] Efficient Streaming Language Models with Attention Sinksβ7,196Jul 11, 2024Updated last year
- YaRN: Efficient Context Window Extension of Large Language Modelsβ1,676Apr 17, 2024Updated last year
- Go ahead and axolotl questionsβ11,395Updated this week
- Awesome list of papers that extend Mamba to various applications.β139Jun 10, 2025Updated 8 months ago
- The official implementation of Self-Play Fine-Tuning (SPIN)β1,235May 8, 2024Updated last year
- Robust recipes to align language models with human and AI preferencesβ5,510Sep 8, 2025Updated 5 months ago
- Implementation of MambaByte in "MambaByte: Token-free Selective State Space Model" in Pytorch and Zetaβ125Feb 6, 2026Updated last month
- β‘ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Plβ¦β2,175Oct 8, 2024Updated last year
- [ICML 2024] LLMCompiler: An LLM Compiler for Parallel Function Callingβ1,825Jul 10, 2024Updated last year
- Training LLMs with QLoRA + FSDPβ1,538Nov 9, 2024Updated last year
- Integrating Mamba/SSMs with Transformer for Enhanced Long Context and High-Quality Sequence Modelingβ215Jan 30, 2026Updated last month
- 20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.β13,206Updated this week
- β208Jan 14, 2026Updated last month
- β62Dec 8, 2023Updated 2 years ago
- Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Headsβ2,710Jun 25, 2024Updated last year
- PyTorch native post-training libraryβ5,691Feb 27, 2026Updated last week
- Official inference library for Mistral modelsβ10,690Feb 26, 2026Updated last week
- Structured state space sequence modelsβ2,854Jul 17, 2024Updated last year
- β35Nov 22, 2024Updated last year
- LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMathβ9,477Jun 7, 2025Updated 9 months ago
- A fast inference library for running LLMs locally on modern consumer-class GPUsβ4,444Dec 9, 2025Updated 2 months ago
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Modelsβ1,663Mar 8, 2024Updated last year
- Run Mixtral-8x7B models in Colab or consumer desktopsβ2,328Apr 8, 2024Updated last year
- β718Mar 6, 2024Updated 2 years ago