Implementation of mamba with rust
☆92Mar 9, 2024Updated last year
Alternatives and similar repositories for mamba-ssm
Users that are interested in mamba-ssm are comparing it to the libraries listed below
Sorting:
- Inference of Mamba and Mamba2 models in pure C☆197Jan 22, 2026Updated last month
- win32 native frontend for llama-cli☆12Nov 2, 2024Updated last year
- Official Repository for Efficient Linear-Time Attention Transformers.☆18Jun 2, 2024Updated last year
- Implementation of the Mamba SSM with hf_integration.☆55Aug 31, 2024Updated last year
- GPT-4 Level Conversational QA Trained In a Few Hours☆68Aug 21, 2024Updated last year
- Some preliminary explorations of Mamba's context scaling.☆218Feb 8, 2024Updated 2 years ago
- A rust wrapper for the spoa C++ partial order alignment library☆10Jun 11, 2025Updated 8 months ago
- Yet another frontend for LLM, written using .NET and WinUI 3☆10Sep 14, 2025Updated 5 months ago
- Rust binding for WFA2-lib☆10Jun 7, 2022Updated 3 years ago
- FMS Model Optimizer is a framework for developing reduced precision neural network models.☆21Feb 23, 2026Updated last week
- ☆12May 30, 2025Updated 9 months ago
- an auto-sleeping and -waking framework around llama.cpp☆12Feb 8, 2025Updated last year
- Desktop application for instant AI-powered text transformation. Translate, correct, summarize, and change the tone of any text, anywhere,…☆28Dec 29, 2025Updated 2 months ago
- LLM CLI Interface - Extremely Convenient and Fast☆12Sep 22, 2025Updated 5 months ago
- Modified Mamba code to run on CPU☆30Jan 14, 2024Updated 2 years ago
- My Implementation of Q-Sparse: All Large Language Models can be Fully Sparsely-Activated☆33Aug 14, 2024Updated last year
- Yet another LLM command line interface☆16Dec 9, 2024Updated last year
- A sleek, customizable interface for managing LLMs with responsive design and easy agent personalization.☆17Aug 30, 2024Updated last year
- Mic-controlled mouse clicks☆17Oct 6, 2025Updated 4 months ago
- A LaTeX package for drawing genetic construct pictograms using TikZ.☆11Jan 26, 2022Updated 4 years ago
- ☆12Apr 4, 2024Updated last year
- A C++ implementation of tinyllama inference on CPU.☆12Feb 28, 2024Updated 2 years ago
- a lightweight, open-source blueprint for building powerful and scalable LLM chat applications☆28Jun 7, 2024Updated last year
- A high-performance FastAPI-based server that provides OpenAI-compatible Text-to-Speech (TTS) endpoints using the Orpheus TTS https://gith…☆30Nov 15, 2025Updated 3 months ago
- TaCo: Enhancing Cross-Lingual Transfer for Low-Resource Languages in LLMs through Translation-Assisted Chain-of-Thought Processes☆13Jul 1, 2025Updated 8 months ago
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆61Apr 8, 2024Updated last year
- a character-ai like UI for LLM☆10Dec 3, 2024Updated last year
- Inference Llama 2 with a model compiled to native code by TorchInductor☆14Feb 8, 2024Updated 2 years ago
- Experience the power of AI with this free AI voice generator demo. Utilizing Deepgram and Groq, we transform text into voice seamlessly. …☆37Jun 12, 2024Updated last year
- Code repository for Black Mamba☆263Feb 8, 2024Updated 2 years ago
- Load and run Llama from safetensors files in C☆15Oct 24, 2024Updated last year
- Grokking on modular arithmetic in less than 150 epochs in MLX☆14Oct 24, 2024Updated last year
- Python Wrappings for exploring Set Substitution Systems (Wolfram Models)☆15Jun 3, 2020Updated 5 years ago
- Inference Llama 2 in one file of pure Haskell (A port of llama2.c from Andrej Karpathy)☆14Oct 17, 2025Updated 4 months ago
- BUD-E (Buddy) is an open-source voice assistant framework that facilitates seamless interaction with AI models and APIs, enabling the cre…☆42Jul 18, 2024Updated last year
- Note about running ollama 🦙☆36May 2, 2024Updated last year
- These agents work based on any local model. You ask your question and simply indicate the number of agents and experts who will answer it…☆19Feb 25, 2024Updated 2 years ago
- An efficient pytorch implementation of selective scan in one file, works with both cpu and gpu, with corresponding mathematical derivatio…☆103Oct 14, 2025Updated 4 months ago
- Implementation of a modular, high-performance, and simplistic mamba for high-speed applications☆40Nov 11, 2024Updated last year