kyegomez / MultiModalMamba
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
☆446Updated this week
Alternatives and similar repositories for MultiModalMamba:
Users that are interested in MultiModalMamba are comparing it to the libraries listed below
- Code repository for Black Mamba☆238Updated last year
- Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"☆841Updated this week
- PyTorch Implementation of "V* : Guided Visual Search as a Core Mechanism in Multimodal LLMs"☆568Updated last year
- Build high-performance AI models with modular building blocks☆469Updated this week
- Implementation of DoRA☆290Updated 8 months ago
- ☆253Updated 5 months ago
- From scratch implementation of a vision language model in pure PyTorch☆194Updated 9 months ago
- 👁️ + 💬 + 🎧 = 🤖 Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]☆601Updated 11 months ago
- This repository provides the code and model checkpoints for AIMv1 and AIMv2 research projects.☆1,182Updated 2 months ago
- [ICLR2025 Spotlight🔥] Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆514Updated last week
- PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention…☆286Updated 9 months ago
- Mamba-Chat: A chat LLM based on the state-space model architecture 🐍☆920Updated 11 months ago
- PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"☆157Updated 3 weeks ago
- PyTorch implementation of models from the Zamba2 series.☆176Updated 3 weeks ago
- Training Large Language Model to Reason in a Continuous Latent Space☆877Updated 3 weeks ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆297Updated 2 months ago
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills☆722Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆230Updated 3 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆222Updated 9 months ago
- Embed arbitrary modalities (images, audio, documents, etc) into large language models.☆177Updated 10 months ago
- Integrating Mamba/SSMs with Transformer for Enhanced Long Context and High-Quality Sequence Modeling☆186Updated 3 weeks ago
- ☆706Updated 11 months ago
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆215Updated 3 weeks ago
- ☆502Updated 5 months ago
- Fine-tune mistral-7B on 3090s, a100s, h100s☆705Updated last year
- Reference implementation of Megalodon 7B model☆516Updated 10 months ago
- run paligemma in real time☆130Updated 9 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆167Updated last month
- When do we not need larger vision models?☆368Updated last week