john-cuffe / MAMBA2
Second Generation of the MAMBA Software
☆27Updated last year
Related projects: ⓘ
- A repository for DenseSSMs☆86Updated 5 months ago
- ☆24Updated last month
- [CVPR'24] Multimodal Pathway: Improve Transformers with Irrelevant Data from Other Modalities☆85Updated 6 months ago
- ☆54Updated 2 months ago
- Public code repo for paper "A Single Transformer for Scalable Vision-Language Modeling"☆103Updated last month
- Implementation of ViTaR: ViTAR: Vision Transformer with Any Resolution in PyTorch☆22Updated last week
- [NeurIPS2023] Make Your Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning☆28Updated last year
- [CVPR 2024] The official pytorch implementation of "A General and Efficient Training for Transformer via Token Expansion".☆36Updated 4 months ago
- Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆56Updated this week
- Implementation of "VL-Mamba: Exploring State Space Models for Multimodal Learning"☆75Updated 6 months ago
- ☆46Updated 10 months ago
- Official PyTorch Implementation of "The Hidden Attention of Mamba Models"☆186Updated 3 months ago
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning☆25Updated last month
- PyTorch implementation of "From Sparse to Soft Mixtures of Experts"☆38Updated last year
- The source code of the EMNLP 2023 main conference paper: Sparse Low-rank Adaptation of Pre-trained Language Models.☆62Updated 6 months ago
- Awsome works based on SSM and Mamba☆14Updated 5 months ago
- [ICML 2024] This repository includes the official implementation of our paper "Rejuvenating image-GPT as Strong Visual Representation Lea…☆96Updated 4 months ago
- [ICML 2024] Memory-Space Visual Prompting for Efficient Vision-Language Fine-Tuning☆41Updated 4 months ago
- Visual self-questioning for large vision-language assistant.☆22Updated 3 weeks ago
- My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing o…☆36Updated 9 months ago
- [ICCV2023] Official code for "VL-PET: Vision-and-Language Parameter-Efficient Tuning via Granularity Control"☆51Updated last year
- ☆100Updated last month
- Repo for the paper `ControlMLLM: Training-Free Visual Prompt Learning for Multimodal Large Language Models'☆44Updated 3 weeks ago
- ☆40Updated 4 months ago
- ✨✨Beyond LLaVA-HD: Diving into High-Resolution Large Multimodal Models☆128Updated last month
- Visual CoT: Advancing Multi-Modal Language Models with a Comprehensive Dataset and Benchmark for Chain-of-Thought Reasoning☆93Updated 2 months ago
- MoVA: Adapting Mixture of Vision Experts to Multimodal Context☆116Updated 2 weeks ago
- Introduce Mamba2 to Vision.☆70Updated 3 weeks ago
- Official implementation of the Law of Vision Representation in MLLMs☆93Updated last week
- Code for NOLA, an implementation of "nola: Compressing LoRA using Linear Combination of Random Basis"☆46Updated 3 weeks ago