Yangyi-Chen / SOLOView external linksLinks
[TMLR] Public code repo for paper "A Single Transformer for Scalable Vision-Language Modeling"
☆147Nov 14, 2024Updated last year
Alternatives and similar repositories for SOLO
Users that are interested in SOLO are comparing it to the libraries listed below
Sorting:
- EVE Series: Encoder-Free Vision-Language Models from BAAI☆367Jul 24, 2025Updated 6 months ago
- DenseFusion-1M: Merging Vision Experts for Comprehensive Multimodal Perception☆159Dec 6, 2024Updated last year
- Public code repo for EMNLP 2024 Findings paper "MACAROON: Training Vision-Language Models To Be Your Engaged Partners"☆14Sep 28, 2024Updated last year
- [COLM'25] Official implementation of the Law of Vision Representation in MLLMs☆176Oct 6, 2025Updated 4 months ago
- Generalizing from SIMPLE to HARD Visual Reasoning: Can We Mitigate Modality Imbalance in VLMs?☆15Jun 3, 2025Updated 8 months ago
- Adapting LLaMA Decoder to Vision Transformer☆30May 20, 2024Updated last year
- [TACL] Do Vision and Language Models Share Concepts? A Vector Space Alignment Study☆16Nov 22, 2024Updated last year
- [NeurIPS 2024] Stabilize the Latent Space for Image Autoregressive Modeling: A Unified Perspective☆79Oct 31, 2024Updated last year
- LaVIT: Empower the Large Language Model to Understand and Generate Visual Content☆602Oct 6, 2024Updated last year
- Code for "CAFe: Unifying Representation and Generation with Contrastive-Autoregressive Finetuning"☆32Mar 26, 2025Updated 10 months ago
- [ICLR 2025] VILA-U: a Unified Foundation Model Integrating Visual Understanding and Generation☆418Apr 25, 2025Updated 9 months ago
- VL-GPT: A Generative Pre-trained Transformer for Vision and Language Understanding and Generation☆86Sep 12, 2024Updated last year
- [CVPR 2025] 🔥 Official impl. of "TokenFlow: Unified Image Tokenizer for Multimodal Understanding and Generation".☆433Aug 8, 2025Updated 6 months ago
- Implementation for "The Scalability of Simplicity: Empirical Analysis of Vision-Language Learning with a Single Transformer"☆78Oct 29, 2025Updated 3 months ago
- ☆54Jan 17, 2025Updated last year
- ☆46Nov 8, 2024Updated last year
- Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation☆823Jun 16, 2025Updated 7 months ago
- High-performance Image Tokenizers for VAR and AR☆302Apr 25, 2025Updated 9 months ago
- [ICML 2025] This is the official repository of our paper "What If We Recaption Billions of Web Images with LLaMA-3 ?"☆149Jun 13, 2024Updated last year
- This repo contains evaluation code for the paper "BLINK: Multimodal Large Language Models Can See but Not Perceive". https://arxiv.or…☆159Sep 27, 2025Updated 4 months ago
- This repository provides the code and model checkpoints for AIMv1 and AIMv2 research projects.☆1,396Aug 4, 2025Updated 6 months ago
- A Novel Semantic Segmentation Network using Enhanced Boundaries in Cluttered Scenes (WACV 2025)☆11Aug 11, 2025Updated 6 months ago
- [CVPR2025] Code Release of F-LMM: Grounding Frozen Large Multimodal Models☆108May 29, 2025Updated 8 months ago
- [MICCAI 2025] Hierarchical Self-Supervised Adversarial Training for Robust Vision Models in Histopathology☆12Jun 17, 2025Updated 7 months ago
- ☆15Jan 9, 2026Updated last month
- The code for "TokenPacker: Efficient Visual Projector for Multimodal LLM", IJCV2025☆276May 26, 2025Updated 8 months ago
- ☆19Jan 10, 2025Updated last year
- (ACL-IJCNLP 2021) Convolutions and Self-Attention: Re-interpreting Relative Positions in Pre-trained Language Models.☆21Jul 13, 2022Updated 3 years ago
- A Video Tokenizer Evaluation Dataset☆151Jan 13, 2025Updated last year
- When do we not need larger vision models?☆412Feb 8, 2025Updated last year
- This repo contains the code for 1D tokenizer and generator☆1,109Mar 20, 2025Updated 10 months ago
- Cambrian-1 is a family of multimodal LLMs with a vision-centric design.☆1,985Nov 7, 2025Updated 3 months ago
- 🐟 Code and models for the NeurIPS 2023 paper "Generating Images with Multimodal Language Models".☆471Jan 19, 2024Updated 2 years ago
- a family of highly capabale yet efficient large multimodal models☆191Aug 23, 2024Updated last year
- Code for "AVG-LLaVA: A Multimodal Large Model with Adaptive Visual Granularity"☆33Oct 12, 2024Updated last year
- Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation☆1,928Aug 15, 2024Updated last year
- SEED-Voken: A Series of Powerful Visual Tokenizers☆992Nov 25, 2025Updated 2 months ago
- Code for MetaMorph Multimodal Understanding and Generation via Instruction Tuning☆234Jan 22, 2026Updated 3 weeks ago
- [ICCV 2023] Going Beyond Nouns With Vision & Language Models Using Synthetic Data☆14Sep 30, 2023Updated 2 years ago