SkalskiP / awesome-foundation-and-multimodal-modelsView external linksLinks
๐๏ธ + ๐ฌ + ๐ง = ๐ค Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]
โ637Feb 29, 2024Updated last year
Alternatives and similar repositories for awesome-foundation-and-multimodal-models
Users that are interested in awesome-foundation-and-multimodal-models are comparing it to the libraries listed below
Sorting:
- Must-have resource for anyone who wants to experiment with and build on the OpenAI vision API ๐ฅโ1,684Jan 14, 2025Updated last year
- streamline the fine-tuning process for multimodal models: PaliGemma 2, Florence-2, and Qwen2.5-VLโ2,659Updated this week
- โ15Dec 7, 2023Updated 2 years ago
- [CVPR 2024] VCoder: Versatile Vision Encoders for Multimodal Large Language Modelsโ279Apr 17, 2024Updated last year
- Each week I create sketches covering key Computer Vision concepts. If you want to learn more about CV stick around!โ150Mar 13, 2023Updated 2 years ago
- โ717Mar 6, 2024Updated last year
- ใTMM 2025๐ฅใ Mixture-of-Experts for Large Vision-Language Modelsโ2,302Jul 15, 2025Updated 7 months ago
- The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained modeโฆโ12Jul 30, 2024Updated last year
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skillsโ763Feb 1, 2024Updated 2 years ago
- Gradio UI for a Cog APIโ70Apr 8, 2024Updated last year
- This repository is a curated collection of the most exciting and influential CVPR 2024 papers. ๐ฅ [Paper + Code + Demo]โ742Jun 2, 2025Updated 8 months ago
- NeurIPS 2025 Spotlight; ICLR2024 Spotlight; CVPR 2024; EMNLP 2024โ1,812Nov 27, 2025Updated 2 months ago
- Official codebase used to develop Vision Transformer, SigLIP, MLP-Mixer, LiT and more.โ3,355May 19, 2025Updated 8 months ago
- โ547Nov 7, 2024Updated last year
- Recipes for shrinking, optimizing, customizing cutting edge vision models. ๐โ1,875Jan 9, 2026Updated last month
- a state-of-the-art-level open visual language model | ๅคๆจกๆ้ข่ฎญ็ปๆจกๅโ6,724May 29, 2024Updated last year
- Official Code for Tracking Any Object Amodallyโ120Jul 11, 2024Updated last year
- Images to inference with no labeling (use foundation models to train supervised models).โ2,624May 14, 2025Updated 9 months ago
- A collection of tutorials on state-of-the-art computer vision models and techniques. Explore everything from foundational architectures lโฆโ9,172Feb 3, 2026Updated 2 weeks ago
- Radiantloom Email Assist 7B is an email-assistant large language model fine-tuned from Zephyr-7B-Beta, over a custom-curated dataset of 1โฆโ14Jan 19, 2024Updated 2 years ago
- โ135Nov 24, 2023Updated 2 years ago
- Testing and evaluating the capabilities of Vision-Language models (PaliGemma) in performing computer vision tasks such as object detectioโฆโ85May 29, 2024Updated last year
- Unofficial implementation and experiments related to Set-of-Mark (SoM) ๐๏ธโ88Oct 20, 2023Updated 2 years ago
- ๐ฅ๐ฅ LLaVA++: Extending LLaVA with Phi-3 and LLaMA-3 (LLaVA LLaMA-3, LLaVA Phi-3)โ848Aug 5, 2025Updated 6 months ago
- [CVPR 2024] Real-Time Open-Vocabulary Object Detectionโ6,208Feb 26, 2025Updated 11 months ago
- List of resources, libraries and more for developers who would like to build with open-source machine learning off-the-shelfโ198Apr 1, 2024Updated last year
- YOLOExplorer : Iterate on your YOLO / CV datasets using SQL, Vector semantic search, and more within secondsโ140Feb 2, 2026Updated 2 weeks ago
- This repository is a curated collection of the most exciting and influential CVPR 2023 papers. ๐ฅ [Paper + Code]โ652Jun 2, 2025Updated 8 months ago
- Run Mixtral-8x7B models in Colab or consumer desktopsโ2,325Apr 8, 2024Updated last year
- InternLM-XComposer2.5-OmniLive: A Comprehensive Multimodal System for Long-term Streaming Video and Audio Interactionsโ2,919May 26, 2025Updated 8 months ago
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.โ24,446Aug 12, 2024Updated last year
- [CVPR2024] ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual Promptsโ336Jul 17, 2024Updated last year
- A family of lightweight multimodal models.โ1,051Nov 18, 2024Updated last year
- โก Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Plโฆโ2,174Oct 8, 2024Updated last year
- 4M: Massively Multimodal Masked Modelingโ1,789Jun 2, 2025Updated 8 months ago
- Latest Advances on Multimodal Large Language Modelsโ17,337Feb 7, 2026Updated last week
- โ444Apr 1, 2024Updated last year
- [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parametersโ5,936Mar 14, 2024Updated last year
- TorchMultimodal is a PyTorch library for training state-of-the-art multimodal multi-task models at scale.โ1,699Updated this week