lyuchenyang / Macaw-LLM
Macaw-LLM: Multi-Modal Language Modeling with Image, Video, Audio, and Text Integration
β1,541Updated last month
Alternatives and similar repositories for Macaw-LLM:
Users that are interested in Macaw-LLM are comparing it to the libraries listed below
- 𦦠Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing impβ¦β3,228Updated 11 months ago
- Large-scale, Informative, and Diverse Multi-round Chat Data (and Models)β2,348Updated 11 months ago
- [TLLM'23] PandaGPT: One Model To Instruction-Follow Them Allβ781Updated last year
- An Open-source Toolkit for LLM Developmentβ2,758Updated last month
- [EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understandingβ2,920Updated 8 months ago
- Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)β2,646Updated 6 months ago
- Emu Series: Generative Multimodal Models from BAAIβ1,683Updated 4 months ago
- GPT4Tools is an intelligent system that can automatically decide, control, and utilize different visual foundation models, allowing the uβ¦β766Updated last year
- Official implementation of paper "MiniGPT-5: Interleaved Vision-and-Language Generation via Generative Vokens"β860Updated 2 months ago
- [NeurIPS 2023] Official implementations of "Cheap and Quick: Efficient Vision-Language Instruction Tuning for Large Language Models"β515Updated last year
- SpeechGPT Series: Speech Large Language Modelsβ1,345Updated 6 months ago
- mPLUG-Owl: The Powerful Multi-modal Large Language Model Familyβ2,413Updated 3 weeks ago
- Tool Learning for Big Models, Open-Source Solutions of ChatGPT-Pluginsβ2,765Updated last year
- BuboGPT: Enabling Visual Grounding in Multi-Modal LLMsβ505Updated last year
- [ICLR'24 spotlight] An open platform for training, serving, and evaluating large language model for tool learning.β4,889Updated 3 months ago
- Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Modelβ3,423Updated 3 months ago
- Mixture-of-Experts for Large Vision-Language Modelsβ2,082Updated 2 months ago
- Multimodal-GPTβ1,488Updated last year
- Codes for "Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models".β1,109Updated last year
- β765Updated 7 months ago
- γICLR 2024π₯γ Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignmentβ780Updated 10 months ago
- Official repo for MM-REACTβ941Updated last year
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.β917Updated 8 months ago
- β‘LLM Zoo is a project that provides data, models, and evaluation benchmark for large language models.β‘β2,932Updated last year
- Data and code for NeurIPS 2022 Paper "Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering".β630Updated 5 months ago
- An open-source framework for training large multimodal models.β3,823Updated 5 months ago
- β767Updated 6 months ago
- Transform Video as a Document with ChatGPT, CLIP, BLIP2, GRIT, Whisper, LangChain.β548Updated last year
- γEMNLP 2024π₯γVideo-LLaVA: Learning United Visual Representation by Alignment Before Projectionβ3,153Updated 2 months ago
- Accelerating the development of large multimodal models (LMMs) with one-click evaluation module - lmms-eval.β2,116Updated this week