mlfoundations / MINT-1TLinks
š MINT-1T: A one trillion token multimodal interleaved dataset.
ā828Updated last year
Alternatives and similar repositories for MINT-1T
Users that are interested in MINT-1T are comparing it to the libraries listed below
Sorting:
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.ā952Updated 10 months ago
- Reference implementation of Megalodon 7B modelā529Updated 8 months ago
- Code release for "LLMs can see and hear without any training"ā457Updated 9 months ago
- TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbonesā1,307Updated last week
- PyTorch Implementation of "V* : Guided Visual Search as a Core Mechanism in Multimodal LLMs"ā688Updated 2 years ago
- Yi-1.5 is an upgraded version of Yi, delivering stronger performance in coding, math, reasoning, and instruction-following capability.ā557Updated last year
- NeurIPS 2025 Spotlight; ICLR2024 Spotlight; CVPR 2024; EMNLP 2024ā1,812Updated 2 months ago
- MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.ā1,409Updated 9 months ago
- Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generationā823Updated 7 months ago
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skillsā763Updated 2 years ago
- Open weights language model from Google DeepMind, based on Griffin.ā663Updated last week
- From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)ā789Updated last year
- Understanding R1-Zero-Like Training: A Critical Perspectiveā1,205Updated 5 months ago
- LLM Analyticsā705Updated last year
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAIā1,407Updated last year
- HPT - Open Multimodal LLMs from HyperGAIā314Updated last year
- ā643Updated last year
- [ICML 2024] CLLMs: Consistency Large Language Modelsā411Updated last year
- LLaVA-UHD v3: Progressive Visual Compression for Efficient Native-Resolution Encoding in MLLMsā413Updated last month
- This repository provides the code and model checkpoints for AIMv1 and AIMv2 research projects.ā1,396Updated 6 months ago
- OLMoE: Open Mixture-of-Experts Language Modelsā967Updated 4 months ago
- DataComp for Language Modelsā1,416Updated 5 months ago
- [COLM 2025] LIMO: Less is More for Reasoningā1,062Updated 6 months ago
- Visualize the intermediate output of Mistral 7Bā384Updated last year
- A family of lightweight multimodal models.ā1,051Updated last year
- Inference code for Persimmon-8Bā412Updated 2 years ago
- This repo contains evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Eā¦ā544Updated 8 months ago
- Emu Series: Generative Multimodal Models from BAAIā1,764Updated last month
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Modelsā1,657Updated last year
- LLM2CLIP significantly improves already state-of-the-art CLIP models.ā623Updated last week