mlfoundations / MINT-1TLinks
MINT-1T: A one trillion token multimodal interleaved dataset.
☆819Updated 11 months ago
Alternatives and similar repositories for MINT-1T
Users that are interested in MINT-1T are comparing it to the libraries listed below
Sorting:
- Code release for "LLMs can see and hear without any training"☆447Updated 2 months ago
- Reference implementation of Megalodon 7B model☆522Updated 2 months ago
- TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones☆1,294Updated last year
- ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Expert…☆1,478Updated last week
- From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)☆730Updated 8 months ago
- PyTorch Implementation of "V* : Guided Visual Search as a Core Mechanism in Multimodal LLMs"☆645Updated last year
- MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.☆1,310Updated 3 months ago
- Codebase for Aria - an Open Multimodal Native MoE☆1,059Updated 6 months ago
- Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation☆775Updated last month
- Yi-1.5 is an upgraded version of Yi, delivering stronger performance in coding, math, reasoning, and instruction-following capability.☆554Updated 8 months ago
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.☆934Updated 4 months ago
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI☆1,394Updated last year
- Open weights language model from Google DeepMind, based on Griffin.☆644Updated last month
- OLMoE: Open Mixture-of-Experts Language Models☆814Updated 4 months ago
- Reaching LLaMA2 Performance with 0.1M Dollars☆983Updated last year
- LLM Analytics☆673Updated 9 months ago
- HPT - Open Multimodal LLMs from HyperGAI☆316Updated last year
- ☆616Updated last year
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills☆751Updated last year
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆897Updated 2 months ago
- Understanding R1-Zero-Like Training: A Critical Perspective☆1,039Updated 3 weeks ago
- [COLM 2025] LIMO: Less is More for Reasoning☆983Updated 2 weeks ago
- [ICML 2024] CLLMs: Consistency Large Language Models☆397Updated 8 months ago
- LLaVA-UHD v2: an MLLM Integrating High-Resolution Semantic Pyramid via Hierarchical Window Transformer☆382Updated 3 months ago
- Unleashing the Power of Reinforcement Learning for Math and Code Reasoners☆685Updated last month
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Models☆1,567Updated last year
- Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.☆2,037Updated 11 months ago
- [ICLR2025 Spotlight🔥] Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆565Updated 5 months ago
- LLM2CLIP makes SOTA pretrained CLIP model more SOTA ever.☆531Updated 3 weeks ago
- This repository provides the code and model checkpoints for AIMv1 and AIMv2 research projects.☆1,331Updated 3 months ago