mbzuai-oryx / AINLinks
AIN - The First Arabic Inclusive Large Multimodal Model. It is a versatile bilingual LMM excelling in visual and contextual understanding across diverse domains.
β46Updated 5 months ago
Alternatives and similar repositories for AIN
Users that are interested in AIN are comparing it to the libraries listed below
Sorting:
- [ACL 2025 π₯] A Comprehensive Multi-Domain Benchmark for Arabic OCR and Document Understandingβ49Updated 3 months ago
- [NAACL 2025 π₯] CAMEL-Bench is an Arabic benchmark for evaluating multimodal models across eight domains with 29,000 questions.β32Updated 4 months ago
- [CVPR 2025 π₯] ALM-Bench is a multilingual multi-modal diverse cultural benchmark for 100 languages across 19 categories. It assesses theβ¦β45Updated 3 months ago
- [EMNLP'23] ClimateGPT: a specialized LLM for conversations related to Climate Change and Sustainability topics in both English and Arabiβ¦β79Updated 11 months ago
- (WACV 2025 - Oral) Vision-language conversation in 10 languages including English, Chinese, French, Spanish, Russian, Japanese, Arabic, Hβ¦β83Updated 3 weeks ago
- Bilingual Medical Mixture of Experts LLMβ31Updated 9 months ago
- Bio-Medical EXpert LMM with English and Arabic Language Capabilitiesβ70Updated 4 months ago
- Composition of Multimodal Language Models From Scratchβ15Updated last year
- vision language models finetuning notebooks & use cases (Medgemma - paligemma - florence .....)β48Updated last month
- [BMVC 2025] Official Implementation of the paper "PerSense: Personalized Instance Segmentation in Dense Images"β26Updated last week
- A minimal implementation of LLaVA-style VLM with interleaved image & text & video processing ability.β96Updated 8 months ago
- ARB: A Comprehensive Arabic Multimodal Reasoning Benchmarkβ15Updated 3 months ago
- [ACL 2025 π₯] Time Travel is a Comprehensive Benchmark to Evaluate LMMs on Historical and Cultural Artifactsβ18Updated 3 months ago
- VideoMathQA is a benchmark designed to evaluate mathematical reasoning in real-world educational videosβ16Updated 2 months ago
- [Fully open] [Encoder-free MLLM] Vision as LoRAβ333Updated 2 months ago
- Official Implementation of DiffCLIP: Differential Attention Meets CLIPβ42Updated 5 months ago
- This is the repo for the paper "PANGEA: A FULLY OPEN MULTILINGUAL MULTIMODAL LLM FOR 39 LANGUAGES"β110Updated 2 months ago
- [InterSpeech 2024] Official code repository of paper titled "Bird Whisperer: Leveraging Large Pre-trained Acoustic Model for Bird Call Clβ¦β36Updated 8 months ago
- [ACCV 2024] ObjectCompose: Evaluating Resilience of Vision-Based Models on Object-to-Background Compositional Changes πππβ37Updated 7 months ago
- β44Updated last year
- β68Updated 2 months ago
- β38Updated 3 months ago
- [CVPRW 2025] Official repository of paper titled "Towards Evaluating the Robustness of Visual State Space Models"β24Updated 2 months ago
- This repository contains codes for fine-tuning LLAVA-1.6-7b-mistral (Multimodal LLM) model.β40Updated 9 months ago
- Code for "Enhancing In-context Learning via Linear Probe Calibration"β35Updated last year
- β42Updated last year
- An implementation of "M3DOCRAG: Multi-modal Retrieval is What You Need for Multi-page Multi-document Understanding" by Jaemin Cho, Debanjβ¦β44Updated 9 months ago
- Official code repository for ICML 2025 paper: "ExPLoRA: Parameter-Efficient Extended Pre-training to Adapt Vision Transformers under Domaβ¦β43Updated last week
- [EMNLP 2024] Official code repository of paper titled "PALM: Few-Shot Prompt Learning for Audio Language Models" accepted in EMNLP 2024 cβ¦β26Updated 8 months ago
- This is the official implementation to the EMNLP 2024 paper: Modeling Layout Reading Order as Ordering Relations for Visually-rich Documeβ¦β26Updated 9 months ago