QwenLM / Qwen3-VL-EmbeddingLinks
☆385Updated this week
Alternatives and similar repositories for Qwen3-VL-Embedding
Users that are interested in Qwen3-VL-Embedding are comparing it to the libraries listed below
Sorting:
- Official Repository for "Glyph: Scaling Context Windows via Visual-Text Compression"☆544Updated 2 months ago
- A Scientific Multimodal Foundation Model☆623Updated 3 months ago
- ☆462Updated 3 weeks ago
- Fully Open Framework for Democratized Multimodal Training☆682Updated 2 weeks ago
- This repository collects and organises state‑of‑the‑art papers on spatial reasoning for Multimodal Vision–Language Models (MVLMs).☆260Updated last week
- A reproduction of the Deepseek-OCR model including training☆201Updated last month
- OmniVinci is an omni-modal LLM for joint understanding of vision, audio, and language.☆620Updated 2 months ago
- ACL 2025: Synthetic data generation pipelines for text-rich images.☆152Updated 10 months ago
- MiMo-VL☆619Updated 4 months ago
- This is the official Python version of Vision-Zero: Scalable VLM Self-Improvement via Strategic Gamified Self-Play.☆107Updated 2 months ago
- Cook up amazing multimodal AI applications effortlessly with MiniCPM-o☆231Updated last month
- NextFlow🚀: Unified Sequential Modeling Activates Multimodal Understanding and Generation☆128Updated this week
- WeDLM: The fastest diffusion language model with standard causal attention and native KV cache compatibility, delivering real speedups ov…☆550Updated this week
- ☆498Updated 3 weeks ago
- 🚀ReVisual-R1 is a 7B open-source multimodal language model that follows a three-stage curriculum—cold-start pre-training, multimodal rei…☆191Updated last month
- Official implementation of "Grasp Any Region: Towards Precise, Contextual Pixel Understanding for Multimodal LLMs".☆96Updated 2 months ago
- ☆182Updated 5 months ago
- The official repository of "R-4B: Incentivizing General-Purpose Auto-Thinking Capability in MLLMs via Bi-Mode Integration"☆134Updated 4 months ago
- UniversalRAG: Retrieval-Augmented Generation over Corpora of Diverse Modalities and Granularities☆131Updated 7 months ago
- [ACL 2025 🔥] Rethinking Step-by-step Visual Reasoning in LLMs☆310Updated 7 months ago
- Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model Reasoning Ability in VibeThinker-1.5B☆559Updated last month
- An official implementation of "CapRL: Stimulating Dense Image Caption Capabilities via Reinforcement Learning"☆169Updated 2 weeks ago
- The official repository of the dots.vlm1 instruct models proposed by rednote-hilab.☆277Updated 3 months ago
- MMSearch-R1 is an end-to-end RL framework that enables LMMs to perform on-demand, multi-turn search with real-world multimodal search too…☆380Updated 4 months ago
- [ArXiv 2025] DiffusionVL: Translating Any Autoregressive Models into Diffusion Vision Language Models☆121Updated 2 weeks ago
- Ling-V2 is a MoE LLM provided and open-sourced by InclusionAI.☆250Updated 3 months ago
- [ACL2025 Oral & Award] Evaluate Image/Video Generation like Humans - Fast, Explainable, Flexible☆114Updated 5 months ago
- ☆56Updated last year
- OpenVLThinker: An Early Exploration to Vision-Language Reasoning via Iterative Self-Improvement☆126Updated 5 months ago
- DiffThinker: Towards Generative Multimodal Reasoning with Diffusion Models☆150Updated last week