Qwen3-omni is a natively end-to-end, omni-modal LLM developed by the Qwen team at Alibaba Cloud, capable of understanding text, audio, images, and video, as well as generating speech in real time.
☆3,488Jan 8, 2026Updated 2 months ago
Alternatives and similar repositories for Qwen3-Omni
Users that are interested in Qwen3-Omni are comparing it to the libraries listed below
Sorting:
- Qwen2.5-Omni is an end-to-end multimodal model by Qwen team at Alibaba Cloud, capable of understanding text, audio, vision, video, and pe…☆3,947Jun 12, 2025Updated 9 months ago
- Qwen3-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.☆18,585Jan 30, 2026Updated last month
- Step-Audio 2 is an end-to-end multi-modal large language model designed for industry-strength audio understanding and speech conversation…☆1,361Feb 13, 2026Updated 3 weeks ago
- The official repo of Qwen2-Audio chat & pretrained large audio language model proposed by Alibaba Cloud.☆2,060Apr 21, 2025Updated 10 months ago
- MiMo-Audio: Audio Language Models are Few-Shot Learners☆981Mar 3, 2026Updated last week
- Kimi-Audio, an open-source audio foundation model excelling in audio understanding, generation, and conversation☆4,508Jun 21, 2025Updated 8 months ago
- Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audi…☆9,799Mar 4, 2026Updated last week
- PyTorch implementation of Audio Flamingo: Series of Advanced Audio Understanding Language Models☆1,009Dec 15, 2025Updated 2 months ago
- Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.☆26,852Jan 9, 2026Updated 2 months ago
- Open-source unified multimodal model☆5,723Oct 27, 2025Updated 4 months ago
- LLaMA-Omni is a low-latency and high-quality end-to-end speech interaction model built upon Llama-3.1-8B-Instruct, aiming to achieve spee…☆3,128May 19, 2025Updated 9 months ago
- A Framework for Speech, Language, Audio, Music Processing with Large Language Model☆995Jan 15, 2026Updated last month
- GLM-4-Voice | 端到端中英语音对话模型☆3,144Dec 5, 2024Updated last year
- Qwen-Image is a powerful image generation foundation model capable of complex text rendering and precise image editing.☆7,505Feb 10, 2026Updated last month
- A Gemini 2.5 Flash Level MLLM for Vision, Speech, and Full-Duplex Multimodal Live Streaming on Your Phone☆24,027Feb 23, 2026Updated 2 weeks ago
- ✨✨[NeurIPS 2025] VITA-1.5: Towards GPT-4o Level Real-Time Vision and Speech Interaction☆2,494Mar 28, 2025Updated 11 months ago
- Next-Token Prediction is All You Need☆2,367Jan 12, 2026Updated 2 months ago
- GLM-4.6V/4.5V/4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning☆2,201Jan 27, 2026Updated last month
- Kimi-VL: Mixture-of-Experts Vision-Language Model for Multimodal Reasoning, Long-Context Understanding, and Strong Agent Capabilities☆1,164Jul 15, 2025Updated 7 months ago
- 🤗 R1-AQA Model: mispeech/r1-aqa☆314Mar 28, 2025Updated 11 months ago
- The official repo of MiniMax-Text-01 and MiniMax-VL-01, large-language-model & vision-language-model based on Linear Attention☆3,362Jul 7, 2025Updated 8 months ago
- The official repo of Qwen-Audio (通义千问-Audio) chat & pretrained large audio language model proposed by Alibaba Cloud.☆1,879Jul 5, 2024Updated last year
- A native-PyTorch library for large scale M-LLM (text/audio) training with tp/cp/dp.☆225Aug 6, 2025Updated 7 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆72,827Updated this week
- ☆4,619Feb 13, 2026Updated 3 weeks ago
- Baichuan-Audio: A Unified Framework for End-to-End Speech Interaction☆218Feb 28, 2025Updated last year
- open-source multimodal large language model that can hear, talk while thinking. Featuring real-time end-to-end speech input and streaming…☆3,530Nov 5, 2024Updated last year
- LLaSA: Scaling Train-time and Inference-time Compute for LLaMA-based Speech Synthesis☆659Jan 21, 2026Updated last month
- Fast and memory-efficient exact attention☆22,719Updated this week
- ✨✨Freeze-Omni: A Smart and Low Latency Speech-to-speech Dialogue Model with Frozen LLM☆368May 27, 2025Updated 9 months ago
- text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)☆12,477Nov 4, 2025Updated 4 months ago
- SALMONN family: A suite of advanced multi-modal LLMs☆1,391Feb 3, 2026Updated last month
- Agent framework and applications built upon Qwen>=3.0, featuring Function Calling, MCP, Code Interpreter, RAG, Chrome extension, etc.☆15,126Mar 4, 2026Updated last week
- DeepSeek-VL2: Mixture-of-Experts Vision-Language Models for Advanced Multimodal Understanding☆5,249Feb 26, 2025Updated last year
- OmniVinci is an omni-modal LLM for joint understanding of vision, audio, and language.☆637Feb 26, 2026Updated 2 weeks ago
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.☆24,543Aug 12, 2024Updated last year
- Wan: Open and Advanced Large-Scale Video Generative Models☆15,498Mar 5, 2026Updated last week
- State-of-the-art audio codec with 90x compression factor. Supports 44.1kHz, 24kHz, and 16kHz mono/stereo audio.☆1,723Jan 26, 2026Updated last month
- Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs (Qwen3.5, DeepSeek-R1, GLM-5, InternLM3, Llama4, ...) and 300+ MLLMs (Qwen3-VL, …☆12,956Updated this week