Qwen logo

Qwen2.5-Omni-7B

Overview

Overview

Qwen2.5-Omni is the flagship end-to-end multimodal model in the Qwen series. It processes diverse inputs including text, images, audio, and video, delivering real-time streaming responses through text generation and natural speech synthesis using a novel Thinker-Talker architecture.

Qwen2.5-Omni-7B was released on March 27, 2025.

Performance

Timeline

ReleasedUnknown
Knowledge CutoffUnknown

Specifications

Parameters
7.0B
License
Apache 2.0
Training Data
Unknown

Benchmarks

Benchmarks

Qwen2.5-Omni-7B Performance Across Datasets

Scores sourced from the model's scorecard, paper, or official blog posts

LLM Stats Logollm-stats.com - Wed Jan 28 2026
Notice missing or incorrect data?Start an Issue discussion

Pricing

Pricing

Pricing, performance, and capabilities for Qwen2.5-Omni-7B across different providers:

No pricing information available for this model.

API Access

API Access Coming Soon

API access for Qwen2.5-Omni-7B will be available soon through our gateway.

Recent Posts

Recent Reviews

FAQ

Common questions about Qwen2.5-Omni-7B

Qwen2.5-Omni-7B was released on March 27, 2025 by Qwen.
Qwen2.5-Omni-7B was created by Qwen.
Qwen2.5-Omni-7B has 7.0 billion parameters.
Qwen2.5-Omni-7B is released under the Apache 2.0 license. This is an open-source/open-weight license.
Yes, Qwen2.5-Omni-7B is a multimodal model that can process both text and images as input.