qwen3-235b-a22b-thinking-2507 by deepinfra - AI Model Details, Pricing, and Performance Metrics

qwen
qwen3-235b-a22b-thinking-2507
qwen

qwen3-235b-a22b-thinking-2507

completions
On:deepinfraparasail

Qwen3-235B-A22B-Thinking-2507 is a high-performance, open-weight Mixture-of-Experts (MoE) language model optimized for complex reasoning tasks. It activates 22B of its 235B parameters per forward pass and natively supports up to 262,144 tokens of context. This "thinking-only" variant enhances structured logical reasoning, mathematics, science, and long-form generation, showing strong benchmark performance across AIME, SuperGPQA, LiveCodeBench, and MMLU-Redux. It enforces a special reasoning mode (</think>) and is designed for high-token outputs (up to 81,920 tokens) in challenging domains. The model is instruction-tuned and excels at step-by-step reasoning, tool use, agentic workflows, and multilingual tasks. This release represents the most capable open-source variant in the Qwen3-235B series, surpassing many closed models in structured reasoning use cases.

ProviderInputOutput
deepinfra
deepinfra
$0.13 / 1M tokens$0.6 / 1M tokens
parasail
parasail
$0.65 / 1M tokens$3 / 1M tokens
Released
Apr 29, 2025
Knowledge
Oct 31, 2024
Context
262144
Input
$0.13 / 1M tokens
Output
$0.6 / 1M tokens
Accepts: text
Returns: text

Access qwen3-235b-a22b-thinking-2507 through LangDB AI Gateway

Recommended

Integrate with qwen's qwen3-235b-a22b-thinking-2507 and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.

Unified API
Cost Optimization
Enterprise Security
Get Started Now

Free tier available • No credit card required

Instant Setup
99.9% Uptime
10,000+Monthly Requests
Available from 2 providers
Provider:

Category Scores

Benchmark Tests

View Other Benchmarks
GPQA
81.1
STEM (Physics, Chemistry, Biology)
MMLU-Pro
84.4
General Knowledge

Code Examples

Integration samples and API usage