deepseek-chat-v3.1 by deepinfra - AI Model Details, Pricing, and Performance Metrics

deepseek
deepseek-chat-v3.1
deepseek

deepseek-chat-v3.1

completions
bydeepinfra

DeepSeek-V3.1 is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes via prompt templates. It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) The model improves tool use, code generation, and reasoning efficiency, achieving performance comparable to DeepSeek-R1 on difficult benchmarks while responding more quickly. It supports structured tool calling, code agents, and search agents, making it suitable for research, coding, and agentic workflows. It succeeds the [DeepSeek V3-0324](/deepseek/deepseek-chat-v3-0324) model and performs well on a variety of tasks.

Released
Aug 21, 2025
Knowledge
Feb 22, 2025
Context
163840
Input
$0.3 / 1M tokens
Output
$1 / 1M tokens
Capabilities: tools
Accepts: text
Returns: text

Access deepseek-chat-v3.1 through LangDB AI Gateway

Recommended

Integrate with deepseek's deepseek-chat-v3.1 and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.

Unified API
Cost Optimization
Enterprise Security
Get Started Now

Free tier available • No credit card required

Instant Setup
99.9% Uptime
10,000+Monthly Requests
Request Volume
Daily API requests
154
Performance (TPS)
Tokens per second
1276.95 tokens/s

Category Scores

Benchmark Tests

View Other Benchmarks
AA Coding Index
47.2
Programming
AAII
44.8
General
GPQA
73.5
STEM (Physics, Chemistry, Biology)
HLE
6.3
General Knowledge
LiveCodeBench
57.7
Programming
MMLU-Pro
83.3
General Knowledge
SciCode
36.7
Scientific

Code Examples

Integration samples and API usage