deepseek-v3.1-base by openrouter - AI Model Details, Pricing, and Performance Metrics

deepseek
deepseek-v3.1-base
deepseek

deepseek-v3.1-base

completions
byopenrouter

This is a base model, trained only for raw next-token prediction. Unlike instruct/chat models, it has not been fine-tuned to follow user instructions. Prompts need to be written more like training text or examples rather than simple requests (e.g., “Translate the following sentence…” instead of just “Translate this”). DeepSeek-V3.1 Base is a 671B parameter open Mixture-of-Experts (MoE) language model with 37B active parameters per forward pass and a context length of 128K tokens. Trained on 14.8T tokens using FP8 mixed precision, it achieves high training efficiency and stability, with strong performance across language, reasoning, math, and coding tasks.

Released
Dec 25, 2024
Knowledge
Jun 28, 2024
License
MIT + Model License (Commercial use allowed)
Context
163840
Input
$0.2 / 1M tokens
Output
$0.8 / 1M tokens
Accepts: text
Returns: text

Access deepseek-v3.1-base through LangDB AI Gateway

Recommended

Integrate with deepseek's deepseek-v3.1-base and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.

Unified API
Cost Optimization
Enterprise Security
Get Started Now

Free tier available • No credit card required

Instant Setup
99.9% Uptime
10,000+Monthly Requests
Request Volume
Daily API requests
11
Performance (TPS)
Tokens per second
109.56 tokens/s

Category Scores

Benchmark Tests

View Other Benchmarks
AIME
25.3
Mathematics
AA Coding Index
35.6
Programming
AAII
32.5
General
AA Math Index
26.0
Mathematics
DROP
91.6
General Knowledge
GPQA
57.4
STEM (Physics, Chemistry, Biology)
HLE
3.6
General Knowledge
LiveCodeBench
35.9
Programming
MATH-500
88.7
Mathematics
MMLU
88.5
General Knowledge
MMLU-Pro
75.6
General Knowledge
SciCode
35.4
Scientific

Code Examples

Integration samples and API usage