deepseek-r1-distill-llama-70b by openrouter - AI Model Details, Pricing, and Performance Metrics

deepseek
deepseek-r1-distill-llama-70b
Try
deepseek

deepseek-r1-distill-llama-70b

completions
byopenrouter

DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: - AIME 2024 pass@1: 70.0 - MATH-500 pass@1: 94.5 - CodeForces Rating: 1633 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.

Released
Jan 20, 2025
Knowledge
Jul 24, 2024
License
MIT
Context
131072
Input
$0.11 / 1M tokens
Output
$0.38 / 1M tokens
Accepts: text
Returns: text

Access deepseek-r1-distill-llama-70b through LangDB AI Gateway

Recommended

Integrate with deepseek's deepseek-r1-distill-llama-70b and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.

Unified API
Cost Optimization
Enterprise Security
Get Started Now

Free tier available • No credit card required

Instant Setup
99.9% Uptime
10,000+Monthly Requests
Request Volume
Daily API requests
6
Performance (TPS)
Tokens per second
115.83 tokens/s

Category Scores

Benchmark Tests

View Other Benchmarks
HLE
6.1
General Knowledge
AIME
67.0
Mathematics
GPQA
65.2
STEM (Physics, Chemistry, Biology)
SciCode
31.2
Scientific
MATH-500
93.5
Mathematics
MMLU-Pro
79.5
General Knowledge
LiveCodeBench
26.6
Programming
AA Math Index
53.7
Mathematics
AA Coding Index
19.7
Programming
AAII
29.9
General

Code Examples

Integration samples and API usage