deepseek-r1-distill-llama-8b by openrouter - AI Model Details, Pricing, and Performance Metrics

deepseek
deepseek-r1-distill-llama-8b
deepseek

deepseek-r1-distill-llama-8b

completions
byopenrouter

DeepSeek R1 Distill Llama 8B is a distilled large language model based on [Llama-3.1-8B-Instruct](/meta-llama/llama-3.1-8b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: - AIME 2024 pass@1: 50.4 - MATH-500 pass@1: 89.1 - CodeForces Rating: 1205 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models. Hugging Face: - [Llama-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B) - [DeepSeek-R1-Distill-Llama-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B) |

Released
Jan 20, 2025
Knowledge
Jul 24, 2024
License
MIT
Context
32K
Input
$0.04 / 1M tokens
Output
$0.04 / 1M tokens
Accepts: text
Returns: text

Access deepseek-r1-distill-llama-8b through LangDB AI Gateway

Recommended

Integrate with deepseek's deepseek-r1-distill-llama-8b and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.

Unified API
Cost Optimization
Enterprise Security
Get Started Now

Free tier available • No credit card required

Instant Setup
99.9% Uptime
10,000+Monthly Requests
Request Volume
Daily API requests
9
Performance (TPS)
Tokens per second
208.31 tokens/s

Category Scores

Benchmark Tests

View Other Benchmarks
AIME
33.3
Mathematics
AA Coding Index
17.6
Programming
AAII
19.5
General
AA Math Index
59.3
Mathematics
GPQA
49.0
STEM (Physics, Chemistry, Biology)
HLE
4.2
General Knowledge
LiveCodeBench
23.3
Programming
MATH-500
85.3
Mathematics
MMLU-Pro
54.3
General Knowledge
SciCode
11.9
Scientific

Code Examples

Integration samples and API usage