deepseek-r1-distill-qwen-14b by openrouter - AI Model Details, Pricing, and Performance Metrics
deepseek-r1-distill-qwen-14b
completionsdeepseek-r1-distill-qwen-14b
DeepSeek R1 Distill Qwen 14B is a distilled large language model based on [Qwen 2.5 14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. Other benchmark results include: - AIME 2024 pass@1: 69.7 - MATH-500 pass@1: 93.9 - CodeForces Rating: 1481 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
DeepSeek R1 Distill Qwen 14B is a distilled large language model based on [Qwen 2.5 14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. Other benchmark results include: - AIME 2024 pass@1: 69.7 - MATH-500 pass@1: 93.9 - CodeForces Rating: 1481 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
Access deepseek-r1-distill-qwen-14b through LangDB AI Gateway
Integrate with deepseek's deepseek-r1-distill-qwen-14b and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.
Free tier available • No credit card required
Statistics
Category Scores
Benchmark Tests
Metric | AIME | AA Coding Index | AAII | AA Math Index | GPQA | HLE | LiveCodeBench | MATH-500 | MMLU-Pro | SciCode |
---|---|---|---|---|---|---|---|---|---|---|
Score | 66.7 | 30.7 | 29.7 | 80.8 | 59.1 | 4.4 | 37.6 | 94.9 | 74.0 | 23.9 |
Compare with Similar Models
Code Examples
Integration samples and API usage
Related Models
Similar models from openrouter