DeepSeek logo

DeepSeek R1 Distill Qwen 14B

DeepSeek

DeepSeek-R1 is a first-generation reasoning model built on DeepSeek-V3 (671 billion total parameters, 37 billion activated per token). It uses large-scale reinforcement learning (RL) to improve chain-of-thought reasoning and logical thinking abilities, demonstrating high performance in math, coding, and multi-step reasoning tasks.

Key Specifications

Parameters
14.8B
Context
-
Release Date
January 20, 2025
Average Score
71.5%

Timeline

Key dates in the model's history
Announcement
January 20, 2025
Last Update
July 19, 2025
Today
March 25, 2026

Technical Specifications

Parameters
14.8B
Training Tokens
14.8T tokens
Knowledge Cutoff
-
Family
-
Capabilities
MultimodalZeroEval

Benchmark Results

Model performance metrics across various tests and benchmarks

Reasoning

Logical reasoning and analysis
GPQA
# Diamond, Pass@1 Pass@1 — this measure that, how well successfully model solves tasks at one attempts. In given case "Pass@1" proportion tasks, which model successfully solves with first attempts. This score for evaluation abilities model correctly solve tasks without necessity several iterations. In context mathematical tasks Pass@1 means percentage tasks, for which model with first attempts provides correct solution and answer. This score especially in real scenarios use, where users often on correct answer immediately, without necessity repeated queries or In difference from other metrics, such how Pass@k (where solutions at k attempts), Pass@1 is more since not allows capabilities corrections errors or approach to solving tasksSelf-reported
59.1%

Other Tests

Specialized benchmarks
AIME 2024
Cons@64 AI: I with method "Cons@64". Method "Cons@64" 64 from model and their for formation final answer, (Chen et al., 2023). We we use following instructions for sample from model: : Solve following task [task]. Please, thoroughly step for step solution and answer in format ": X". We 64 answer from model. How in we we use first value, which after ":" in capacity final answer model. Then, we we determine majority from these 64 answers how answer. In case we manner we choose answer from those, that have majoritySelf-reported
80.0%
LiveCodeBench
Pass@1 AI, and especially large language model, often solve tasks method and errors. Model tries solve task several times, and if it find correct solution in one from attempts, we we consider, that model capable solve this task. Pass@1 evaluates probability that, that model correct answer with first attempts. This metric can be on basis Pass@k for cases, when model generates k various solutions. For example, if model solves task correctly in 40% cases from 10 attempts, we we can evaluate probability success with first attempts how 40%Self-reported
53.1%
MATH-500
Pass@1 Pass@1 - this metric, which measures proportion tasks, solved model with first attempts. For model, set tasks and Pass@1 is calculated following manner: 1) For each tasks in set model generates one solution, which then is evaluated. tasks, for which model correct solution with first attempts, and is Pass@1. Pass@1 useful for evaluation that, how well well model handles with tasks "with ", without capabilities do several attempts. However this metric can significantly abilities model, since she/it not accounts for, that model can several different approaches to solving tasks or have some in even at solutionsSelf-reported
93.9%

License & Metadata

License
mit
Announcement Date
January 20, 2025
Last Updated
July 19, 2025

Similar Models

All Models

Recommendations are based on similarity of characteristics: developer organization, multimodality, parameter size, and benchmark performance. Choose a model to compare or go to the full catalog to browse all available AI models.