DeepSeek R1 Distill Qwen 7B
DeepSeek-R1 is a first-generation reasoning model built on DeepSeek-V3 (671 billion total parameters, 37 billion activated per token). It uses large-scale reinforcement learning to improve step-by-step thinking and reasoning abilities, demonstrating high performance in math, coding, and multi-step reasoning tasks.
Key Specifications
Parameters
7.6B
Context
-
Release Date
January 20, 2025
Average Score
65.7%
Timeline
Key dates in the model's history
Announcement
January 20, 2025
Last Update
July 19, 2025
Today
March 25, 2026
Technical Specifications
Parameters
7.6B
Training Tokens
14.8T tokens
Knowledge Cutoff
-
Family
-
Capabilities
MultimodalZeroEval
Benchmark Results
Model performance metrics across various tests and benchmarks
Reasoning
Logical reasoning and analysis
GPQA
Diamond, Pass@1 Methodology evaluation, on abilities model generate although would one correct answer for several attempts. For each tasks model generates several independent solutions (usually 5-200), and then solution from total set. Pass@1 evaluates probability that, that model correct answer with first attempts. This metric especially useful for tasks, where exists correct answer, for example, for mathematical tasks, programming or Methodology accounts for data large language models and measures their performance at solving complex tasks • Self-reported
Other Tests
Specialized benchmarks
AIME 2024
Cons@64 Cons@64 — this method improvement accuracy computations in LLM, which consists in intermediate tokens output for obtaining more correct answers. In situations, when model performs computation, complex logical tasks or numerical computation, answers often from-for errors in chain reasoning. Cons@64 solves this problem by means of limitations output model to 64 tokens, actually chains reasoning, which can lead to to When model number tokens, she/it give and answers, that can lead to to more high accuracy. This especially effectively for tasks, requiring numerical computations or specific conclusions. explanation consists in that, that number tokens probability errors in intermediate which then on final answer. Cons@64 not requires model or her/its — this simply output, which can apply to LLM • Self-reported
LiveCodeBench
Pass@1 In research models Pass@1 — this metric, used for evaluation performance language models in tasks, requiring exact output, especially in and mathematical tasks. Definition: Pass@1 measures probability that, that model will solve task with first attempts without necessity generate several options. In difference from metrics, on (such how Pass@k, where k > 1), Pass@1 evaluates ability model generate one solution immediately, that important for real scenarios use, where users correct answer with first times. Application: - Evaluation solutions tasks programming, where one solution can lead to to or work - tasks, where model should not only find answer, but and provide step-by-step solution - Evaluation reliability model in where attempts or : Pass@1 = (Number tasks, solved with first attempts) / (number tasks) Pass@1 especially how score reliability model and her/its abilities to exact reasoning at capabilities set • Self-reported
MATH-500
Pass@1 Metric performance, which measures probability that, that model correctly will solve task with first attempts. In context reasoning with help (sampling-based reasoning) this means probability obtaining correct answer from one output model. Pass@1 is calculated by means of model with different (seeds) and solutions. This metric especially important in scenarios, where at user is only one attempt obtain answer from model, that reflects majority real with LLM. High score Pass@1 indicates on then, that model and solves tasks without necessity in attempts • Self-reported
License & Metadata
License
mit
Announcement Date
January 20, 2025
Last Updated
July 19, 2025
Similar Models
All ModelsDeepSeek R1 Distill Llama 8B
DeepSeek
8.0B
Best score:0.5 (GPQA)
Released:Jan 2025
DeepSeek R1 Distill Qwen 1.5B
DeepSeek
1.8B
Best score:0.3 (GPQA)
Released:Jan 2025
Llama 3.1 Nemotron Nano 8B V1
NVIDIA
8.0B
Best score:0.5 (GPQA)
Released:Mar 2025
Phi 4 Mini Reasoning
Microsoft
3.8B
Best score:0.5 (GPQA)
Released:Apr 2025
DeepSeek R1 Distill Qwen 14B
DeepSeek
14.8B
Best score:0.6 (GPQA)
Released:Jan 2025
DeepSeek R1 Distill Llama 70B
DeepSeek
70.6B
Best score:0.7 (GPQA)
Released:Jan 2025
Price:$0.10/1M tokens
DeepSeek R1 Distill Qwen 32B
DeepSeek
32.8B
Best score:0.6 (GPQA)
Released:Jan 2025
Price:$0.12/1M tokens
DeepSeek-V3.2-Speciale
DeepSeek
685.0B
Released:Nov 2025
Price:$0.28/1M tokens
Recommendations are based on similarity of characteristics: developer organization, multimodality, parameter size, and benchmark performance. Choose a model to compare or go to the full catalog to browse all available AI models.