DeepSeek-R1
DeepSeek-R1 is a first-generation reasoning model built on DeepSeek-V3 (671 billion total parameters, 37 billion activated per token). It incorporates large-scale reinforcement learning (RL) to improve chain-of-thought reasoning and logical thinking abilities, demonstrating high performance in mathematical tasks, coding, and multi-step reasoning.
Key Specifications
Timeline
Technical Specifications
Pricing & Availability
Benchmark Results
Model performance metrics across various tests and benchmarks
General Knowledge
Programming
Reasoning
Other Tests
License & Metadata
Articles about DeepSeek-R1

The Two Loops: How China's Open-Source AI Strategy Is Outpacing America
A new USCC report warns that China's open AI models now dominate global downloads. 80% of US startups use Chinese models. Washington is scrambling.

OpenAI Watched Millions of Agent Conversations. Here's What Went Wrong.
OpenAI's internal monitoring caught coding agents deceiving users, bypassing restrictions, and trying to manipulate other AI systems. Less than 1% of the time.

Unsloth Studio Wants to Be the IDE for Local AI — Training Included
The open-source tool combines inference and fine-tuning in one interface, with 70% less VRAM and no-code training for 500+ models. LM Studio should be nervous.
DeepSeek Core Researcher Daya Guo Rumored to Have Left
Reports suggest Daya Guo, a key researcher behind DeepSeek's code intelligence work, has resigned from the Chinese AI lab.
Similar Models
All ModelsDeepSeek R1 Zero
DeepSeek
DeepSeek-V2.5
DeepSeek
DeepSeek-V3.2 (Thinking)
DeepSeek
DeepSeek-V3.2-Exp
DeepSeek
DeepSeek-V3.1
DeepSeek
DeepSeek-V3
DeepSeek
DeepSeek-V3.2 (Non-thinking)
DeepSeek
DeepSeek-V3.2-Speciale
DeepSeek
Recommendations are based on similarity of characteristics: developer organization, multimodality, parameter size, and benchmark performance. Choose a model to compare or go to the full catalog to browse all available AI models.