Mistral AI logo

Mistral Large 3 (675B Instruct 2512)

Multimodal
Mistral AI

Mistral Large 3 is a multimodal Mixture-of-Experts model with 41B active parameters and 675B total parameters, trained from scratch on 3,000 H200 GPUs. This is an FP8 instruct version fine-tuned for instruction tasks, ideal for chat, agentic, and instruction-based scenarios. Designed for reliability and long-context understanding, suitable for production-grade assistants, RAG systems, scientific workloads, and complex enterprise workflows.

Key Specifications

Parameters
675.0B
Context
262.1K
Release Date
December 3, 2025
Average Score
47.8%

Timeline

Key dates in the model's history
Announcement
December 3, 2025
Last Update
January 22, 2026
Today
March 26, 2026

Technical Specifications

Parameters
675.0B
Training Tokens
-
Knowledge Cutoff
-
Family
-
Capabilities
MultimodalZeroEval

Pricing & Availability

Input (per 1M tokens)
$0.50
Output (per 1M tokens)
$1.50
Max Input Tokens
262.1K
Max Output Tokens
262.1K
Supported Features
Function CallingStructured OutputCode ExecutionWeb SearchBatch InferenceFine-tuning

Benchmark Results

Model performance metrics across various tests and benchmarks

Reasoning

Logical reasoning and analysis
GPQA
Self-reported
44.0%

Other Tests

Specialized benchmarks
MMMLU
Self-reported
85.0%
AMC 2022-23
Self-reported
52.0%
LiveCodeBench
Self-reported
34.0%
SimpleQA
Self-reported
24.0%

License & Metadata

License
apache-2.0
Announcement Date
December 3, 2025
Last Updated
January 22, 2026

Similar Models

All Models

Recommendations are based on similarity of characteristics: developer organization, multimodality, parameter size, and benchmark performance. Choose a model to compare or go to the full catalog to browse all available AI models.