Phi-3.5-MoE-instruct
Phi-3.5-MoE-instruct is a Mixture-of-Experts model with approximately 42 billion total parameters (6.6 billion active) and a 128K token context window. It excels at reasoning, math, coding, and multilingual tasks, outperforming larger dense models on many benchmarks. The model underwent a thorough post-training process for safety (SFT + DPO) and is licensed under MIT. This model is ideal for scenarios requiring both efficiency and high performance, especially in multilingual tasks or reasoning-intensive tasks.
Key Specifications
Timeline
Technical Specifications
Benchmark Results
Model performance metrics across various tests and benchmarks
General Knowledge
Programming
Mathematics
Reasoning
Other Tests
License & Metadata
Similar Models
All ModelsPhi 4
Microsoft
Phi 4 Reasoning Plus
Microsoft
Phi 4 Reasoning
Microsoft
Gemma 2 27B
Jamba 1.5 Mini
AI21 Labs
Llama-3.3 Nemotron Super 49B v1
NVIDIA
ERNIE 4.5
Baidu
Llama 3.3 70B Instruct
Meta
Recommendations are based on similarity of characteristics: developer organization, multimodality, parameter size, and benchmark performance. Choose a model to compare or go to the full catalog to browse all available AI models.