Mistral Small 4 logo
Updated April 2026

Review of Mistral Small 4

Mistral Small 4 is the new hybrid model from Mistral AI, blending Magistral's reasoning, Pixtral's multimodal, and Devstral's agentic coding capabilities. The Mixture of Experts architecture activates 4 of 128 experts per token, totalling 119B parameters with 6B active. Released under Apache 2.0, it ships through the Mistral API, La Plateforme, and Le Chat, and powers everything from conversational assistants to autonomous developer agents.

4.8/5(91)
enfr#AI Assistant#Code Generation#Open Source#Autonomous Agents

Mistral Small 4: Le modèle Mistral qui unifie chat, raisonnement, code et multimodal dans une seule architecture.

Try Mistral Small 4

Best for

  • Developers building custom AI agents
  • Companies needing a sovereign European model
  • ML teams running on-premise or private cloud
  • Use cases mixing reasoning and high-end coding

Not ideal for

  • Users seeking a plug-and-play chatbot
  • Non-technical profiles with no AI pipeline
  • Small teams without fine-tuning resources
  • Workflows focused only on pure media creation
  • MoE architecture with 128 experts for top efficiency
  • Native multimodal text and image in a single model
  • Configurable reasoning effort from fast to deep
  • Open source under Apache 2.0, deployable on-premise
  • Specialized for agentic coding and developer tasks
  • Official support via La Plateforme, Le Chat, and API
  • ⚠️ Local inference still demanding despite sparse activation
  • ⚠️ Documentation skewed toward advanced engineers
  • ⚠️ No native no-code interface bundled with the model
  • ⚠️ API costs add up on high-volume workloads

Mistral Small 4 reaffirms Mistral AI's ability to ship competitive open models against US giants. Folding Magistral, Pixtral, and Devstral into a unified Mixture of Experts delivers a rare versatility: deep reasoning, native text-image multimodal, and agentic coding in a single model. The Apache 2.0 license guarantees full deployment freedom, especially valuable for organizations focused on digital sovereignty. The architecture's efficiency (6B active out of 119B total) keeps inference costs in check for serious workloads. Some caveats remain: documentation still leans toward ML experts, and local inference requires a solid GPU stack. For European technical teams looking for a flexible sovereign model, Mistral Small 4 stands out as a major reference in the open-source AI landscape today.

Is Mistral Small 4 truly open source?

Yes, the model is released under Apache 2.0, which allows commercial use, modification, and on-premise deployment.

What multimodal capabilities does it offer?

Mistral Small 4 natively handles text and image inputs without requiring an additional vision model.

How does the MoE architecture work?

The Mixture of Experts activates 4 of 128 experts per token, using 6 billion active parameters out of 119 billion total.

Where can I test Mistral Small 4?

The model is available on Le Chat, La Plateforme Mistral, the official API, and via Apache 2.0 download for on-premise deployments.

Is Mistral Small 4 fit for AI agents?

Yes, thanks to Devstral's heritage, it is particularly well suited to agentic coding and developer-oriented workflows.

⚠️ Disclosure: some links are affiliate links (no impact on your price).