Easy methods to Entry Ministral 3 through API
TL;DR
Ministral 3 is a household of open-weight, reasoning-optimized fashions accessible in each 3B and 14B variants. The fashions help multimodal reasoning, native perform and gear calling, and an enormous 256K token context window, all launched beneath an Apache 2.0 license.
You may run Ministral 3 instantly on Clarifai utilizing the Playground for interactive testing or combine it into your purposes by way of Clarifai’s OpenAI-compatible API.
This information explains the Ministral 3 structure, the right way to entry it by way of Clarifai, and the way to decide on the correct variant in your manufacturing workloads.
Introduction
Fashionable AI purposes more and more rely on fashions that may motive reliably, keep lengthy context, and combine cleanly into current instruments and APIs. Whereas closed-source fashions have traditionally led in these capabilities, open-source alternate options are quickly closing the hole.Â
Amongst globally accessible open fashions, Ministral 3 ranks alongside DeepSeek and the GPT OSS household on the prime tier. Slightly than concentrating on leaderboard efficiency on benchmarks, Ministral prioritises performances that matter in manufacturing, reminiscent of producing structured outputs, processing giant paperwork, and executing perform calls inside reside programs.
This makes Ministral 3 well-suited for the calls for of actual enterprise purposes, as organisations are more and more adopting open-weight fashions for his or her transparency, deployment flexibility, and talent to run throughout numerous infrastructure setups, from cloud platforms to on-premise programs.
Ministral 3 Structure
Ministral 3 is a household of dense, edge-optimised multimodal fashions designed for environment friendly reasoning, long-context processing, and native or personal deployment. The household at the moment consists of 3B and 14B parameter fashions, every accessible in base, instruct, and reasoning variants.
Ministral 3 14B
The most important mannequin within the Ministral household is a dense, reasoning-post-trained structure optimised for math, coding, STEM, and different multi-step reasoning duties. It combines a ~13.5B-parameter language mannequin with a ~0.4B-parameter imaginative and prescient encoder, enabling native textual content and picture understanding. The 14B reasoning variant achieves 85% accuracy on AIME ’25, delivering state-of-the-art efficiency inside its weight class whereas remaining deployable on real looking {hardware}. It helps context home windows of as much as 256k tokens, making it appropriate for lengthy paperwork and sophisticated reasoning workflows.
Ministral 3 3B
The 3B mannequin is a compact, reasoning-post-trained variant designed for extremely environment friendly deployment. It pairs a ~3.4B-parameter language mannequin with a ~0.4B-parameter imaginative and prescient encoder (~4B whole parameters), offering multimodal capabilities. Just like the 14B mannequin, it helps 256k-token context lengths, enabling long-context reasoning and doc evaluation on constrained {hardware}.
Key Technical Options
- Multimodal Capabilities: All Ministral 3 fashions use a hybrid language-and-vision structure, permitting them to course of textual content and pictures concurrently for duties reminiscent of doc understanding and visible reasoning.
- Lengthy-Context Reasoning: Reasoning variants help as much as 256k tokens, enabling prolonged conversations, giant doc ingestion, and multi-step analytical workflows.
- Environment friendly Inference: The fashions are optimised for edge and personal deployments. The 14B mannequin runs in BF16 on ~32 GB VRAM, whereas the 3B mannequin runs in BF16 on ~16 GB VRAM, with quantised variations requiring considerably much less reminiscence.
- Agentic Workflows: Ministral 3 is designed to work effectively with structured outputs, perform calling, and tool-use, making it appropriate for automation and agent-based programs.
- License: All Ministral 3 variants are launched beneath the Apache 2.0 license, enabling unrestricted industrial use, fine-tuning, and customisation.
Pretraining Benchmark Efficiency

Ministral 3 14B demonstrates robust reasoning capabilities and multilingual efficiency in comparison with equally sized open fashions, whereas sustaining aggressive outcomes on common data duties. It notably excels in reasoning-heavy benchmarks and reveals strong factual recall and multilingual understanding.
Â
|
Benchmark |
Ministral 3 14B |
Gemma 3 12B Base |
Qwen3 14B Base |
Notes |
|
MATH CoT |
67.6 |
48.7 |
62.0 |
Sturdy lead on structured reasoning |
|
MMLU Redux |
82.0 |
76.6 |
83.7 |
Aggressive common data |
|
TriviaQA |
74.9 |
78.8 |
70.3 |
Stable factual recall |
|
Multilingual MMLU |
74.2 |
69.0 |
75.4 |
Sturdy multilingual efficiency |
Â
Accessing Ministral 3 through Clarifai
Stipulations
Earlier than runing Ministral 3 with the Clarifai API, you’ll want to finish a couple of fundamental setup steps:
- Clarifai Account: Create a Clarifai account to entry hosted AI fashions and APIs.
- Private Entry Token (PAT): All API requests require a Private Entry Token. You may generate or copy one from the Settings > Secrets and techniques part of your Clarifai dashboard.
For added SDKs and setup steering, discuss with the Clarifai Quickstart documentation.
Utilizing the API
The examples under use Ministral-3-14B-Reasoning-2512, the biggest mannequin within the Ministral 3 household. It’s optimised for multi-step reasoning, mathematical downside fixing, and long-context workloads, making it well-suited for long-document useecases and agentic purposes. Right here’s the right way to make your first API name to the mannequin utilizing completely different strategies.
Python (OpenAI-Suitable)
Python (Clarifai SDK)
You too can use the Clarifai Python SDK for inference with extra management over era settings. Right here’s the right way to make a prediction and generate streaming output utilizing the SDK:
Node.js (Clarifai SDK)
Right here’s the right way to carry out inference with the Node.js SDK:
Playground
The Clarifai Playground permits you to shortly experiment with prompts, structured outputs, reasoning workflows, and performance calling with out writing any code.
Go to the Playground and select both:
- Ministral-3-3B-Reasoning‑2512

- Ministral-3-14B-Reasoning‑2512

Functions and Use Circumstances
Ministral 3 is designed for groups constructing clever programs that require robust reasoning, long-context understanding, and dependable structured outputs. It performs effectively throughout agentic, technical, multimodal, and business-critical workflows.
Agentic UtilityÂ
Ministral 3 is effectively suited to AI brokers that have to plan, motive, and act throughout a number of steps. It could possibly orchestrate instruments and APIs utilizing structured JSON outputs, which makes it dependable for automation pipelines the place consistency issues.Â
Lengthy Context
Ministral 3 can analyze giant paperwork utilizing its prolonged 256K token context, making it efficient for summarization, info extraction, and query answering over lengthy technical texts.Â
Multimodal Reasoning
Ministral 3 helps multimodal reasoning, permitting purposes to mix textual content and visible inputs in a single workflow. This makes it helpful for image-based queries, doc understanding, or assistants that have to motive over blended inputs.
Conclusion
Ministral 3 offers reasoning-optimized, open-weight fashions which might be prepared for manufacturing use. With a 256K token context window, multimodal inputs, native instrument calling, and OpenAI-compatible API entry by way of Clarifai, it provides a sensible basis for constructing superior AI programs.
The 3B variant is good for low-latency, cost-sensitive deployments, whereas the 14B variant helps deeper analytical workflows. Mixed with Apache 2.0 licensing, Ministral 3 provides groups flexibility, efficiency, and long-term management.
To get began, discover the fashions within the Clarifai Playground or combine them instantly into your purposes utilizing the API.
