January 31, 2025

How to Download and Run Locally Mistral Small 3 on Windows 11

  • In this tutorial, we explain how to download, install, and run on a local computer the Mistral Small 3 large language model (LLM).
  • For people new to LLMs: The Mistral series of models is published by the French company called Mistral AI. The Mistrial AI company is in fourth place in the global AI race and in first place outside of the San Francisco Bay Area (did they consider DeepSeek?).
  • Mistral Small 3 is a relatively small model that is optimized for latency. The model has “only” 24B parameters and its quantized version is only about 14 GB large.
  • Mistral Small 3 is comparable in terms of performance with larger models such as Llama 3.3 70B and Qwen 32B.
  • It is released under Apache 2.0 license which is a a free, permissive license that allows users to modify, use, and distribute software.
  • It is an excellent low-cost replacement for GPT4o-mini.
  • The performance of Mistral Small 3 is similar to Llama 3.3 70B instruct while being more than 3x faster on the same hardware.

The YouTube tutorial explaining how to install and run Mistral Small 3 is given below.

When to Use Mistral Small 3

  • On consumer-level hardware. It works well on our NVIDIA 3090 GPU.
  • When quick and accurate responses are required since the model is relatively small.
  • Low-latency function calling – ideal for RAG, Internet agents, math reasoning, etc.
  • To fine-tune to create subject matter experts. The model can be “easily” fine tuned to create subject matter experts, such as in the fields of medial diagnositcs, technical support, troubleshooting, etc.

We were able to successfully run Mistral Small 3 on a desktop computer with the following specifications

  • GPU: NVIDIA 3090
  • 64 GB of regular RAM
  • Intel i9 processor

Installation Instructions

To install Mistral Small 3 we are going to use Ollama. The first step is to download and install Ollama. To download and install Ollama, go to the official Ollama website

https://www.ollama.com

and click on the download button to download the installation file

Once the file is downloaded, run it to install Ollama. After Ollama is installed, open a Command Prompt and type the following command

ollama

to verify the installation of Ollama. If Ollama is properly installed, the response should look like this

The next step is to download and install the model. To download the model, go back to the Ollama website, and in the search menu search for “mistral-small

then, click on the model link, and the following webpage will appear

Then, copy the command for installing and running the Mistral Small 3 model as shown in the figure below.

Then, open the command prompt and execute the command in the command prompt

ollama run mistral-small

This command will download and install the model. After the model is installed it will automatically run and you can start asking questions. To exit the model, press CTRL+d. If you want to run the model again, then type

ollama list

to list the exact model name. To run the model, you need to type “ollama run <model name>”, where <model name> should be replaced with the model. The command for running the model should look like this

ollama run mistral-small

that is the same command that was initially used to download the model. However, this time, since the model is downloaded, the model will only be executed.