Overview
Ollama enables you to run AI models locally on your machine or on a private network endpoint. This gives you complete control over your data, eliminates API costs, and enables offline development. Ollama can automatically download and manage models for you, making local inference accessible and straightforward.Prerequisites
Before configuring Ollama in cmd, you need to have Ollama installed and running on your system.Install Ollama
Download and install Ollama from the official website or follow the Quick Start guide.
Start Ollama
Ensure Ollama is running. The default installation sets up autostart, but you can also start it manually:Ollama will run on
http://localhost:11434 by default.Install Models
Install one or more AI models. Ollama will automatically download the models when you first request them:You can browse available models at https://ollama.com/search.
Ollama automatically loads models on-demand. You don’t need to manually load models before using them with cmd.
Configuring Ollama in cmd
Once Ollama is installed and running, you can enable it in cmd:Add Ollama Provider
Click “Add Provider” or “Configure” for Ollama.By default, cmd connects to Ollama on
http://localhost:11434. If you’re running Ollama on a different port or on a remote server, you can customize the base URL.Example: Default local configuration
Example: Default local configuration
For a standard local Ollama installation, simply toggle “Enable”
Example: Custom port
Example: Custom port
If you’re running Ollama on a different port:
Example: Remote server
Example: Remote server
If you’re running Ollama on a remote server in your private network:
Model discovery
cmd will automatically discover the available models from your Ollama installation when the provider is configured.
Using Ollama
Once enabled and models are selected, you can use Ollama models just like any other AI provider:Select an Ollama Model
In the cmd interface, open the model selector and choose one of your Ollama models (e.g., “devstral-small-2:24b”)
Tested Models
The following models have been successfully tested with cmd:- qwen3-coder:30b - High-performance coding model
- devstral-small-2:24b - Efficient development assistant
- deepseek-r1:14b - Reasoning-focused model
- qwen2.5-coder:7b - Lightweight coding model
Benefits of Local Inference
Using Ollama with cmd provides several advantages:Data Privacy
All processing happens locally - your code never leaves your machine
No API Costs
Run unlimited inference without per-token charges
Offline Development
Work without internet connectivity
Custom Models
Use specialized or fine-tuned models for your specific needs

