You can try out the latest Ollama models on VS Code for free.
We are using Ollama, which is a free local AI model running application developed by the Llama community.
Installing and using Ollama
You can download Ollama from its website.
Now you'll be able to access ollama using your terminal.
- Open your terminal.
Run a model you like(depending on your hardware), using the command:
ollama run qwen3:4b
This command will pull and run the model.

Change the model name to your preferred model and install it.
To view all the available models, go to ollama.com/search
ㅤIf you want to run a high-end AI model, you can use Ollama Cloud for free.
Run them like this:ollama pull qwen3-coder:480b-cloud
ㅤ
ㅤ
Integrating with VS Code
Make sure the Ollama server is running in the background.
Verification: Check this URL, localhost:11434, and see if Ollama is running.
If not: Run it using the command ollama serve.
ㅤ
- Open VS Code -> Copilot chat sidebar.
- Select the model dropdown -> Manage models -> Select Ollama Select the desired models. Open the model dropdown and choose the model.
The option will disappear once you turn the Ollama server off.


Top comments (0)