Ollama - Package Updates
-
[1.4.0]
- Update ollama to 0.16.1
- Full Changelog
- Installing Ollama via the
curlinstall script on macOS will now only prompt for your password if its required - Installing Ollama via the
ieminstall script in Windows will now show progress - Image generation models will now respect the
OLLAMA_LOAD_TIMEOUTvariable - GLM-5: A strong reasoning and agentic model from Z.ai with 744B total parameters (40B active), built for complex systems engineering and long-horizon tasks.
- MiniMax-M2.5: a new state-of-the-art large language model designed for real-world productivity and coding tasks.
- The new
ollamacommand makes it easy to launch your favorite apps with models using Ollama - Launch Pi with
ollama launch pi - Improvements to Ollama's MLX runner to support GLM-4.7-Flash
- Ctrl+G will now allow for editing text prompts in a text editor when running a model
-
[1.4.1]
- Update ollama to 0.16.2
- Full Changelog
ollama launch claudenow supports searching the web when using:cloudmodels- Fixed rendering issue when running
ollamain PowerShell - New setting in Ollama's app makes it easier to disable cloud models for sensitive and private tasks where data cannot leave your computer. For Linux or when running
ollama servemanually, setOLLAMA_NO_CLOUD=1. - Fixed issue where experimental image generation models would not run in 0.16.0 and 0.16.1
-
[1.4.2]
- Update ollama to 0.16.3
- Full Changelog
- New
ollama launch clineadded for the Cline CLI ollama launch <integration>will now always show the model picker- Added Gemma 3, Llama and Qwen 3 architectures to MLX runner
-
[1.5.0]
- Update ollama to 0.17.0
- Full Changelog
- OpenClaw can now be installed and configured automatically via Ollama, making it the easiest way to get up and running with OpenClaw with open models like Kimi-K2.5, GLM-5, and Minimax-M2.5.
- When using cloud models, websearch is enabled - allowing OpenClaw to search the internet.
- Improved tokenizer performance
- Ollama's macOS and Windows apps will now default to a context length based on available VRAM
-
[1.5.1]
- Update ollama to 0.17.4
- Full Changelog
- Tool call indices will now be included in parallel tool calls
- Fixed issue where tool calls in the Qwen 3 and Qwen 3.5 model families would not be parsed correctly if emitted during thinking
- Fixed issue where Ollama's app on Windows would crash when a new update has been downloaded
- Nemotron architecture support in Ollama's engine
- MLX engine now has improved memory usage
- Ollama's app will now allow models that support tools to use web search capabilities
- Improved LFM2 and LFM2.5 models in Ollama's engine
ollama createwill no longer default to affine quantization for unquantized models when using the MLX engine- Added configuration for disabling automatic update downloading
-
[1.5.2]
- Update ollama to 0.17.5
- Full Changelog
- Qwen3.5: the small Qwen 3.5 model series is now available in 0.8B, 2B, 4B and 9B parameter sizes.
- Fixed crash in Qwen 3.5 models when split over GPU & CPU
- Fixed issue where Qwen 3.5 models would repeat themselves due to no presence penalty (note: you may have to redownload the
qwen3.5models:ollama pull qwen3.5:35bfor example) ollama run --verbosewill now show peak memory usage when using Ollama's MLX engine- Fixed memory issues and crashes in MLX runner
- Fixed issue where Ollama would not be able to run models imported from Qwen3.5 GGUF files
-
[1.5.3]
- Update ollama to 0.17.6
- Full Changelog
- Fixed issue where GLM-OCR would not work due to incorrect prompt rendering
- Fixed tool calling parsing and rendering for Qwen 3.5 models
-
[1.5.4]
- Update ollama to 0.17.7
- Full Changelog
- Allow thinking levels such as
"medium"to correctly interpreted in Ollama's API for all thinking models - Add context length to support compaction when using
ollama launch
-
[1.6.1]
- Update ollama to 0.18.1
- Full Changelog
- Web Search and Fetch in OpenClaw
- Ollama now ships with web search and web fetch plugin for OpenClaw. This allows Ollama's models (local or cloud) to search the web for the latest content and news. This also allows OpenClaw with Ollama to be able to fetch the web and extract readable content for processing. This feature does not execute JavaScript.
- When using local models with web search in OpenClaw, ensure you are signed into Ollama with
ollama signin - You can install web search directly into OpenClaw as a plugin if you already have OpenClaw configured and working:
-
[1.6.2]
- Update ollama to 0.18.2
- Full Changelog
- Add extra check to ensure npm and git are installed before installing OpenClaw
- Claude Code will now be faster when run locally, due to preventing cache breakages
- Fix to correctly support
ollama launch openclaw --model <model> - Register Ollama's websearch package correctly for OpenClaw
Hello! It looks like you're interested in this conversation, but you don't have an account yet.
Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.
With your input, this post could be even better 💗
Register Login