[1.0.1]
Update ollama to 0.12.10
Full Changelog
ollama run now works with embedding models
Fixed errors when running qwen3-vl:235b and qwen3-vl:235b-instruct
Enable flash attention for Vulkan (currently needs to be built from source)
Add Vulkan memory detection for Intel GPU using DXGI+PDH
Ollama will now return tool call IDs from the /api/chat API
Fixed hanging due to CPU discovery
Ollama will now show login instructions when switching to a cloud model in interactive mode
Fix reading stale VRAM data