News: 0001591890

  ARM Give a man a fire and he's warm for a day, but set fire to him and he's warm for the rest of his life (Terry Pratchett, Jingo)

ollama 0.12.11 Brings Vulkan Acceleration

([AI] 3 Hours Ago ollama 0.12.11)


ollama 0.12.11 released this week as the newest feature update to this easy-to-run method of deploying OpenAI GPT-OSS, DeepSeek-R1, Gemma 3, and other large language models. Exciting with ollama 0.12.11 is that it's now supporting the Vulkan API.

Launching ollama with the OLLAMA_VULKAN=1 environment variable set will now enable Vulkan API support as an alternative to the likes of AMD ROCm and NVIDIA CUDA acceleration. This is great for open-source Vulkan drivers, older AMD graphics cards lacking ROCm support, or even any AMD setup with the RADV driver present but not having installed ROCm. As we've seen when testing Llama.cpp with Vulkan, in some cases using Vulkan can be faster than using the likes of ROCm.

[1]This commit in ollama 0.12.11 lays out all the details on the Vulkan API support for ollama. Over the past few weeks [2]ollama offered the Vulkan support as experimental .

The ollama 0.12.11 release also adds API support for Logprobs, support for WebP images within their new app, improved rendering performance, preferring discrete GPUs over integrated GPUs when scheduling models, and various other fixes and enhancements.

Downloads and more details on the ollama 0.12.11 release via [3]GitHub .



[1] https://github.com/ollama/ollama/commit/2aba569a2a593f56651ded7f5011480ece70c80f

[2] https://www.phoronix.com/news/ollama-Experimental-Vulkan

[3] https://github.com/ollama/ollama/releases/tag/v0.12.11



Volcanoes have a grandeur that is grim
And earthquakes only terrify the dolts,
And to him who's scientific
There is nothing that's terrific
In the pattern of a flight of thunderbolts!
-- W. S. Gilbert, "The Mikado"