Vulkan + Mesa Drivers For AI Inferencing? It's Already Showing Potential On Radeon RADV
- Reference: 0001563801
- News link: https://www.phoronix.com/news/Mesa-Vulkan-AI-RADV-Potential
- Source link:
David Airlie shared a blog post today outlining his experiences exploring the Mesa Vulkan drivers for AI inferencing. At the same time he and others like Karol Herbst of Red Hat are working on addressing feature gaps in the Mesa Vulkan drivers to make then more suitable for handling AI workloads.
Using the ramalama wrapper to Llama.cpp, Airlie has tested the different open-source and closed-source driver options. With the Mesa NVK driver the performance is a lot slower still than the official NVIDIA closed-source driver stack. No real surprise, especially given our recent NVK graphics benchmarks in [2]Mesa 25.2 NVK vs. NVIDIA R575 Linux Graphics Performance For GeForce RTX 40 Series .
On the Intel side he was able to get the Vulkan AI inferencing working with the ANV driver but unable to get the oneAPI/SYCL stack working nicely. On the AMD side is where it currently shows the most potential. Airlie's results show that the RADV Vulkan performance with ramalama/llama.cpp for token generation could be faster than using the official AMD ROCm compute stack. With prompt processing is where ROCm came out ahead, at least for now but there is hope some Mesa/RADV optimizations could put it ahead of ROCm.
Here is the data shared from Airlie with his GPU driver/hardware comparison:
Read more about his Vulkan AI comparison adventure via [3]his blog . It will be interesting to run some benchmarks on our side once the Mesa Vulkan drivers further mature for AI workloads.
[1] https://www.phoronix.com/news/NVIDIA-Vulkan-AI-ML-Success
[2] https://www.phoronix.com/review/mesa-252-nvk-nvidia
[3] https://airlied.blogspot.com/2025/07/ramalamamesa-benchmarks-on-my-hardware.html
shmerl