Comment by numpad0
fyi: ROCm support status currently isn't crucial for casual AI users - standard proprietary AMD drivers include Vulkan API support going back ~10 years. It's slower, but llama.cpp supports it, and so do many oneclick automagic LLM apps like LM Studio.