https://github.com/ggml-org/llama.cpp/releases
I'm not sure I follow, what alternative to CUDA on Linux offers similar performance?
Nevertheless, worth looking at the Vulkan builds. They work on all GPUs!