CUDA works, ROCm doesn't work well. Very few people want to run stable diffusion inference, fine tune LLaMA, train a large foundation model on AMD cards.
OpenAI has put in some work on Triton, Modular is working on Mojo, and tiny corp is working on their alternative.
Until some of those alternatives work as well as CUDA, people will mostly choose to buy Nvidia cards.
The monopoly is under attack from multiple angles, but they'll be able to print some good cash in the (potentially long) meantime.
Oh, and still significant supply shortages at many cloud providers. And now Nvidia's making more moves to renting GPUs directly. It'll be interesting to see how long it takes them to be able to have their supply meet demand.
https://geohot.github.io//blog/jekyll/update/2023/05/24/the-...
CUDA works, ROCm doesn't work well. Very few people want to run stable diffusion inference, fine tune LLaMA, train a large foundation model on AMD cards.
OpenAI has put in some work on Triton, Modular is working on Mojo, and tiny corp is working on their alternative.
Until some of those alternatives work as well as CUDA, people will mostly choose to buy Nvidia cards.
The monopoly is under attack from multiple angles, but they'll be able to print some good cash in the (potentially long) meantime.
Oh, and still significant supply shortages at many cloud providers. And now Nvidia's making more moves to renting GPUs directly. It'll be interesting to see how long it takes them to be able to have their supply meet demand.