![](https://kyu.de/pictrs/image/aefdbba8-1697-4a52-8c4f-e0c2c624b576.png)
![](https://voyager.lemmy.ml/api/v3/image_proxy?url=https%3A%2F%2Fsh.itjust.works%2Fpictrs%2Fimage%2F045a2049-eb61-4960-88ba-97e7f1ffbf31.jpeg)
For AI and compute… They’re far behind. CUDA just wins. I hope a joint standard will be coming up soon, but until then Nvidia wins
I got a W6800 recently. I know a nvidia model of the same generation would be faster for AI - but that thing is fast enough to run stable diffusion variants with high resolution pictures locally without getting too annoyed.
Admittedly I’m just toying around for entertainment purposes - but I didn’t really have any problems of getting anything I wanted to try out with rocm support. Bigger annoyance was different projects targetting specific distributions or specific software versions (mostly ancient python), but as I’m doing everything in containers anyway that also was manageable.