Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I've been using it for a few years on Gentoo. There were challenges with Python 2 years ago, but over the past year it's stabilized and I can even do img2video which is the most difficult local inference task so far.

Performance-wise, the 7900 xtx is still the most cost effective way of getting 24 gigabytes that isn't a sketchy VRAM mod. And VRAM is the main performance barrier since any LLM is going to barely fit in memory.

Highly suggest checking out TheRock. There's been a big rearchitecting of ROCm to improve the UX/quality.





Bought a Radeon r9700. 32GB vram and it does a good job.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: