Best AI News — Updated Every 3 Hours
Story Page
← All Stories
Home Community Story
Community

Has prompt processing taken a massive hit in llama.cpp for ROCm recently?

Via r/LocalLlama
Tuesday, Mar 24, 2026 · 8:32AM
Summary

ROCm Prefill Performance Drop on 7900XTX I've been looking to set up a dual 7900xtx system and recently put my Power Cooler Hellhound 7900xtx back into the machine to benchmark before PCIe splitting it with my Trio. Annoyingly, prompt processing on llama bench has dropped significantly while token g

Continue reading the full article
Read at r/LocalLlama
www.reddit.com
Back to all stories