Built an Ansible playbook to turn AMD Strix Halo machines into local AI inference servers Hey all, I've been running local LLMs on my Framework Desktop (AMD Strix Halo, 128 GB unified memory) and wanted a reproducible, one-command setup. So I packaged everything into an Ansible playbook and put