This isn't a repo, its just how my Linux workstation is built. My setup was the following: vLLM Docker container - for easy deployment and parallel inference. Claude Code - vibecoding and Agent Teams orchestration. Points at vLLM localhost endpoint instead of a cloud provider. gpt-oss:120