LLMs are awesome but what about running other stuff locally? While I typically need 3b+ parameters to do something useful with an LLM there are a number of other use cases such as stt, tts, embeddings, etc. What are people running or would like to run locally outside of text generation? I am working