I run ollama on my laptop in a VM with open web UI. It works great and I have plenty of models to choose from.
I recently was playing around with TTS and it is pretty solid as well. I am thinking about taking the smaller phi models and throwing it onto my pine64 quartz64 for a portable AI assistant while traveling. The only potential problem is the time it takes to process.