Just got a great first impression running Open WebUI - took one command to get it up and running (uvx --python 3.11 open-webui serve) and it noticed I had Ollama running and automatically gave me a UI to interact with those models https://simonwillison.net/2024/Dec/27/open-webui/
Also got a reminder of how freakishly capable the tiny (1.9GB) Llama 3.2:3B model is at summarization and question answering against documents you feed into it via its impressive 128,000 token context