@girish ya, i think this is a good approach because in real use case, I think most people prefer to run inference models not in the same machine with openWebUI instance. OpenWebUI is good enough bundled with chromadb , sql lite or postgressql database and redis as one package and the inference models (including ollama) running in different servers.