Ollama is now available
-
Ollama is now packaged as a standalone app.
The app package is as usual still published as unstable until we get some feedback how it works and iron out any initial issues.
Forum category is at https://forum.cloudron.io/category/212/ollama so please report any issues there
The docs will be at https://docs.cloudron.io/packages/ollama/
Package repo is https://git.cloudron.io/packages/ollama-app -
Firstly, this is wonderful news. I see that your Ollama package appears to include an authentication capability, which is something that vanilla Ollama is missing. When coupled with Cloudron's app management features, that makes it extremely helpful.
However... Without GPU support or crazy fast CPUs and a big load of fast RAM, isn't Ollama self hosting almost useless?
Note - Here's an old thread where we discussed some related topics.
-
Hello @abuyuy
Currently, the Ollama app does not include the capability to access vaapi devices (gpus).
We can add this capability to the Ollama app but how to configure it needs to be researched then.@james In our case (we're have some old enterprise Nvidia GPUs), we found that installing the Nvidia drivers and CUDA toolkit is ultimately not difficult once you know what to do. Our problem with making that useful on Cloudron (per this old thread) was that we couldn't start Cloudron's OpenWebUI package with GPU support. So even though we had our GPU hardware working on a Cloudron server, we could never use it with a Cloudron instance of OpenWebUI/Ollama.
I suspect the process of installing the Nvidia drivers and CUDA toolkit is generic enough that it could be supported by Cloudron somehow. However we don't have any GPUs from other enterprise brands (e.g. AMD, Intel) so we never tried that, and I imagine supporting drivers for consumer GPUs is an impossible maze of complexity.
We would LOVE it if you could support the Nvidia GPU use case as a basic option.
-
Firstly, this is wonderful news. I see that your Ollama package appears to include an authentication capability, which is something that vanilla Ollama is missing. When coupled with Cloudron's app management features, that makes it extremely helpful.
However... Without GPU support or crazy fast CPUs and a big load of fast RAM, isn't Ollama self hosting almost useless?
Note - Here's an old thread where we discussed some related topics.
