Cloudron makes it easy to run web apps like WordPress, Nextcloud, GitLab on your server. Find out more or install now.


Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Bookmarks
  • Search
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

Cloudron Forum

Apps | Demo | Docs | Install
  1. Cloudron Forum
  2. OpenWebUI
  3. Should ollama be part of this app package?

Should ollama be part of this app package?

Scheduled Pinned Locked Moved OpenWebUI
19 Posts 7 Posters 4.4k Views 7 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • KubernetesK Offline
    KubernetesK Offline
    Kubernetes
    App Dev
    wrote on last edited by
    #6

    At least as long as the App throws errors because Ollama is unavailable - but OpenAI could be still used - it should be included in the package.

    1 Reply Last reply
    3
    • timconsidineT timconsidine

      @girish great !
      I know all the big engines flaunt their speed, but for me it's not the primary issue.
      Within reason of course.
      How did you install ollama on netcup ? manually ? docker run ? docker-compose ?
      I'm looking for a decent docker-compose.yaml as I prefer that for my non-cloudron docker apps (using traefik which I have got sussed, mostly).

      girishG Offline
      girishG Offline
      girish
      Staff
      wrote on last edited by
      #7

      @timconsidine said in Should ollama be part of this app package?:

      How did you install ollama on netcup ? manually ? docker run ? docker-compose ?

      ollama is just a single go binary. I just downloaded it to /app/data and chmod +x and ran it 🙂 you can see the openwebui-app repo where it is now integrated.

      L 1 Reply Last reply
      3
      • girishG girish

        Was wondering if people had thoughts on this. Should https://github.com/ollama/ollama be packaged as part of this app? Maybe this can be an optional component.

        The models have some heavy requirements: You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models.

        Also, Cloudron does not have GPU support for containers yet, so I don't know how well these work.

        L Offline
        L Offline
        LoudLemur
        wrote on last edited by LoudLemur
        #8

        @girish said in Should ollama be part of this app package?:

        You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models.

        This is ordinary RAM, right, not VRAM? SSDnodes is great for lavish amounts of RAM.

        We would have to go into the application's settings in Cloudron and increase the resource allocation to e.g. 16GB?

        1 Reply Last reply
        1
        • girishG girish

          @timconsidine said in Should ollama be part of this app package?:

          How did you install ollama on netcup ? manually ? docker run ? docker-compose ?

          ollama is just a single go binary. I just downloaded it to /app/data and chmod +x and ran it 🙂 you can see the openwebui-app repo where it is now integrated.

          L Offline
          L Offline
          LoudLemur
          wrote on last edited by
          #9

          @girish I would much rather have ollama as part of the app package than having to do that!

          One further suggestion: I think it should ship with a small but functioning model and prompt.

          On first run, "nothing is there", which is nice for people who are used to having to configure everything, but it is not so encouraging if you are just getting started. You have to "leave" the instance and go to the openwebui website and then try and find a model and prompt.

          I guess most people won't want that, so I think we should anyway include a step by step in the documentation on how to get find and install a model and prompt that would work.

          Also, I think the documentation could benefit from your above guidelines on what size model to use depending on your available RAM.

          girishG 1 Reply Last reply
          1
          • girishG Offline
            girishG Offline
            girish
            Staff
            wrote on last edited by girish
            #10

            Local ollama is now integrated. You have to reinstall the app though.

            Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

            coniunctioC 1 Reply Last reply
            3
            • L LoudLemur

              @girish I would much rather have ollama as part of the app package than having to do that!

              One further suggestion: I think it should ship with a small but functioning model and prompt.

              On first run, "nothing is there", which is nice for people who are used to having to configure everything, but it is not so encouraging if you are just getting started. You have to "leave" the instance and go to the openwebui website and then try and find a model and prompt.

              I guess most people won't want that, so I think we should anyway include a step by step in the documentation on how to get find and install a model and prompt that would work.

              Also, I think the documentation could benefit from your above guidelines on what size model to use depending on your available RAM.

              girishG Offline
              girishG Offline
              girish
              Staff
              wrote on last edited by
              #11

              @LoudLemur said in Should ollama be part of this app package?:

              One further suggestion: I think it should ship with a small but functioning model and prompt.

              Now with the local ollama integration, you can download whichever models you want from the UI itself. The models are quite big so pre-installing them is not an option.

              KubernetesK 1 Reply Last reply
              6
              • girishG girish

                @LoudLemur said in Should ollama be part of this app package?:

                One further suggestion: I think it should ship with a small but functioning model and prompt.

                Now with the local ollama integration, you can download whichever models you want from the UI itself. The models are quite big so pre-installing them is not an option.

                KubernetesK Offline
                KubernetesK Offline
                Kubernetes
                App Dev
                wrote on last edited by
                #12

                @girish great job!

                one question, is it possible to enable external use of the ollama api ?

                coniunctioC 1 Reply Last reply
                2
                • girishG Offline
                  girishG Offline
                  girish
                  Staff
                  wrote on last edited by
                  #13

                  @Kubernetes yes, https://docs.cloudron.io/apps/openwebui/#ollama

                  KubernetesK 1 Reply Last reply
                  2
                  • girishG girish

                    @Kubernetes yes, https://docs.cloudron.io/apps/openwebui/#ollama

                    KubernetesK Offline
                    KubernetesK Offline
                    Kubernetes
                    App Dev
                    wrote on last edited by
                    #14

                    @girish ah, amazing, was not aware that the documentation has also been updated already. thanks!

                    1 Reply Last reply
                    2
                    • girishG Offline
                      girishG Offline
                      girish
                      Staff
                      wrote on last edited by
                      #15

                      I learnt a lite version might be appearing soon (which contains just the chat frontend) - https://github.com/open-webui/open-webui/discussions/260

                      1 Reply Last reply
                      3
                      • girishG girish forked this topic on
                      • KubernetesK Kubernetes

                        @girish great job!

                        one question, is it possible to enable external use of the ollama api ?

                        coniunctioC Offline
                        coniunctioC Offline
                        coniunctio
                        wrote on last edited by coniunctio
                        #16

                        @Kubernetes I am not sure if this is what you are asking, but I am currently running Ollama separately via docker with a dedicated OLD GPU (8GB NVIDIA) on my NAS (working shockingly good on 7B/11B GGUFs and moderately good on 13B ones) and Cloudron on a VM on the same NAS. I use Ollama externally (technically it's still local on the machine's hardware, but is configured as though it is not) and deactivated this app's localhost Ollama. This can be done by going into the Cloudron's Open-Webui File Manager through the settings and configuring "env.sh" -

                        # Change this to false to disable local ollama and use your own
                        export LOCAL_OLLAMA_ENABLED=false
                        
                        # When using remote ollama, change this to the ollama's base url
                        export OLLAMA_API_BASE_URL="http://changethis:11434"
                        
                        # When local ollama is enabled, this is location for the downloaded models.
                        # If the path is under /app/data, models will be backed up. Note that models
                        # can be very large. To skip backup of models, move the models to a volume (https://docs.cloudron.io/volumes/)
                        # export OLLAMA_MODELS=/app/data/ollama-home/models
                        

                        Is this what you are referring to?

                        KubernetesK 1 Reply Last reply
                        0
                        • coniunctioC coniunctio

                          @Kubernetes I am not sure if this is what you are asking, but I am currently running Ollama separately via docker with a dedicated OLD GPU (8GB NVIDIA) on my NAS (working shockingly good on 7B/11B GGUFs and moderately good on 13B ones) and Cloudron on a VM on the same NAS. I use Ollama externally (technically it's still local on the machine's hardware, but is configured as though it is not) and deactivated this app's localhost Ollama. This can be done by going into the Cloudron's Open-Webui File Manager through the settings and configuring "env.sh" -

                          # Change this to false to disable local ollama and use your own
                          export LOCAL_OLLAMA_ENABLED=false
                          
                          # When using remote ollama, change this to the ollama's base url
                          export OLLAMA_API_BASE_URL="http://changethis:11434"
                          
                          # When local ollama is enabled, this is location for the downloaded models.
                          # If the path is under /app/data, models will be backed up. Note that models
                          # can be very large. To skip backup of models, move the models to a volume (https://docs.cloudron.io/volumes/)
                          # export OLLAMA_MODELS=/app/data/ollama-home/models
                          

                          Is this what you are referring to?

                          KubernetesK Offline
                          KubernetesK Offline
                          Kubernetes
                          App Dev
                          wrote on last edited by
                          #17

                          @coniunctio Yes, exactly this was what I was referring to. Thank you for bringing this example up.

                          1 Reply Last reply
                          0
                          • girishG girish

                            Local ollama is now integrated. You have to reinstall the app though.

                            Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

                            coniunctioC Offline
                            coniunctioC Offline
                            coniunctio
                            wrote on last edited by
                            #18

                            @girish said in Should ollama be part of this app package?:

                            Local ollama is now integrated. You have to reinstall the app though.

                            Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

                            I have been using the workaround of disabling local Ollama with the Cloudron app and running a separate (external) docker container installation of Ollama with a dedicated GPU on the same hardware and then linking that instance of Ollama to the Cloudron instance of Open-WebUI. Somehow, this configuration is faster on a NAS purchased in 2018 with an add-on NVIDIA 8GB GPU than my M1 MacBook Pro with 16GB RAM and integrated GPU purchased more recently. The additional bonus of running the Cloudron Open-WebUI vs the localhost version on my Apple silicon MBP is that I can use my local LLMs on my mobile devices in transit when my laptop is shut down.

                            1 Reply Last reply
                            1
                            • JOduMonTJ Offline
                              JOduMonTJ Offline
                              JOduMonT
                              wrote on last edited by
                              #19

                              personally I disabled the ollama local, because my Cloudron doesnt have GPU and on CPU it is too painfull.
                              in exchange; I activated a bunch of Providers API compatible with OpenAI
                              but at the end I realized that I just need OpenRouter to access all of them.
                              image.png

                              with OpenRouter, you could even block providers that logs your queries;
                              which I will Feature Request for Open-WebUI
                              image.png

                              1 Reply Last reply
                              3
                              Reply
                              • Reply as topic
                              Log in to reply
                              • Oldest to Newest
                              • Newest to Oldest
                              • Most Votes


                              • Login

                              • Don't have an account? Register

                              • Login or register to search.
                              • First post
                                Last post
                              0
                              • Categories
                              • Recent
                              • Tags
                              • Popular
                              • Bookmarks
                              • Search