Cloudron makes it easy to run web apps like WordPress, Nextcloud, GitLab on your server. Find out more or install now.


Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Bookmarks
  • Search
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

Cloudron Forum

Apps | Demo | Docs | Install
  1. Cloudron Forum
  2. OpenWebUI
  3. Should ollama be part of this app package?

Should ollama be part of this app package?

Scheduled Pinned Locked Moved OpenWebUI
19 Posts 7 Posters 4.4k Views 7 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • girishG Offline
    girishG Offline
    girish
    Staff
    wrote on last edited by
    #4

    @timconsidine I tried it locally on my low end netcup, it works atleast. Very slowly but works.

    timconsidineT 1 Reply Last reply
    5
    • girishG girish

      @timconsidine I tried it locally on my low end netcup, it works atleast. Very slowly but works.

      timconsidineT Offline
      timconsidineT Offline
      timconsidine
      App Dev
      wrote on last edited by
      #5

      @girish great !
      I know all the big engines flaunt their speed, but for me it's not the primary issue.
      Within reason of course.
      How did you install ollama on netcup ? manually ? docker run ? docker-compose ?
      I'm looking for a decent docker-compose.yaml as I prefer that for my non-cloudron docker apps (using traefik which I have got sussed, mostly).

      girishG 1 Reply Last reply
      1
      • KubernetesK Offline
        KubernetesK Offline
        Kubernetes
        App Dev
        wrote on last edited by
        #6

        At least as long as the App throws errors because Ollama is unavailable - but OpenAI could be still used - it should be included in the package.

        1 Reply Last reply
        3
        • timconsidineT timconsidine

          @girish great !
          I know all the big engines flaunt their speed, but for me it's not the primary issue.
          Within reason of course.
          How did you install ollama on netcup ? manually ? docker run ? docker-compose ?
          I'm looking for a decent docker-compose.yaml as I prefer that for my non-cloudron docker apps (using traefik which I have got sussed, mostly).

          girishG Offline
          girishG Offline
          girish
          Staff
          wrote on last edited by
          #7

          @timconsidine said in Should ollama be part of this app package?:

          How did you install ollama on netcup ? manually ? docker run ? docker-compose ?

          ollama is just a single go binary. I just downloaded it to /app/data and chmod +x and ran it 🙂 you can see the openwebui-app repo where it is now integrated.

          L 1 Reply Last reply
          3
          • girishG girish

            Was wondering if people had thoughts on this. Should https://github.com/ollama/ollama be packaged as part of this app? Maybe this can be an optional component.

            The models have some heavy requirements: You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models.

            Also, Cloudron does not have GPU support for containers yet, so I don't know how well these work.

            L Offline
            L Offline
            LoudLemur
            wrote on last edited by LoudLemur
            #8

            @girish said in Should ollama be part of this app package?:

            You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models.

            This is ordinary RAM, right, not VRAM? SSDnodes is great for lavish amounts of RAM.

            We would have to go into the application's settings in Cloudron and increase the resource allocation to e.g. 16GB?

            1 Reply Last reply
            1
            • girishG girish

              @timconsidine said in Should ollama be part of this app package?:

              How did you install ollama on netcup ? manually ? docker run ? docker-compose ?

              ollama is just a single go binary. I just downloaded it to /app/data and chmod +x and ran it 🙂 you can see the openwebui-app repo where it is now integrated.

              L Offline
              L Offline
              LoudLemur
              wrote on last edited by
              #9

              @girish I would much rather have ollama as part of the app package than having to do that!

              One further suggestion: I think it should ship with a small but functioning model and prompt.

              On first run, "nothing is there", which is nice for people who are used to having to configure everything, but it is not so encouraging if you are just getting started. You have to "leave" the instance and go to the openwebui website and then try and find a model and prompt.

              I guess most people won't want that, so I think we should anyway include a step by step in the documentation on how to get find and install a model and prompt that would work.

              Also, I think the documentation could benefit from your above guidelines on what size model to use depending on your available RAM.

              girishG 1 Reply Last reply
              1
              • girishG Offline
                girishG Offline
                girish
                Staff
                wrote on last edited by girish
                #10

                Local ollama is now integrated. You have to reinstall the app though.

                Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

                coniunctioC 1 Reply Last reply
                3
                • L LoudLemur

                  @girish I would much rather have ollama as part of the app package than having to do that!

                  One further suggestion: I think it should ship with a small but functioning model and prompt.

                  On first run, "nothing is there", which is nice for people who are used to having to configure everything, but it is not so encouraging if you are just getting started. You have to "leave" the instance and go to the openwebui website and then try and find a model and prompt.

                  I guess most people won't want that, so I think we should anyway include a step by step in the documentation on how to get find and install a model and prompt that would work.

                  Also, I think the documentation could benefit from your above guidelines on what size model to use depending on your available RAM.

                  girishG Offline
                  girishG Offline
                  girish
                  Staff
                  wrote on last edited by
                  #11

                  @LoudLemur said in Should ollama be part of this app package?:

                  One further suggestion: I think it should ship with a small but functioning model and prompt.

                  Now with the local ollama integration, you can download whichever models you want from the UI itself. The models are quite big so pre-installing them is not an option.

                  KubernetesK 1 Reply Last reply
                  6
                  • girishG girish

                    @LoudLemur said in Should ollama be part of this app package?:

                    One further suggestion: I think it should ship with a small but functioning model and prompt.

                    Now with the local ollama integration, you can download whichever models you want from the UI itself. The models are quite big so pre-installing them is not an option.

                    KubernetesK Offline
                    KubernetesK Offline
                    Kubernetes
                    App Dev
                    wrote on last edited by
                    #12

                    @girish great job!

                    one question, is it possible to enable external use of the ollama api ?

                    coniunctioC 1 Reply Last reply
                    2
                    • girishG Offline
                      girishG Offline
                      girish
                      Staff
                      wrote on last edited by
                      #13

                      @Kubernetes yes, https://docs.cloudron.io/apps/openwebui/#ollama

                      KubernetesK 1 Reply Last reply
                      2
                      • girishG girish

                        @Kubernetes yes, https://docs.cloudron.io/apps/openwebui/#ollama

                        KubernetesK Offline
                        KubernetesK Offline
                        Kubernetes
                        App Dev
                        wrote on last edited by
                        #14

                        @girish ah, amazing, was not aware that the documentation has also been updated already. thanks!

                        1 Reply Last reply
                        2
                        • girishG Offline
                          girishG Offline
                          girish
                          Staff
                          wrote on last edited by
                          #15

                          I learnt a lite version might be appearing soon (which contains just the chat frontend) - https://github.com/open-webui/open-webui/discussions/260

                          1 Reply Last reply
                          3
                          • girishG girish forked this topic on
                          • KubernetesK Kubernetes

                            @girish great job!

                            one question, is it possible to enable external use of the ollama api ?

                            coniunctioC Offline
                            coniunctioC Offline
                            coniunctio
                            wrote on last edited by coniunctio
                            #16

                            @Kubernetes I am not sure if this is what you are asking, but I am currently running Ollama separately via docker with a dedicated OLD GPU (8GB NVIDIA) on my NAS (working shockingly good on 7B/11B GGUFs and moderately good on 13B ones) and Cloudron on a VM on the same NAS. I use Ollama externally (technically it's still local on the machine's hardware, but is configured as though it is not) and deactivated this app's localhost Ollama. This can be done by going into the Cloudron's Open-Webui File Manager through the settings and configuring "env.sh" -

                            # Change this to false to disable local ollama and use your own
                            export LOCAL_OLLAMA_ENABLED=false
                            
                            # When using remote ollama, change this to the ollama's base url
                            export OLLAMA_API_BASE_URL="http://changethis:11434"
                            
                            # When local ollama is enabled, this is location for the downloaded models.
                            # If the path is under /app/data, models will be backed up. Note that models
                            # can be very large. To skip backup of models, move the models to a volume (https://docs.cloudron.io/volumes/)
                            # export OLLAMA_MODELS=/app/data/ollama-home/models
                            

                            Is this what you are referring to?

                            KubernetesK 1 Reply Last reply
                            0
                            • coniunctioC coniunctio

                              @Kubernetes I am not sure if this is what you are asking, but I am currently running Ollama separately via docker with a dedicated OLD GPU (8GB NVIDIA) on my NAS (working shockingly good on 7B/11B GGUFs and moderately good on 13B ones) and Cloudron on a VM on the same NAS. I use Ollama externally (technically it's still local on the machine's hardware, but is configured as though it is not) and deactivated this app's localhost Ollama. This can be done by going into the Cloudron's Open-Webui File Manager through the settings and configuring "env.sh" -

                              # Change this to false to disable local ollama and use your own
                              export LOCAL_OLLAMA_ENABLED=false
                              
                              # When using remote ollama, change this to the ollama's base url
                              export OLLAMA_API_BASE_URL="http://changethis:11434"
                              
                              # When local ollama is enabled, this is location for the downloaded models.
                              # If the path is under /app/data, models will be backed up. Note that models
                              # can be very large. To skip backup of models, move the models to a volume (https://docs.cloudron.io/volumes/)
                              # export OLLAMA_MODELS=/app/data/ollama-home/models
                              

                              Is this what you are referring to?

                              KubernetesK Offline
                              KubernetesK Offline
                              Kubernetes
                              App Dev
                              wrote on last edited by
                              #17

                              @coniunctio Yes, exactly this was what I was referring to. Thank you for bringing this example up.

                              1 Reply Last reply
                              0
                              • girishG girish

                                Local ollama is now integrated. You have to reinstall the app though.

                                Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

                                coniunctioC Offline
                                coniunctioC Offline
                                coniunctio
                                wrote on last edited by
                                #18

                                @girish said in Should ollama be part of this app package?:

                                Local ollama is now integrated. You have to reinstall the app though.

                                Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

                                I have been using the workaround of disabling local Ollama with the Cloudron app and running a separate (external) docker container installation of Ollama with a dedicated GPU on the same hardware and then linking that instance of Ollama to the Cloudron instance of Open-WebUI. Somehow, this configuration is faster on a NAS purchased in 2018 with an add-on NVIDIA 8GB GPU than my M1 MacBook Pro with 16GB RAM and integrated GPU purchased more recently. The additional bonus of running the Cloudron Open-WebUI vs the localhost version on my Apple silicon MBP is that I can use my local LLMs on my mobile devices in transit when my laptop is shut down.

                                1 Reply Last reply
                                1
                                • JOduMonTJ Offline
                                  JOduMonTJ Offline
                                  JOduMonT
                                  wrote on last edited by
                                  #19

                                  personally I disabled the ollama local, because my Cloudron doesnt have GPU and on CPU it is too painfull.
                                  in exchange; I activated a bunch of Providers API compatible with OpenAI
                                  but at the end I realized that I just need OpenRouter to access all of them.
                                  image.png

                                  with OpenRouter, you could even block providers that logs your queries;
                                  which I will Feature Request for Open-WebUI
                                  image.png

                                  1 Reply Last reply
                                  3
                                  Reply
                                  • Reply as topic
                                  Log in to reply
                                  • Oldest to Newest
                                  • Newest to Oldest
                                  • Most Votes


                                  • Login

                                  • Don't have an account? Register

                                  • Login or register to search.
                                  • First post
                                    Last post
                                  0
                                  • Categories
                                  • Recent
                                  • Tags
                                  • Popular
                                  • Bookmarks
                                  • Search