Cloudron makes it easy to run web apps like WordPress, Nextcloud, GitLab on your server. Find out more or install now.


Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Bookmarks
  • Search
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

Cloudron Forum

Apps | Demo | Docs | Install
  1. Cloudron Forum
  2. OpenWebUI
  3. Should ollama be part of this app package?

Should ollama be part of this app package?

Scheduled Pinned Locked Moved OpenWebUI
19 Posts 7 Posters 4.4k Views 7 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • girishG girish

    @timconsidine I tried it locally on my low end netcup, it works atleast. Very slowly but works.

    timconsidineT Offline
    timconsidineT Offline
    timconsidine
    App Dev
    wrote on last edited by
    #5

    @girish great !
    I know all the big engines flaunt their speed, but for me it's not the primary issue.
    Within reason of course.
    How did you install ollama on netcup ? manually ? docker run ? docker-compose ?
    I'm looking for a decent docker-compose.yaml as I prefer that for my non-cloudron docker apps (using traefik which I have got sussed, mostly).

    girishG 1 Reply Last reply
    1
    • KubernetesK Offline
      KubernetesK Offline
      Kubernetes
      App Dev
      wrote on last edited by
      #6

      At least as long as the App throws errors because Ollama is unavailable - but OpenAI could be still used - it should be included in the package.

      1 Reply Last reply
      3
      • timconsidineT timconsidine

        @girish great !
        I know all the big engines flaunt their speed, but for me it's not the primary issue.
        Within reason of course.
        How did you install ollama on netcup ? manually ? docker run ? docker-compose ?
        I'm looking for a decent docker-compose.yaml as I prefer that for my non-cloudron docker apps (using traefik which I have got sussed, mostly).

        girishG Offline
        girishG Offline
        girish
        Staff
        wrote on last edited by
        #7

        @timconsidine said in Should ollama be part of this app package?:

        How did you install ollama on netcup ? manually ? docker run ? docker-compose ?

        ollama is just a single go binary. I just downloaded it to /app/data and chmod +x and ran it 🙂 you can see the openwebui-app repo where it is now integrated.

        L 1 Reply Last reply
        3
        • girishG girish

          Was wondering if people had thoughts on this. Should https://github.com/ollama/ollama be packaged as part of this app? Maybe this can be an optional component.

          The models have some heavy requirements: You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models.

          Also, Cloudron does not have GPU support for containers yet, so I don't know how well these work.

          L Offline
          L Offline
          LoudLemur
          wrote on last edited by LoudLemur
          #8

          @girish said in Should ollama be part of this app package?:

          You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models.

          This is ordinary RAM, right, not VRAM? SSDnodes is great for lavish amounts of RAM.

          We would have to go into the application's settings in Cloudron and increase the resource allocation to e.g. 16GB?

          1 Reply Last reply
          1
          • girishG girish

            @timconsidine said in Should ollama be part of this app package?:

            How did you install ollama on netcup ? manually ? docker run ? docker-compose ?

            ollama is just a single go binary. I just downloaded it to /app/data and chmod +x and ran it 🙂 you can see the openwebui-app repo where it is now integrated.

            L Offline
            L Offline
            LoudLemur
            wrote on last edited by
            #9

            @girish I would much rather have ollama as part of the app package than having to do that!

            One further suggestion: I think it should ship with a small but functioning model and prompt.

            On first run, "nothing is there", which is nice for people who are used to having to configure everything, but it is not so encouraging if you are just getting started. You have to "leave" the instance and go to the openwebui website and then try and find a model and prompt.

            I guess most people won't want that, so I think we should anyway include a step by step in the documentation on how to get find and install a model and prompt that would work.

            Also, I think the documentation could benefit from your above guidelines on what size model to use depending on your available RAM.

            girishG 1 Reply Last reply
            1
            • girishG Offline
              girishG Offline
              girish
              Staff
              wrote on last edited by girish
              #10

              Local ollama is now integrated. You have to reinstall the app though.

              Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

              coniunctioC 1 Reply Last reply
              3
              • L LoudLemur

                @girish I would much rather have ollama as part of the app package than having to do that!

                One further suggestion: I think it should ship with a small but functioning model and prompt.

                On first run, "nothing is there", which is nice for people who are used to having to configure everything, but it is not so encouraging if you are just getting started. You have to "leave" the instance and go to the openwebui website and then try and find a model and prompt.

                I guess most people won't want that, so I think we should anyway include a step by step in the documentation on how to get find and install a model and prompt that would work.

                Also, I think the documentation could benefit from your above guidelines on what size model to use depending on your available RAM.

                girishG Offline
                girishG Offline
                girish
                Staff
                wrote on last edited by
                #11

                @LoudLemur said in Should ollama be part of this app package?:

                One further suggestion: I think it should ship with a small but functioning model and prompt.

                Now with the local ollama integration, you can download whichever models you want from the UI itself. The models are quite big so pre-installing them is not an option.

                KubernetesK 1 Reply Last reply
                6
                • girishG girish

                  @LoudLemur said in Should ollama be part of this app package?:

                  One further suggestion: I think it should ship with a small but functioning model and prompt.

                  Now with the local ollama integration, you can download whichever models you want from the UI itself. The models are quite big so pre-installing them is not an option.

                  KubernetesK Offline
                  KubernetesK Offline
                  Kubernetes
                  App Dev
                  wrote on last edited by
                  #12

                  @girish great job!

                  one question, is it possible to enable external use of the ollama api ?

                  coniunctioC 1 Reply Last reply
                  2
                  • girishG Offline
                    girishG Offline
                    girish
                    Staff
                    wrote on last edited by
                    #13

                    @Kubernetes yes, https://docs.cloudron.io/apps/openwebui/#ollama

                    KubernetesK 1 Reply Last reply
                    2
                    • girishG girish

                      @Kubernetes yes, https://docs.cloudron.io/apps/openwebui/#ollama

                      KubernetesK Offline
                      KubernetesK Offline
                      Kubernetes
                      App Dev
                      wrote on last edited by
                      #14

                      @girish ah, amazing, was not aware that the documentation has also been updated already. thanks!

                      1 Reply Last reply
                      2
                      • girishG Offline
                        girishG Offline
                        girish
                        Staff
                        wrote on last edited by
                        #15

                        I learnt a lite version might be appearing soon (which contains just the chat frontend) - https://github.com/open-webui/open-webui/discussions/260

                        1 Reply Last reply
                        3
                        • girishG girish forked this topic on
                        • KubernetesK Kubernetes

                          @girish great job!

                          one question, is it possible to enable external use of the ollama api ?

                          coniunctioC Offline
                          coniunctioC Offline
                          coniunctio
                          wrote on last edited by coniunctio
                          #16

                          @Kubernetes I am not sure if this is what you are asking, but I am currently running Ollama separately via docker with a dedicated OLD GPU (8GB NVIDIA) on my NAS (working shockingly good on 7B/11B GGUFs and moderately good on 13B ones) and Cloudron on a VM on the same NAS. I use Ollama externally (technically it's still local on the machine's hardware, but is configured as though it is not) and deactivated this app's localhost Ollama. This can be done by going into the Cloudron's Open-Webui File Manager through the settings and configuring "env.sh" -

                          # Change this to false to disable local ollama and use your own
                          export LOCAL_OLLAMA_ENABLED=false
                          
                          # When using remote ollama, change this to the ollama's base url
                          export OLLAMA_API_BASE_URL="http://changethis:11434"
                          
                          # When local ollama is enabled, this is location for the downloaded models.
                          # If the path is under /app/data, models will be backed up. Note that models
                          # can be very large. To skip backup of models, move the models to a volume (https://docs.cloudron.io/volumes/)
                          # export OLLAMA_MODELS=/app/data/ollama-home/models
                          

                          Is this what you are referring to?

                          KubernetesK 1 Reply Last reply
                          0
                          • coniunctioC coniunctio

                            @Kubernetes I am not sure if this is what you are asking, but I am currently running Ollama separately via docker with a dedicated OLD GPU (8GB NVIDIA) on my NAS (working shockingly good on 7B/11B GGUFs and moderately good on 13B ones) and Cloudron on a VM on the same NAS. I use Ollama externally (technically it's still local on the machine's hardware, but is configured as though it is not) and deactivated this app's localhost Ollama. This can be done by going into the Cloudron's Open-Webui File Manager through the settings and configuring "env.sh" -

                            # Change this to false to disable local ollama and use your own
                            export LOCAL_OLLAMA_ENABLED=false
                            
                            # When using remote ollama, change this to the ollama's base url
                            export OLLAMA_API_BASE_URL="http://changethis:11434"
                            
                            # When local ollama is enabled, this is location for the downloaded models.
                            # If the path is under /app/data, models will be backed up. Note that models
                            # can be very large. To skip backup of models, move the models to a volume (https://docs.cloudron.io/volumes/)
                            # export OLLAMA_MODELS=/app/data/ollama-home/models
                            

                            Is this what you are referring to?

                            KubernetesK Offline
                            KubernetesK Offline
                            Kubernetes
                            App Dev
                            wrote on last edited by
                            #17

                            @coniunctio Yes, exactly this was what I was referring to. Thank you for bringing this example up.

                            1 Reply Last reply
                            0
                            • girishG girish

                              Local ollama is now integrated. You have to reinstall the app though.

                              Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

                              coniunctioC Offline
                              coniunctioC Offline
                              coniunctio
                              wrote on last edited by
                              #18

                              @girish said in Should ollama be part of this app package?:

                              Local ollama is now integrated. You have to reinstall the app though.

                              Keep your expectations in check. It probably won't work great if you don't have a good CPU and we have no GPU integration yet. It's very slow with low end CPUs. I am not an expert on the RAM/CPU/GPU requirements. Feel free to experiment.

                              I have been using the workaround of disabling local Ollama with the Cloudron app and running a separate (external) docker container installation of Ollama with a dedicated GPU on the same hardware and then linking that instance of Ollama to the Cloudron instance of Open-WebUI. Somehow, this configuration is faster on a NAS purchased in 2018 with an add-on NVIDIA 8GB GPU than my M1 MacBook Pro with 16GB RAM and integrated GPU purchased more recently. The additional bonus of running the Cloudron Open-WebUI vs the localhost version on my Apple silicon MBP is that I can use my local LLMs on my mobile devices in transit when my laptop is shut down.

                              1 Reply Last reply
                              1
                              • JOduMonTJ Offline
                                JOduMonTJ Offline
                                JOduMonT
                                wrote on last edited by
                                #19

                                personally I disabled the ollama local, because my Cloudron doesnt have GPU and on CPU it is too painfull.
                                in exchange; I activated a bunch of Providers API compatible with OpenAI
                                but at the end I realized that I just need OpenRouter to access all of them.
                                image.png

                                with OpenRouter, you could even block providers that logs your queries;
                                which I will Feature Request for Open-WebUI
                                image.png

                                1 Reply Last reply
                                3
                                Reply
                                • Reply as topic
                                Log in to reply
                                • Oldest to Newest
                                • Newest to Oldest
                                • Most Votes


                                • Login

                                • Don't have an account? Register

                                • Login or register to search.
                                • First post
                                  Last post
                                0
                                • Categories
                                • Recent
                                • Tags
                                • Popular
                                • Bookmarks
                                • Search