Cloudron makes it easy to run web apps like WordPress, Nextcloud, GitLab on your server. Find out more or install now.


Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Bookmarks
  • Search
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

Cloudron Forum

Apps | Demo | Docs | Install
  1. Cloudron Forum
  2. Support
  3. Since I added a storage volume to my infrastructure, I've been getting this backup error message very regularly

Since I added a storage volume to my infrastructure, I've been getting this backup error message very regularly

Scheduled Pinned Locked Moved Solved Support
backups
21 Posts 4 Posters 1.2k Views 5 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • D Dont-Worry

    I have something like 200gb of Emby data (and it's growing every day) you're right about that. On top we have a lot of music and a lot of youtube videos. Some files are in MP3 but the majority are in .WAV and for videos we're on .MP4.

    Do you have any idea what I should do to be able to restore the backups system?

    girishG Offline
    girishG Offline
    girish
    Staff
    wrote on last edited by
    #11

    @Dont-Worry It's failing pretty much at 100GB. It seems that the provider may not support > 100GB files. Which backup provider are you using (Is it OVH?) ?

    1 Reply Last reply
    1
    • D Dont-Worry

      I have something like 200gb of Emby data (and it's growing every day) you're right about that. On top we have a lot of music and a lot of youtube videos. Some files are in MP3 but the majority are in .WAV and for videos we're on .MP4.

      Do you have any idea what I should do to be able to restore the backups system?

      necrevistonnezrN Offline
      necrevistonnezrN Offline
      necrevistonnezr
      wrote on last edited by
      #12

      @Dont-Worry said in Since I added a storage volume to my infrastructure, I've been getting this backup error message very regularly:

      I have something like 200gb of Emby data (and it's growing every day) you're right about that. On top we have a lot of music and a lot of youtube videos. Some files are in MP3 but the majority are in .WAV and for videos we're on .MP4.

      Using the encrypted tar backend for this mass of - already compressed - data is asking for trouble, IMHO. As girish pointed out, you end up with a single file of huge proportions, from the docs:

      "The tgz format stores all the backup information in a single tarball whereas the rsync format stores all backup information as files inside a directory."

      I could be wrong but many B2C providers will probably have difficulties supporting a 200 GB + file?!

      1 Reply Last reply
      2
      • D Offline
        D Offline
        Dont-Worry
        wrote on last edited by Dont-Worry
        #13

        Thank you for your answer @necrevistonnezr
        @girish To answer your question, I am using AWS for backups. I tried to use OVH when the problem started but idk why i cant connect OVH Object Storage with my cloudron.

        @necrevistonnezr When you say "Using the encrypted tar backend for this mass of - already compressed - data is asking for trouble, IMHO.". I don't think 200gb is a huuuge mass of data (especially for AWS that is not a B2C focused provider). For our Company this is a little bit concerning if we cannot backup anything above 100gb.... Because we use cloudron for Entertainment (Emby), but also for business purposes (Files storages, apps ...) and we planned to use cloudron for long term purposes and forcasted more than 5 to 10 TB of data in a year. There is any way to Backup large amount of data w/cloudron in an automated way, as the system usually worked ?

        1 Reply Last reply
        0
        • necrevistonnezrN Offline
          necrevistonnezrN Offline
          necrevistonnezr
          wrote on last edited by
          #14

          Use the rsync backend?

          1 Reply Last reply
          0
          • D Offline
            D Offline
            Dont-Worry
            wrote on last edited by Dont-Worry
            #15

            The problem has been solved. I have increased the size of the packages which are sent to the S3 and the whole was less than 10,000 packages so it was getting in. I read the documentation and really been over all my Cloudron panel many times, but I really didn't think about reviewing the advanced configurations of my backup, that is the thing I do not often change.

            This solution works in the short term, but according to my calculations, once it reaches 10Tb, the maximum cut proposed by Cloudron, namely the cut by 1gb per package, will not be enough. Would it be possible to increase the maximum in the Cloudron application for people who use and consider Cloudron as a long-term mid-scale solution ?

            This would not change anything if it was adjustable, but for people like us, who will have to deploy Cloudron on a powerful infrastructure, the double could be supported without any problem (2gb/Package).

            image.png

            girishG 1 Reply Last reply
            1
            • D Dont-Worry

              The problem has been solved. I have increased the size of the packages which are sent to the S3 and the whole was less than 10,000 packages so it was getting in. I read the documentation and really been over all my Cloudron panel many times, but I really didn't think about reviewing the advanced configurations of my backup, that is the thing I do not often change.

              This solution works in the short term, but according to my calculations, once it reaches 10Tb, the maximum cut proposed by Cloudron, namely the cut by 1gb per package, will not be enough. Would it be possible to increase the maximum in the Cloudron application for people who use and consider Cloudron as a long-term mid-scale solution ?

              This would not change anything if it was adjustable, but for people like us, who will have to deploy Cloudron on a powerful infrastructure, the double could be supported without any problem (2gb/Package).

              image.png

              girishG Offline
              girishG Offline
              girish
              Staff
              wrote on last edited by
              #16

              @Dont-Worry Thanks for the follow-up. Can you tell me which region of OVH Object Storage you are using?

              D 1 Reply Last reply
              1
              • girishG girish

                @Dont-Worry Thanks for the follow-up. Can you tell me which region of OVH Object Storage you are using?

                D Offline
                D Offline
                Dont-Worry
                wrote on last edited by
                #17

                @girish I am using Graveline for OVH Object Storage. First I choosed Roubaix, but it wasn't in the Cloudron list so I created a new bucket in Graveline. But it never worked. Now it is still working on the old AWS S3 bucket.
                Didn't found any way to make the backup to an OVH bucket work.

                girishG 1 Reply Last reply
                1
                • humptydumptyH Offline
                  humptydumptyH Offline
                  humptydumpty
                  wrote on last edited by
                  #18

                  @girish Can you please expand the doc page to cover what the "advanced settings" in the backups section offers and how they can be best used/optimized for a few use-cases like (general-use, heavy media use, etc.). Thank you!

                  girishG 1 Reply Last reply
                  1
                  • D Dont-Worry

                    @girish I am using Graveline for OVH Object Storage. First I choosed Roubaix, but it wasn't in the Cloudron list so I created a new bucket in Graveline. But it never worked. Now it is still working on the old AWS S3 bucket.
                    Didn't found any way to make the backup to an OVH bucket work.

                    girishG Offline
                    girishG Offline
                    girish
                    Staff
                    wrote on last edited by
                    #19

                    @Dont-Worry thanks, I have added RBX to the region list.

                    D 1 Reply Last reply
                    1
                    • humptydumptyH humptydumpty

                      @girish Can you please expand the doc page to cover what the "advanced settings" in the backups section offers and how they can be best used/optimized for a few use-cases like (general-use, heavy media use, etc.). Thank you!

                      girishG Offline
                      girishG Offline
                      girish
                      Staff
                      wrote on last edited by
                      #20

                      @humptydumpty I think the issue is that there are too many providers, each with their own region. Back in the day, we hardcoded this and the advanced setting was not even exposed to the user because we thought it's impossible for the user to guess correct values. We ended up exposing the values because we couldn't figure a way to guess values . Even now, I have no concrete values to suggest for each provider, they keep changing (since upstream providers are also deploying updates) and each network/app/cpu is different. It's a bit of trial and error 🤔 . But I will discuss internally and add something to the docs to help users.

                      1 Reply Last reply
                      1
                      • girishG girish has marked this topic as solved on
                      • girishG girish

                        @Dont-Worry thanks, I have added RBX to the region list.

                        D Offline
                        D Offline
                        Dont-Worry
                        wrote on last edited by
                        #21

                        @girish Thank you very much !

                        1 Reply Last reply
                        0
                        Reply
                        • Reply as topic
                        Log in to reply
                        • Oldest to Newest
                        • Newest to Oldest
                        • Most Votes


                        • Login

                        • Don't have an account? Register

                        • Login or register to search.
                        • First post
                          Last post
                        0
                        • Categories
                        • Recent
                        • Tags
                        • Popular
                        • Bookmarks
                        • Search