Object Storage or Block Storage for backups of growing 60+ GB?
-
@d19dotca said in Object Storage or Block Storage for backups of growing 60+ GB?:
more storage
Depends if you store backup for 2 days or for 1 year.
Rsync is an incremental backup, you will never store the same file 2 times.
But is disadvantage is that I'd you even get 1 corrupted snapshot you will lose everything after that. -
@MooCloud_Matt Would you suggest Backblaze over Wasabi in that case? Backblaze charges on command types too so I assumed I’d be better with Wasabi but point taken as their 3-month lifespan requirement struck me as very strange. Went with Wasabi still as it seemed unlikely I’d hit the 1TB limit even with deleted files, plus it has a Canadian Datacentre which will be more performant latency-wise then Backblaze’s California Datacentre when my VPS is hosted in Toronto, Canada.
-
@d19dotca On Vultr, object storage starts at $5/month for 250GB and 1TB of transfer. Block storage is $25/month for 250GB.
-
@LoudLemur Seems to be half that price on Contabo, even less on Wasabi.
-
@robi said in Object Storage or Block Storage for backups of growing 60+ GB?:
@LoudLemur Seems to be half that price on Contabo.
I was surprised as the Vultr Object Storage is nvme, but the block storage can be HDD, so I thought the block storage would be cheaper.
What is the best toolto browse through block storage files to find one you want?
-
@LoudLemur there are many tools, depends what OS you use.. WinSCP or say Cyberduck on MacOS.
-
@LoudLemur Yeah I used Vultr for a little bit when testing but their Datacentre location for their object storage was far away from my VPS Datacentre in Vultr, and it's not too cheap at least compared to what sort of storage you get with Wasabi or Backblaze for example. Not too bad though, for sure, and worth consideration for some.
-
@d19dotca said in Object Storage or Block Storage for backups of growing 60+ GB?:
@LoudLemur Yeah I used Vultr for a little bit when testing but their Datacentre location for their object storage was far away from my VPS Datacentre in Vultr, and it's not too cheap at least compared to what sort of storage you get with Wasabi or Backblaze for example. Not too bad though, for sure, and worth consideration for some.
-
@d19dotca
You must understand what is more valuable for you: speed, reliability, or price.
reliability mostly Backblaze’s services are one of the best.
Speed, Wasabi is good enough if is near to your data center, but the is no duplication of data, and is something that we have to use more than want I like to admit. -
Interesting. I decided to test something with a local Datacentre closer to my VPS called IDrive e2 (seems like a recent s3 competitor from mid-2022 which promises high speeds).
Backblaze is good although I find it quite slow (mostly because my VPS is in a very far away Datacentre from Backblaze's California (us-west) location. Speed isn't critical since it's just backups but definitely helps still.
Backblaze's pricing for their API calls scares me a little bit, makes me think it'll be much more pricey than I'm anticipating. May just need to test it out for a while to verify.
I see what you mean about Wasabi's weird 90-day storage policy which means even deleted files are still counted for 90 days, and my current estimate is quickly adding up, so I think despite initially happy with Wasabi's performance I may need to abandon that provider.
Still experimenting. Currently in the middle of a large 60+GB backup to IDrive e2 (using rsync instead of tarball for now) and have to say I'm super impressed with the speeds. Their pricing is also quite minimal. Will see if they end up being the one I use.
-
@MooCloud_Matt said in Object Storage or Block Storage for backups of growing 60+ GB?:
@d19dotca
You must understand what is more valuable for you: speed, reliability, or price.
reliability mostly Backblaze’s services are one of the best.
Speed, Wasabi is good enough if is near to your data center, but the is no duplication of data, and is something that we have to use more than want I like to admit.And price? Which one?
-
@LoudLemur probably contabo, they use ceph.
This means that u have replication by default and a really good support for S3. -
FWIW, I’m quite impressed with the pricing and performance of the newer IDrive e2 storage (with s3 API).
Curious though on a related but different topic… when uploading to s3, do you find yourself using rsync for larger backups or do you opt to just use tgz? I so far tend to find rsync a bit more performant (likely because of the concurrency settings) but the downside is it takes forever to delete files out of the bucket when there’s so many of them, the deletion process even with s3 API calls is incredibly slow when so many files exist. Makes me think it may be better to just stick to tarball images instead.
-
@d19dotca do you know much about these idrive folks / would you recommend? the pricing almost makes them look like a scam - 2TB of storage for $8/year unlimited egress? i mean, they LOOK legit, but wow that's a helluva deal.
-
@doodlemania2 IDrive has been around for many years for the computer backup solution (competitors to Backblaze and Carbonate for example). This year they released their s3 storage competitor, so they're basically following in the same shoes as Backblaze. It's all legitimate. They're newer so they're trying to sweeten the deal to attract people away from Backblaze and such, hence the lower pricing if you buy their promo for the year plan. It's only that price for the first year though, definitely not a scam IMO.
I'm currently just trialling with their monthly plan as it's only $0.004 USD per GB, so to store even 800 GB of backups in a month would only be $3.20 USD.
-
@d19dotca I keep thinking that between these newcomers with their free egress and cloudflare shaming them, that Azure and AWS et al will drop the egress fees (or at least lower them)!
-
@doodlemania2 Egress fees? If you want to take your data away, you have to pay?
-
@d19dotca You might not believe it, but my idrive account goes back to 2009!
-
@scooke funny story - i just migrated all my stuff to idrive for my mastodon instance only to discover - twoops - they don't appear to allow public buckets!?
-
@doodlemania2 I also signed up (the regular idrive.com account didn't work with the newer idrive e2), activated S3, made a bucket, checked out the user info, and then saw the buckets were private (for now). Bad timing, hey!
-
@doodlemania2 I think that's a setting before you create the bucket.
-
@robi Unfortunately it is so. Not sure when they will enable the setting.
-
@scooke anyone know a contact? Would like to stay with them but no public buckets...lol - maybe they can bit flip it for me?
-
went with their main tech support email
-
@scooke all regions?
-
Noticing that tgz of about 35-40 GB GB (well that’s the size after compression, it’s really compressing closer to 60+ GB uncompressed data) and it’s nearly double what it consumes in storage because at least the first backup has that snapshot directory too which is a duplicate of the latest backup. My IDrive e2 bucket shows a total usage of 70 GB currently after one system backup.
I notice if I use rsync the backup times are anywheres between about 20-70 minutes, however when using tgz it’s closer to 120 minutes. That’s about the same performance when using Wasabi and even better than Backblaze in my tests so far.
The downside to rsync is it takes forever to delete so many (tens of thousands of) different files from s3 storage, which means Cloudron cleaning it up or even just me doing it manually to purge older data is a real time sink. Makes me want to use tarballs instead however the upload rate seems to be much less performant. I suppose it is maybe a moot point to some degree because it’d normally be done in the middle of the night when time isn’t really a factor, although it will mess things up for two hours during Cloudron version updates if I (which I would) do system backup prior to that task as the maintenance would be so much longer.
I’m curious though… is there a way to improve the tarball performance? Why is it so dang slow? I assume concurrency but what prevents concurrency from taking place when using tarball? @staff, any suggestions here? Anything I can do or any specific recommendations?
I should note I’ve seen next to zero improvements or changes at all whether I use a part size of 128 MB or 1 GB, doesn’t seem to change anything performance wise.
-
@d19dotca said in Object Storage or Block Storage for backups of growing 60+ GB?:
I’m curious though… is there a way to improve the tarball performance? Why is it so dang slow? I assume concurrency but what prevents concurrency from taking place when using tarball? @staff, any suggestions here? Anything I can do or any specific recommendations?
One idea might be to measure just uploading via other tools to see how much the performance difference is. While the tgz code uploader is slow, it's not substantially slower than native tools (from what I have tested). Try to do a tar cvf - upload of the same data directory as the app to the backup storage and measure.
-
@robi I checked most (all non-US ones, 4 US ones) and the Public radio button stayed greyed out.
-
@scooke let's see what support says.. unless they have a status page where one can see a portion of their services being offline (which provides this public access to a private bucket).
-
@scooke So I heard back from email based support - they turned public buckets on for my account!
-
@d19dotca Thanks for your R&D sharing on this, just trialing it, and the 1 year deals save a lot of money compared to any other S3, including Wasabi.
Really like the UI/UK, much easier for anyone to work with than Wasabi too, as you don't need to create Policies in JSON like you do with other.
from me, definitely recommended for all home users, and most SME needs too!
@d19dotca said in Object Storage or Block Storage for backups of growing 60+ GB?:
Interesting. I decided to test something with a local Datacentre closer to my VPS called IDrive e2 (seems like a recent s3 competitor from mid-2022 which promises high speeds).
-
FYI, got an email today from IDrive which announced the public buckets for people looking for that, mentioning it here as I know a few people were asking about that feature.
They’ve also made some serious performance tweaks. I noticed before I was seeing it take about 55-75 minutes for a 30+ GB upload in tarball, and now it’s closer to 45-50 minutes for the same size (in fact possibly even larger of a file now to boot as I have a couple more apps deployed now too and email continues to grow larger).
-
@d19dotca said in Object Storage or Block Storage for backups of growing 60+ GB?:
I noticed before I was seeing it take about 55-75 minutes for a 30+ GB upload in tarball, and now it’s closer to 45-50 minutes for the same
It may be related to lower traffic and usage over the end of year holidays.
-
@robi said in Object Storage or Block Storage for backups of growing 60+ GB?:
It may be related to lower traffic and usage over the end of year holidays.
Entirely possible although I was discussing with them in a support case about the general speed of things and they did say they had recently implemented a change to their service which should speed things up and in my experience it does seem to be improved.
-
@d19dotca I also noticed an increase in performance. One thing I did notice though - my public buckets were part of the beta and included anonymous root "viewing" of sorts - it generated an XML file of the contents of each public bucket. A quick email to support and they had me turn the bucket to private, then back to public to correct. Something to check if you or someone you love may be an early public bucket adopter there!
-
@d19dotca We're noticing some significant slowdowns with IDrive backup uploads now. (We have lots and lots of small files and over 1TB). Just wondering if you'd experienced anything that might seem like throttling?
-
@marcusquinn Oh oh... too good to be true?
-
@scooke Maybe, trying their support.
-
@marcusquinn
idk what tech is behind their stack, but if it is just HDD with no NVMe for caching metadata or any kind of index.
100% that it will be slow if they grow too much. -
Changing strategy to Tarball and that seems to complete in a few minutes. Might just have to be the trade-off for our apps, that have lots and lots of small files. More S3 storage usage but faster to backup and restore, and more self-contained for each backup not relying on files from others.
-
@marcusquinn yes, that is the more optimal format for object storage, plus incremental diffs.
We discussed elsewhere on the forum a hybrid option inbetween rsync by file and large tgz.
-
Yeah, I'm gonna call this, any backups >100GB or >100,000 files is likely to get impractically slower for most budget S3 storage. Might be worth a note/tooltip in the settings to suggest tarball for Cloudron servers over these numbers.
TBH I think the compression and minimal numbers of files being uploaded, with a sensible retention policy, is going to offset any storage-saving from using rsync. Rsync is a nice idea in theory for smaller directories, but I feel the file count costs are higher than storage costs for local compression and uploading that for each backup run.
-
@marcusquinn I haven’t noticed that myself but been using tgz the whole time and it’s been quite fast overall so far.
-
To confirm on my R&D inspired by this thread. We're happy now with:
- IDrive on the introduction offer pricing
- Tarball for the backup method
- IDrive and Tarball encryption
- 7 Daily, 4 Weekly, 12 Monthly retention policy