Top
Best
New

Posted by rrreese 14 hours ago

Backblaze has stopped backing up OneDrive and Dropbox folders and maybe others(rareese.com)
889 points | 540 comments
julianozen 7 hours ago|
We are going to drop blackblaze over this

We discovered this change recently because my dad was looking for a file that Dropbox accidentally overwrote which at first we said “no problem. This is why we pay for backblaze”

We had learned that this policy had changed a few months ago, and we were never notified. File was unrecoverable

If anyone at backblaze is reading this, I pay for your product so I can install you on my parents machine and never worry about it again. You decided saving on cloud storage was worth breaking this promise. Bad bad call

wafflebot 7 hours ago||
I'm going to drop Backblaze for my entire company over this.

I need it to capture local data, even though that local data is getting synced to Google Drive. Where we sync our data really has nothing to do with Backblaze backing up the endpoint. We don't wholly trust sync, that's why we have backup.

On my personal Mac I have iCloud Drive syncing my desktop, and a while back iCloud ate a file I was working on. Backblaze had it captured, thankfully. But if they are going to exclude iCloud Drive synced folders, and sounds like that is their intention, Backblaze is useless to me.

therealpygon 1 hour ago|||
Same. Specifically I was considering Backblaze for our company’s backups (both products, computers and their bucket for server backups. That is no longer the case as of the news.
moffkalast 6 hours ago||||
Bidirectional auto file sync is a fundamentally broken pattern and I'm tired of pretending it's not. It's just complete chaos with wrong files constantly getting overridden on both ends.

I have no clue why people still use it and I'd cut my losses if I were you, either backup to the cloud or pull from it, not both at the same time like an absolute tictac.

Aurornis 4 hours ago|||
> I have no clue why people still use it

This is an instance of someone familiar with complex file access patterns not understanding the normal use case for these services.

The people using these bidirectional sync services want last writer wins behavior. The mild and moderately technical people I work with all get it and work with it. They know how to use the UI to look for old versions if someone accidentally overwrites their file.

Your characterization as complete chaos with constant problems does not mesh with the reality of the countless low-tech teams I've seen use Dropbox type services since they were launched.

Anamon 3 hours ago||
This would be half OK if it worked, but you can't trust it to. OneDrive, for instance, has an open bug for years now where it will randomly revert some of your files to a revision from several months earlier. You can detect and recover this from the history, but only if you know that it happened and where, which you usually won't because it happens silently. I only noticed because it happened to an append-only text file I use daily.
sunnybeetroot 2 hours ago||
Even crazier is one drive has a limit on the total length of a file path, how is this even a thing that exists.
ahhhhnoooo 2 hours ago|||
Unlimited strings are a problem. People will use it as storage.

No, I'm not joking. We used to allow arbitrary paths in a cloud API I owned. Within about a month someone had figured out that the cost to store a single byte file was effectively zero, and they could encode arbitrary files into the paths of those things. It wasn't too long before there was a library to do it on Github. We had to put limits on it because otherwise people would store their data in the path, not the file.

garaetjjte 1 hour ago||
Just include filename size in file size for billing purposes?
thebrain 4 minutes ago||||
Everything needs limits otherwise someone will figure out how to or accidentally break it.
dns_snek 36 minutes ago|||
What do you expect to happen when your cloud storage file path is 5000 characters long and your local filesystem only supports a maximum of 4096?
omnimus 5 hours ago||||
I also have no clue why people use it.

You can build such a system yourself quite trivially by getting an FTP account, mounting it locally with curlftpfs, and then using SVN or CVS on the mounted filesystem. From Windows or Mac, this FTP account could be accessed through built-in software.

snowwrestler 3 hours ago|||
This reference is 19 years old this month, in case anyone who recognized it was still feeling young.
morganf 58 seconds ago||
noooooooooooo!!!!!!!!
j1elo 5 hours ago||||
Wait a moment, you just gave me an idea for a product
zymhan 4 hours ago||||
This cannot be a serious proposal. You should probably talk to people who don't use technology because they love it, but because they need it.
burnte 5 hours ago||||
1 out of a thousand people might do that, the others will buy the product. That's why people use it, most people don't want to build everything themselves.
srdjanr 5 hours ago||
It's a reference to https://news.ycombinator.com/item?id=8863
Dylan16807 2 hours ago||
But as usual it forgets the "For a Linux user" part.

If we remove the whole linux section and just ask "why not map a folder in Explorer" it's a reasonable question, probably even more reasonable in 2026 than in 2007. The network got faster and more reliable, and the dropbox access got slower.

lazide 3 hours ago||||
Obvious. Explorer even has support built in for transparent ‘native’ gui support. I’m not even sure why you felt the need to explain it in detail. Next you’ll be explaining how to walk. (/s, I loved it)
PunchyHamster 5 hours ago|||
Slow as fuck compared to 2 synced dirs
PunchyHamster 5 hours ago||||
It works perfectly fine if you're user that know how it works. I use it with Syncthing and it works coz I know to not edit same file at the same time on 2 devices (my third and fourth device is always on internet so chances propagate reasonably fast even if the 2 devices aren't on on the same time)

But the moment that hits normal users, yeah, mess

babypuncher 6 hours ago||||
I think this is a case of people using bidirectional file sync wrong. The point is to make the most up to date version of a file available across multiple devices, not to act as a backup or for collaboration between multiple users.

It works perfectly fine as long as you keep how it works in mind, and probably most importantly don't have multiple users working directly on the same file at once.

I've been using these systems for over a decade at this point and never had a problem. And if I ever do have one, my real backup solution has me covered.

senkora 5 hours ago|||
+1. It works perfectly if your mental model is:

“Every file is only ever written to from a single client, and will be asynchronously made available to all other clients, and after some period of time has elapsed you can safely switch to always writing to the file from a different client”.

01HNNWZ0MV43FF 5 hours ago||
Bidirectional file sync is also in hot demand from people who don't know the words, "file", "client", "write", "async", "available", or "time"

:P

babypuncher 4 hours ago||
The fact that lay people can and will use a tool incorrectly does not mean said tool is not useful
donmcronald 3 hours ago||||
> And if I ever do have one, my real backup solution has me covered.

What do you use and how do you test / reconcile to make sure it’s not missing files? I find OneDrive extremely hard to deal with because the backup systems don’t seem to be 100% reliable.

I think there are a lot of solutions these days that error on the side of claiming success.

andrejserafim 5 hours ago|||
I agree. I use syncthing for syncing phones and laptops. For data like photos, which aren't really updated. It works very nice. And for documents updated by one user, moving between devices is totally seamless.

That being said i understand how it works at a high level.

lazide 6 hours ago||||
Throw some clock skew into the mix and it’s even more hilarious!
Gud 6 hours ago|||
Why is this downvoted?
adamsb6 6 hours ago||
The insult to tictacs.
huflungdung 2 hours ago|||
[dead]
varenc 25 minutes ago|||
Dropbox itself should already keep version history of files for 30 days with the free plan, or more if you pay.
petercooper 6 hours ago|||
I'm going to join the exodus, though for a different reason. Switched to Orbstack and ever since Backblaze refuses to back up saying "disk full" as Orbstack uses a 8TB sparse disk image. You can exclude it, but if they won't (very easily) fix a known issue by measuring file sizes properly I don't feel confident about the product.
julianozen 7 hours ago|||
Also taking recommendations for a simple services I can install on my dads windows machine and my moms Mac that will just automatically backup the main drive to the cloud just in case
jamie_ca 4 hours ago|||
I've been extremely happy with Arq https://www.arqbackup.com/ for several years as a quiet backup solution, bring your own storage. I've done a few small restores and it's been just fine, and it automatically thins your backups to constrain storage costs.

Managing exclusions is something to keep vaguely on top of (I've accidentally had a few VM disk images get backed up when I don't need/want them) but the default exclusions are all very reasonable.

noja 2 hours ago||||
I like https://www.borgbase.com

It's set it and forget.

You will need to set it up for them, then you get an email (from borgbackup, not the client so it works when the client is not running) when a backup hasn't happened for a while.

As client there are more options now (like Vorta, from them), but I have had success with https://github.com/garethgeorge/backrest and the Restic backend.

web007 2 hours ago||||
`rclone` with AWS credentials. Go make a bucket and a key that can read/write to it.

Set up your config to exclude common non-file dirs, or say "only `/Applications` and `Home` and that's about it. If it's a file then it's a file, and it will be synced up.

physicles 3 hours ago|||
Installed Carbonite on my parents’ computer something like 15 years ago, and it still works (every now and then my dad tells me he used it to recover from a bug or a mistake).

But I have no idea where the company currently sits on the spectrum from good actor to fully enshittified.

ibizaman 7 hours ago|||
Why not just use backblaze as cold storage and use restic or another tool with a GUI to backup to it?
julianozen 7 hours ago|||
Well this wasn’t the promise backblaze made a decade ago when we started using their products.

Now I need a new solution that will work for my parents

ibizaman 6 hours ago||
Of course, I wouldn’t use their client anymore. Actually, I would have never used it from the start as it’s not open source. I think for backups there’s no better guarantee than that. I don’t mean because you could look at the source code, I mean because in my experience open source products tend to care more about their users than not. At least for such foundational tools.
reddalo 6 hours ago||||
Backblaze Computer Backup != B2 Cloud Storage

You can't connect to their Computer Backup service through third-party software.

jsw97 5 hours ago||||
When an org quietly degrades one of their products, you should expect this behavior to occur again.
cortesoft 5 hours ago|||
That is a lot more expensive if you have more than a small amount of data.
sieabahlpark 1 hour ago||
[dead]
Neil44 10 hours ago||
The issue with a client app backing up dropbox and onedrive folders on your computer is the files on demand feature, you could sync a 1tb onedrive to your 250gb laptop but it's OK because of smart/selective sync aka files on demand. Then backblaze backup tries to back the folder up and requests a download of every single file and now you have zero bytes free, still no backup and a sick laptop. You could oauth the backblaze app to access onedrive directly, but if you want to back your onedrive up you need a different product IMO.
appreciatorBus 9 hours ago||
Shoutout to Arq backup which simply gives you an option in backup plans for what to do with cloud only files:

- report an error

- ignore

- materialize

Regardless, if you make it back up software that doesn’t give this level of control to users, and you make a change about which files you’re going to back up, you should probably be a lot more vocal with your users about the change. Vanishingly few people read release notes.

vunderba 6 hours ago|||
I honestly didn't even realize Backblaze had a clientside app. Very happy user of Arq - been running a daily scheduled dual backup of my HDD to an external NAS and Backblaze B2 for years with zero issues.
justusthane 6 hours ago||
That was their whole business originally. The block storage is a newer offering.
decadefade 8 hours ago||||
Love Arq!
Lord_Zero 7 hours ago|||
Why no linux support?
sreitshamer 3 hours ago|||
(Arq developer here) Haven't gotten many requests for it at all over the years. I presume it's because there are so many free options for Linux.
bartvk 3 hours ago|||
Just wanted to say that for many years, Arq has been my backup solution. It's amazing and I advise it to everyone I know.
dreamcompiler 2 hours ago|||
I for one would pay for Arq on Linux as I now do on Mac. It would be fantastic to be able to use the same "it just works" backup solution on all my computers.
SSLy 35 minutes ago||||
we already have restic which is really really great.
CamperBob2 6 hours ago|||
If it's open-source, Linux support is only a few hours with Claude away.

If it's not open-source, but the protocol is documented, see above.

If it's not open-source, and the protocol isn't documented, well... that makes the decision easy, doesn't it?

cobertos 6 hours ago||
Backups software written by Claude? No thanks.

I've used enough Claude coded applications that I wouldn't trust that with a backup, unless it had extensive tests along with it.

CamperBob2 6 hours ago||||
And I've used enough "gold standard" commercial applications, like the one being discussed in this very article, that I don't trust those either. If you recoil in horror at code written by LLMs, I'm afraid that the vendors you're already working with have some really bad news for you. You can get over it now or get over it later. You will get over it.

I can audit and verify Claude's output. Code running at BackBlaze, not so much. Take some responsibility for your data. Rest assured, nobody else will.

samspot 5 hours ago|||
You are not wrong, but I just don't have time. My choices are pay someone or throw my hands up. I have been paying backblaze. But I recently had a drive die, and discovered the backups are missing .exe and .dll files, and so that part of the restore was worthless.

What time I do have, I've been using to try and figure out photo libraries. Nothing is working the way I need it to. The providers are a mess of security restrictions and buggy software.

nozzlegear 5 hours ago||||
The choices are maybe eat shit, or spend your own time auditing and polishing shit into something edible before eating it?
CamperBob2 5 hours ago||
That's the general conclusion you will draw after reading the comments on this story, yes.
BizarroLand 5 hours ago|||
>You can get over it now or get over it later. You will get over it

You're forgetting the third option:

You can remain blissfully unaware of it.

CamperBob2 5 hours ago||
You can remain blissfully unaware of it.

And you can read many accounts of the outcome of that strategy in this very thread.

kid64 6 hours ago|||
"Dear Claude, please create an eztensive testing suite for this app. Love, cobertos"
lolc 4 hours ago||
"Great idea wise customer, I will certainly mock one out just for you!"
CamperBob2 3 hours ago||
My favorite Peanuts comic was always the one where Linus is standing at an intersection next to a 'Push Button To Cross Street' sign. He is sucking his thumb and clutching his blanket despondently.

In the last panel, Charlie Brown tells him, "You have to move your feet, too."

ineedasername 7 hours ago|||
That seems like a pretty straightforward issue to solve, to simply backup only those files that are actually on the system, not the stubs. If it's on your computer, it should able to get backed up. If it's just a shadow, a pointer, it doesn't.

Making the change without making it clear though, that's just awful. A clear recipe for catastrophic loss & drip drip drip of news in the vein of "How Backblaze Lost my Stuff"

wrs 6 hours ago|||
The OP’s complaint is that the files were not backed up. If they had discovered that only stubs were backed up, I don’t think they’d be any happier.
ineedasername 20 minutes ago||
Not what I meant: The other cloud storage services connected to the computer, eg OneDrive. Those files, when they are just stubs. I'm saying that Backblaze could simply not backup stubs, if the person isn't syncing the actual file to their drive. If they are, backblaze should back it up.
lazide 7 hours ago|||
The stubs are the thing on your computer?
coldtea 7 hours ago||
Imagine if they could detect stab or real file huh? Space technology, I know! Or just fucking copy them as stubs and what's actually downloaded as actually downloaded! Boggles the mind!

Or maybe just do what they do now, but WARN about that in HUGE RED LETTERS, in the website and the app, instead of burying it in an update note like weasels!

whimblepop 8 hours ago|||
The whole "just sync everything, and if you can't seek everything, pretend to sync everything with fake files and then download the real ones ad-hoc" model of storage feels a bit ill-conceived to me. It tries to present a simple facade but I'm not sure it actually simplifies things. It always results in nasty user surprises and sometimes data loss. I've seen Microsoft OneDrive do the same thing to people at work.
carefulfungi 7 hours ago||
I’ve lost data not realizing I was backing up placeholder files (iCloud).

Hiding the network always ends in pain. But never goes out of style.

eblume 5 hours ago|||
Same. I lost a lot of photos this way. I've recently moved over to Immich + Borg backup with a 3-2-1 backup between a local synology NAS and BorgBase. Painful lesson, but at least now I feel much more confident. I've even built some end-to-end monitoring with Grafana.
chrisweekly 4 hours ago||
Careful w that Synology NAS, mine's now a brick that may also have led to permanent data loss.
whimblepop 6 hours ago|||
My own approach to simplicity generally means "hide complexity behind a simple interface" rather than pushing for simple implementations because I feel that too much emphasis on simplicity of implementations often means sacrificing correctness.

This particular example is a useful one for me to think about, because it's a version of hiding complexity in order to present a simple interface that I actually hate. (WYSIWYG editors is another one, for similar reasons: it always ends up being buggy and unpredictable.)

thecapybara 8 hours ago|||
That would make sense for online-only files, but I have my Dropbox folder set to synchronize everything to my PC, and Backblaze still started skipping over it a few months ago. I reached out to support and they confirmed that they are just entirely skipping Dropbox/OneDrive/etc folders entirely, regardless of if the files are stored locally or not.
signorovitch 8 hours ago|||
The primary trouble I have with backblaze was that this change was not clearly communicated, even if perhaps it could be justified.
bastawhiz 9 hours ago|||
That doesn't really make a lot of sense, though. Reading a file that's not actually on disk doesn't download it permanently. If I have zero of 10TB worth of files stored locally on my 1TB device, read them all serially, and measure my disk usage, there's no reason the disk should be full, or at least it should be cache that can be easily freed. The only time this is potentially a problem is if one of the files exceeds the total disk space available.

Hell, if I open a directory of photos and my OS tries to pull exif data for each one, it would be wild if that caused those files to be fully downloaded and consume disk space.

jrmg 8 hours ago|||
Right, but even if that’s working it breaks the user experience of services like this that ‘files I used recently are on my device’.

After a backup, you’d go out to a coffee shop or on a plane only to find that the files in the synced folder you used yesterday, and expected to still be there, were not - but photos from ten years ago were available!

wtallis 7 hours ago|||
That shouldn't be seen as Backblaze's problem. It's Dropbox's problem that they made their product too complicated for users to reason about. The original Dropbox concept was "a folder that syncs" and there would be nothing problematic about Backblaze or anything else trying to back it up like any other folder.

Today's Dropbox is a network file system with inscrutable cache behavior that seeks to hide from the users the information about which files are actually present. That makes it impossible for normal users to correctly reason about its behavior, to have correct expectations for what will be available offline or what the side effects of opening a file will be, and Backblaze is stuck trying to cope with a situation where there is no right answer.

realo 6 hours ago||
If I backup a file, I need to read that file. The rest is in the management layer underneath that file.

Seems simple enough to do for Backblaze, no?

wtallis 5 hours ago||
Do you really want Backblaze to ignore all the side effects of scanning through the entire contents a badly-designed network filesystem?
realo 4 hours ago||
What I actually want is not a backup. That is just an artefact of the process.

What i want is restores. The ability to restore anything from ideally any point back in time.

How that is achieved is not my concern.

Obviously Backblaze does not achieve that, today.

wtallis 4 hours ago||
> How that is achieved is not my concern.

You're dodging the question. Wanting to ignore the side effects does not mean they won't affect you.

NetMageSCW 8 hours ago|||
There’s no reason to think that would happen - files you had from ten years ago would have been backed up ten years ago and would be skipped over today.
jrmg 8 hours ago||
Good point (I’m assuming you’re right here and it trusts file metadata and doesn’t read files it’s already backed up?)

It would still happen with the first backup - or first connection of the cloud drive - though, which isn’t a great post-setup new user experience. It probably drove complaints and cancellations.

I feel like I’ve accidentally started defending the concept of not backing up these folders, which I didn’t really intend to. I’d also want these backed up. I’m just thinking out loud about the reasons the decision was made.

bombcar 9 hours ago|||
It's generally now handled decently well, but with three or four of these things it can make backups take annoying long as without "smarts" (which are not always present) it may force a download of the entire OneDrive/Box each time - even if it never crashes out.
bastawhiz 7 hours ago||
> it may force a download of the entire OneDrive/Box each time - even if it never crashes out.

I am not aware of any evidence supporting this.

bombcar 6 hours ago|||
The issue really isn't that it's not backing up the folder (which I can see an argument for both sides and various ways to do it) - it's that they changed what they did in a surprising way.

Your backup solution is not something you ever want to be the source of surprises!

danpalmer 10 hours ago|||
This is a complexity that makes it harder, but not insurmountable.

It would be reasonable to say that if you run the file sync in a mode that keeps everything locally, then Backblaze should be backing it up. Arguably they should even when not in that mode, but it'll churn files repeatedly as you stream files in and out of local storage with the cloud provider.

bayindirh 10 hours ago|||
> Arguably they should even when not in that mode, but it'll churn files repeatedly as you stream files in and out of local storage with the cloud provider.

When you have a couple terabytes of data in that drive, is it acceptable to cycle all that data and use all that bandwidth and wear down your SSD at the same time?

Also, high number of small files is a problem for these services. I have a large font collection in my cloud account and oh boy, if I want to sync that thing, the whole thing proverbially overheats from all the queries it's sending.

jtbayly 9 hours ago|||
Reading your comments, it sounds like you are arguing it is impossible to backup files in Dropbox in any reasonable way, and therefore nobody should backup their cloud files. I know you haven’t technically said that, but that’s what it sounds like.

I assume you don’t think that, so I’m curious, what would you propose positively?

bayindirh 9 hours ago||
> I know you haven’t technically said that, but that’s what it sounds like.

Yes, I didn't technically said that.

> It sounds like you are arguing it is impossible to backup files in Dropbox in any reasonable way, and therefore nobody should backup their cloud files.

I don't argue neither, either.

What I said is with "on demand file download", traditional backup software faces a hard problem. However, there are better ways to do that, primary candidate being rclone.

You can register a new application ID for your rclone installation for your Google Drive and Dropbox accounts, and use rclone as a very efficient, rsync-like tool to backup your cloud storage. That's what I do.

I'm currently backing up my cloud storages to a local TrueNAS installation. rclone automatically hash-checks everything and downloads the changed ones. If you can mount Backblaze via FUSE or something similar, you can use rclone as an intelligent MITM agent to smartly pull from cloud and push to Backblaze.

Also, using RESTIC or Borg as a backup container is a good idea since they can deduplicate and/or only store the differences between the snapshots, saving tons of space in the process, plus encrypting things for good measure.

nine_k 7 hours ago|||
This. You should not try to backup your local cache of cloud files as if those were your local files. Use a tool that talks to the cloud storage directly.

Use tools with straightforward, predictable semantics, like rclone, or synching, or restic/Borg. (Deduplication rules, too.)

mroche 5 hours ago|||
My understanding of Backblaze Computer Backup is it is not a general purpose, network accessible filesystem.[0] If you want to use another tool to backup specific files, you'd use their B2 object storage platform.[1] It has an S3 compatible API you can interact with, Computer Backup does not.

But generally speaking, I'd agree with your sentiment.

[0]: https://www.backblaze.com/computer-backup/docs/supported-bac...

[1]: https://www.backblaze.com/docs/cloud-storage-about-backblaze...

vladvasiliu 9 hours ago||||
But if the files are only on the remote storage and not local, chances are they haven't been modified recently, so it shouldn't download them fully, just check the metadata cache for size / modification time and let them be if they didn't change.

So, in practice, you shouldn't have to download the whole remote drive when you do an incremental backup.

bayindirh 9 hours ago||
You can't trust size and modification time all the time, though mdate is a better indicator, it's not foolprooof. The only reliable way will be checksumming.

Interestingly, rclone supports that on many providers, but to be able to backblaze support that, it needs to integrate rclone, connect to the providers via that channel and request checks, which is messy, complicated, and computationally expensive. Even if we consider that you won't be hitting API rate limits on the cloud provider.

NetMageSCW 8 hours ago||
If you can’t trust modification time you are doing something so unusual that you probably need to be handling your backups privately anyway.
bayindirh 8 hours ago||
I don't think so.

Sometimes modification time of a file which is not downloaded on computer A, but modified by computer B is not reflected immediately to computer A.

Henceforth, backup software running on computer A will think that the file has not been modified. This is a known problem in file synchronization. Also, some applications modifying the files revert or protect the mtime of the file for reasons. They are rare, but they're there.

Chaosvex 9 hours ago|||
Then do it in memory, assuming those services allow you to read the files like that. It sounds like they do based on your other comments.
bayindirh 9 hours ago||
The problem is, downloading files and disk management is not in your control, that part is managed by the cloud client (dropbox, google drive, et. al) transparently. The application accessing the file is just waiting akin to waiting for a disk spin up.

The filesystem is a black box for these software since they don't know where a file resides. If you want control, you need to talk with every party, incl. the cloud provider, a-la rclone style.

NetMageSCW 8 hours ago|||
Why would they do new backups of old files all the time? They would just skip those.
Dylan16807 10 hours ago|||
Unless it does something very weird it won't trigger all those files to download at the same time. That shouldn't be a worry.

And, as a separate note, they shouldn't be balking at the amount of data in a virtualized onedrive or dropbox either considering the user could get a many-terabyte hard drive for significantly less money.

bayindirh 10 hours ago||
> Unless it does something very weird it won't trigger all those files to download at the same time. That shouldn't be a worry.

The moment you call read() (or fopen() or your favorite function), the download will be triggered. It's a hook sitting between you and the file. You can't ignore it.

The only way to bypass it is to remount it over rclone or something and use "ls" and "lsd" functions to query filenames. Otherwise it'll download, and it's how it's expected to work.

Dylan16807 10 hours ago||
Why would it use either of those on all the files at once? It should only be opening enough files to fill the upload buffer.
ethin 9 hours ago|||
I think you might be confusing Backblaze reading files and how Dropbox/OneDrive/Nextcloud/etc. work. NC doesn't enable this by default (I don't think), but Windows calls it virtual file support. There is no avoiding filling the upload buffer, because Backblaze has zero control over how Dropbox downloads files. When Backblaze requests that a file be opened and read, Windows will ask Dropbox or whatever to open the file for it, and to read it. How that is done is up to whatever handles the virtual files. To Backblaze, your Dropbox folder is a normal directory with all that that entails, so Backblaze thinks that it can just zip through the directory and it'll read data from disk, even though that isn't really what's happening. I had to exclude my Nextcloud directory from my Duplicati backups for precisely this reason -- my Nextcloud is hosted on my server, and Duplicati was sending it so many requests it would cause my server to start sending back error 500s.

And no, my server isn't behind cloudflare, primarily because I don't have $200 to throw at them to allow me to proxy arbitrary TCP/UDP ports through their network, and I don't know how to tell CF "Hey, only proxy this traffick but let me handle everything else" (assuming that's even possible given that the usual flow is to put your entire domain behind them).

Dylan16807 9 hours ago||
No, I'm not confusing anything.

Dropbox and onedrive can handle backblaze zipping through and opening many files. The risk is getting too many gigabytes at once, but that shouldn't happen because backblaze should only open enough for immediate upload. If it does happen it's very easily fixed.

If it overloads nextcloud by hitting too many files too fast, that's a legitimate issue but it's not what OP was worried about.

bayindirh 10 hours ago|||
Maybe it'll, maybe it won't, but it'll cycle all files in the drive and will stress everything from your cloud provider to Backblaze, incl. everything in between; software and hardware-wise.
Dylan16807 10 hours ago||
That sounds very acceptable to get those files backed up.

It shouldn't stress things to spend a couple weeks relaying a terabyte in small chunks. The most likely strain is on my upload bandwidth and yeah that's the cost of cloud backup, more ISPs need to improve upload.

bayindirh 9 hours ago||
I mean, cycling a couple of terabytes of data over a 512GB drive is at least full 4 writes, which is too much for that kind of thing.

> more ISPs need to improve upload.

I was yelling the same things to the void for the longest time, then I had a brilliant idea of reading the technical specs of the technology coming to my home.

Lo and behold, the numbers I got were the technical limits of the technology that I had at home (PON for the time being), and going higher would need a very large and expensive rewiring with new hardware and technology.

Dylan16807 9 hours ago|||
4 writes out of what, 3000? For something you'll need to do once or twice ever? It's fine. You might not even eat your whole Drive Write Per Day quota for the upload duration, let alone the entire month.

> the technical limits of the technology that I had at home (PON for the time being)

Isn't that usually symmetrical? Is yours not?

bayindirh 9 hours ago|||
> 4 writes out of what, 3000?

Depends on your device capacity and how much is in actual use. Wear leveling things also wear things while it moves things around.

> For something you'll need to do once or twice ever?

I don't know you, but my cloud storage is living, and even if it's not living, if the software can't smartly ignore files, it'll pull everything in, compare and pass without uploading, causing churns in every backup cycle.

> Isn't that usually symmetrical? Is yours not?

GPON (Gigabit PON) is asymmetric. Theoretical limits is 2.4Gbps down, 1.2Gbps up. I have 1000Mbit/75Mbit at home.

Dylan16807 8 hours ago||
> I don't know you, but my cloud storage is living

But you're probably changing less than 1% each day. And new changes are likely already in the cache, no need to download them.

> if the software can't smartly ignore files, it'll

Backblaze checks the modification date.

> GPON (Gigabit PON) is asymmetric. Theoretical limits is 2.4Gbps down, 1.2Gbps up. I have 1000Mbit/75Mbit at home.

2:1 is fine. If you're getting worse than 10:1 then that does sound like your ISP failed you?

jonhohle 9 hours ago|||
How do you know how often those files need to be backed up without reading them? Timestamps and sizes are not reliable, only content hashes. How do you get a content hash? You read the file.
NetMageSCW 8 hours ago|||
If timestamps aren’t reliable, you fall way outside the user that can trust a third party backup provider. Name a time when modification timestamp fails but a cloud provider will catch the need to download the file.
Dylan16807 9 hours ago|||
Backblaze already trusts the modification date.
NetMageSCW 8 hours ago|||
Why would it do that more than once unless you are modifying 4TB of data every day, in which case you are causing the problem.
bayindirh 8 hours ago||
I don't know how your client works, but reading metadata (e.g. requesting size) off any file causes some cloud clients to download it completely.

Of course I'm not modifying 4TB on a cloud drive, every day.

Dylan16807 3 hours ago||
Can you name such a client? That sounds like a terrible experience.
tencentshill 4 hours ago|||
Cloud placeholders have been a feature for years, plenty of programs have mitigations for this behavior.
downrightmike 5 hours ago||
The fault is with the PC manufacturers screwing you on disk space claiming 1TB, when its only 256gb. bait and switch
azalemeth 12 hours ago||
I guess the problem with Backblaze's business model with respect to Backblaze Personal is that it is "unlimited". They specifically exclude linux users because, well, we're nerds, r/datahoarders exists, and we have different ideas about what "unlimited" means. [1]

This is another example in disguise of two people disagreeing about what "unlimited" means in the context of backup, even if they do claim to have "no restrictions on file type or size" [2].

[1] https://www.reddit.com/r/backblaze/comments/jsrqoz/personal_... [2] https://www.backblaze.com/cloud-backup/personal

embedding-shape 12 hours ago||
Any company that does the "unlimited*" shenanigans are automatically out from any selection process I had going, wherever they use it. It's a clear signal that the marketing/financial teams have taken over the businesses, and they'll be quick to offload you from the platform given the chance, and you'll have no recourse.

Always prefer businesses who are upfront and honest about what they can offer their users, in a sustainable way.

ethbr1 12 hours ago|||
> It's a clear signal that the marketing/financial teams have taken over the businesses

Or that they're targeting the mass retail market, where people are technically ignorant, and "unlimited" is required to compete.

And statistically-speaking, is viable as long as a company keeps its users to a normal distribution.

michaelbuckbee 11 hours ago|||
Is there an example of a consumer facing SaaS that's been able to handle the "unlimited" in a way you'd consider positive?
ethbr1 6 hours ago|||
US cellular data plans? Where it's throttled after soft cap?

Although I will say it's been nice to have them give more transparency around their actual soft cap numbers.

bombcar 6 hours ago||
That’s an example of where unlimited can work (because the limit is a number of hours of degraded service which is quantifiable).

Storage was already a hairy beast with the original setup, and it would be much better if they had defined limits you could at least know about (and pay for).

Aerroon 9 hours ago||||
Google and Youtube, especially Youtube.
tedivm 9 hours ago|||
Google does not have unlimited. I had to pay to increase my storage.
samfriedman 7 hours ago|||
Google Drive reneged on unlimited storage for Education accounts once they realized that universities also contain researchers who need to store huge amounts of data.
Dylan16807 2 hours ago|||
Not only did they cut unlimited, they went to insultingly low limits with not much warning after all their nice promises. Moderately large universities ended up with less space per student than the 15GB they give out to anyone for free. It was a pretty bad rug pull.
Barbing 2 hours ago|||
Massive fraud from abroad didn't help there either. A favorite backup spot for terabytes of pirated media, complete with guides on which schools had good @edu addresses for it.

Hadn't even considered your obvious point, a good one!

ValentineC 5 hours ago|||
Google forced everyone off their deprecated G Suite for Business plan (which had unlimited storage) and onto a Workspace plan.

I had to give up and delete plenty of data because of this. That data was important to me, but not important enough to pay their ransom.

Chaosvex 9 hours ago|||
YouTube is constantly reencoding videos to save space at the expense of older content looking like mud, so arguably even they're having their struggles.
bombcar 6 hours ago|||
We all know the "nobody has watched this video in ten years, login at least once or it'll be yeeted" email is coming, someday.
computably 3 hours ago||
YT would have to start declining in growth pretty substantially for that to be the case. All the 360p video from 2010-2015 probably doesn't take up even 1% of the storage new videos added in 2025.
bombcar 3 hours ago||
True, it's more likely to be aimed at stemming the tide of 4k video that nobody watches - but luckily they're worth more than Disney right now so we don't have to confront that ... yet.
UltraSane 8 hours ago|||
YouTube shorts are incredibly highly compressed.
Spooky23 11 hours ago||||
You can only do it during growth phases or if there’s complimentary products with margin. The story I was told about Office 365 was the when they were using spinning disk, exchange was IOPS-bound, so they had lots of high volume, low iops storage to offer for SharePoint. Google has a similar story, although neither are really unlimited, but approaching unlimited with for large customers.

Once growth slows, churn eats much of the organic growth and you need to spend money on marketing.

Imustaskforhelp 8 hours ago|||
Telegram?
NetMageSCW 8 hours ago||||
>and "unlimited" is required to compete.

And there speaks marketing.

MrRadar 6 hours ago||
Or they're selling their product to a market where the purchaser doesn't understand how much they would need to pay if they were paying by the gigabyte (or even how to check how much they would need). Telling those people they don't need to worry about that "detail" is a key selling point. Backblaze has a product for people who understand the limitations of their consumer product and don't find them acceptable: B2, which is priced by the gigabyte.
Barbing 2 hours ago||
>doesn't understand how much they would need to pay...how to check how much they would need...

...even nearly any frame of reference for anything storage related, much less gigabytes

imiric 10 hours ago||||
> And statistically-speaking, is viable as long as a company keeps its users to a normal distribution.

Doing a bait-and-switch on a percentage of your paying customers, no matter how small the percentage is, may be "viable" for the company, but it's a hostile experience for those users, and companies deserve to be called out for it.

BobaFloutist 7 hours ago||
On the other hand, subsidizing high-usage customers with low-usage customers is pretty generous to the high-usage customers, and there's no pricing model that doesn't suck a little.

Pricing tiers suck if your usage needs are at the bottom of a tier, or you need exactly one premium feature but not more. A la carte pricing is always at least a bit steep, since there's no minimum charge/bulk discount (consider a gym or museum's "day pass") so they have to charge you the full one-time costs every time in case that's your only time.

Base cost + extra per usage might be the best overall, but because nobody has solved micro transactions, the usage fees have to be pretty steep too. And frankly, everyone hates being metered - it means you have to think about pricing every time you go to use something.

malvim 11 hours ago|||
> Or that they're targeting the mass retail market, where people are technically ignorant, and "unlimited" is required to compete.

So… Marketing has taken over, just as parent comment said. Got it.

kccqzy 8 hours ago||||
I just read the Reddit post by their developer and my takeaway is that they have a very good understanding of “unlimited” really means. It’s not a shenanigan. It’s just calculated risk. It’s clear to me that they simultaneously intend to offer truly unlimited backups while hoping that what the average user backs up is within a certain limit that they can easily predict and plan for. It’s a statistical game that they are prepared to play.
embedding-shape 8 hours ago|||
> It’s a statistical game that they are prepared to play.

I understand this, many others do too, the only difference seems to be that we're not willing to play those games. Others are, and that's OK, just giving my point of view which I know is shared by many others who are bit stricter about where we host our backups. Instead of "statistical games" we prefer "upfront limitations", as one example.

bombcar 6 hours ago||||
The problem is you have to play with them - and sure, maybe they're willing to be the Costco to the unlimited backup's $1.50 hotdog - but for how long? Will their dedication to unlimited and particular price points mean you have to take Pepsi for awhile instead of Coke, or that your polish sausage dog disappears? Wait, where did the analogy go? I'm hungry.

It's a bit safer when you know your playbook - if there was unlimited (as it is now) and unlimited plus (where they backup "cloud storage cached files") and unlimited pro max premier (where they backup entire cloud storages) you'd at least know where you stand, and you'd change "holy shit my important file I though was backed up isn't and now it's gone forever" to "I have to pay $10 a more a month or take on this risk".

colechristensen 7 hours ago||||
In university we had computer labs, I worked in the office that handled all of engineering computing. You paid the fee for engineering school and you got to use the labs. They had printers. We wanted printing to be free. This didn't mean "you get to take reams of blank paper home with you", it meant "you get as much printing as you reasonably need for academic purposes". Nobody cared if you printed your resume, fliers for your book club, or whatever, we weren't sticklers. Honestly we wanted to think about printers as little as possible.

But we'd always have a few people at the end of the semester print 493 blank pages using up all of their print quota they'd "paid for". No sir, you didn't pay for 500 pages of printing a semester, we'd let you print as much as you needed, we just had to put a quota in place to prevent some joker from wallpapering the lecture hall.

It was hard to express what we meant and "unlimited" didn't cut it.

Majromax 6 hours ago|||
You meant “reasonable,” but you did not apply reason. Situations such as this can be handled with a quota set at something like 150% of median use, but then extended upon a justified request. It can work in a lab where there’s a human touch, but it fails at million-user scale where even that level of human support is too expensive.
redsocksfan45 6 hours ago|||
[dead]
dreamcompiler 2 hours ago||||
Completely agree. This reminds me of the shady companies offering their employees "unlimited vacation" which translates to "you had better never take vacation because if you do it will be a major black mark against you."
littlecranky67 12 hours ago|||
Most home broadband providers offer unlimited network traffic.
hypercube33 11 hours ago|||
If they limit the rate of speed it's technically limited which really makes me wonder how they legally can say these things. I guess it means in a lot of cases it's like Comcast where they also limit the data a month perhaps but dang.
BobaFloutist 7 hours ago|||
They mean that they're not going to limit the total amount of data that you send/receive beyond the natural limit implied by the maximum rate.

When a movie subscription says unlimited movies, we know they're not suggesting that they can break the laws of time, just that they won't turn you away from a screening. It's pretty normal language, used to communicate no additional limit, which is relevant when compared to cell phone data plans (which are actually, in my opinion, fraudulent) that shunt you to a lower tier after a certain amount of usage.

dboreham 10 hours ago|||
In the language of marketing (in the USA at least) the word "unlimited" means "limited".
embedding-shape 12 hours ago||||
They offer "unlimited" where I live, not "unlimited*".
pixl97 10 hours ago||
I mean, in this universe we live in everything is limited somehow.

I do wish it was a word that had to be completely dropped from marketing/adverting.

For example there is not unlimited storage, hell the visible universe has a storage limit. There is not unlimited upload and download speed, and what if when you start using more space they started exponentially slowing the speed you could access the storage? Unlimited CPU time in processing your request? Unlimited execution slots to process your request? Unlimited queue size when processing your requests.

Hence everything turns into the mess of assumptions.

embedding-shape 9 hours ago||
> I mean, in this universe we live in everything is limited somehow.

Yes, indeed, most relevant in this case probably "time" and "bandwidth", put together, even if you saturate the line for a month, they won't throttle you, so for all intents and purposes, the "data cap" is unlimited (or more precise; there is no data cap).

pixl97 6 hours ago|||
In almost all services this tends to get an asterisk that says "unless your usage interferes with other users" which in itself is poorly defined. But typically means once their system gets closer to its usage limit, you're the first to get booted off the service.
embedding-shape 5 hours ago||
No ISP I've had in my adult life had such conditions, it truly is "Whatever you manage to do with the bandwidth we give you". I've done hundreds of TBs for months without any impact to my bandwidth (transferring ML datasets among other things), and I'm pretty sure a ISP in my country would break some law if they'd limit a typical broadband home connection based on data transfer quotas.
mcmcmc 7 hours ago|||
What? You are capped by bandwidth and time is its own limit. You are capped at the max bandwidth in your service contract multiplied by the length of the contract. A bandwidth cap has an implied data cap
bombcar 6 hours ago|||
The point is that you have access to a 100Mb/s connection, and your access to that connection is unlimited. It doesn't become a 10Mb/s connection at some point, and your access isn't cut off - there are no limits on your access.

Of course there are practical limits as you can't make your 100Mb/s connection into a gigabit one (ignoring that you can buy burstable in a datacenter, etc, etc).

Where unlimited falls down is when it refers to a endlessly consumable resource, like storage.

organsnyder 6 hours ago|||
Of course. You're always capped by rate. But you're not capped by the cumulative amount (other than as a function of rate and time).
ThatMedicIsASpy 11 hours ago||||
Doesn't help when you still need a VPN to get rid of Telekom/Vodafones abysmal peering
willis936 11 hours ago||||
And they have the necessary pipes to serve the rate they sell you 24/7.

Nobody has turned the moon into a hard drive yet.

littlecranky67 11 hours ago|||
> And they have the necessary pipes to serve the rate they sell you 24/7

I doubt they have those pipes, at least if every of their customers (or a sufficiently large amount) would actually make use of that.

Second question would be, how long they would allow you to utilize your broadband 24/7 at max capacity without canceling your subscription. Which leads back to the point the person I replied to was making: If you truly make use of what is promised, they cancel you. Hence it is not a faithful offer in the first place.

Dylan16807 10 hours ago||||
> Nobody has turned the moon into a hard drive yet.

Not important here because backblaze only has to match the storage of your single device. Plus some extra versions but one year multiplied by upload speed is also a tractable amount.

deno 11 hours ago|||
Since I know how many of those businesses are run I'll let you in on the very obvious secret: there’s zero chance they have enough uplink to accommodate everyone using 100% of their bandwidth at the same time, and probably much less than that.

Residential network access is oversold as everything else.

The only difference with storage is there’s a theoretical maximum on how much a single person can use.

But you could just as well limit backup upload speed for similar effect. Having something about fair use in ToS is really not that different.

cestith 8 hours ago|||
Residential ISPs don’t work financially unless you oversell peak time full-rate bandwidth. If you do things right, you oversell at a level that your customers don’t actually slow down. Even today, you won’t have 100% of customers using 100% of their full line rate 100% of the time.

Back in the late 1990s we could run a couple dozen 56k lines on a 1.544 Mbps backhaul. We could have those to the same extent today, but there’s still a ratio that works fine.

salawat 8 hours ago|||
Yes, yes. We know. The business environment can't be arsed to maintain it's own integrity by actually building out the capacity they want to charge for. Everyone hides behind statistical multiplexing until the actuarial pants shitting event occurs. Then it's bail out time, or "We're sorry. We used all the money for executive bonuses!"
deno 7 hours ago||
Building out for 100% of theoretical capacity makes no sense but you can still easily accommodate the small handful of power users with plenty to spare. Most ISPs will not drop or throttle users trying to get their money's worth if it’s fiber or similar. LTE of course that’s another thing.

That sort of horrible abuse only happens in areas where some provider has strict monopoly, but that’s an aberration and with Starlink’s availability there’s an upper bound nowadays.

LaGrange 9 hours ago||||
It’s not unlimited. The limit might be very high these days, but it’s at most bandwidth times duration. And while that sounds trivial, it does mean they aren’t selling you an infinity of a resource.
mikepurvis 11 hours ago|||
Unsure if sarcastic but most ISPs will throttle and "traffic" long before you use anything close to <bandwidth rating> times <seconds in a month>.
dmantis 11 hours ago|||
I've been running RPI-based torrent client 24/7 in several countries and never experienced that. Eats a few TBs per month, not the full line, but pretty decent amount. I guess it really depends on the country.
gambiting 10 hours ago||
I'm in the UK with Virgin Media on their 1Gbps package, going through multiple TB a month and I'm yet to be throttled in any way.
Dylan16807 8 hours ago||
Well, multiple TB isn't close to your bandwidth rating. It only takes 2% of your connection in a single direction to hit 6TB a month.
gambiting 4 hours ago||
Ha, yes I suppose that's correct.
Spooky23 11 hours ago||||
I’ve used Spectrum and their predecessors since the 90s. Never ran into this, although the upstream speeds are ridiculously slow, and they used to force Netflix traffic to an undersized peer circuit.
embedding-shape 11 hours ago|||
I'm unsure if you're sarcastic or not, never have I've used a ISP that would throttle you, for any reason, this is unheard of in the countries I've lived, and I'm not sure many people would even subscribe to something like that, that sounds very reverse to how a typical at-home broadband connection works.

Of course, in countries where the internet isn't so developed as in other parts of the world, this might make sense, but modern countries don't tend to do that, at least in my experience.

lelandfe 11 hours ago||
Alas, "isn't so developed" applies to the US: https://arstechnica.com/tech-policy/2020/06/cox-slows-intern...

My parents have gotten hit by this. Dad was downloading huge video files at one point on his WiFi and his ISP silently throttled him.

A common term is "data cap": https://en.wikipedia.org/wiki/Data_cap

embedding-shape 10 hours ago||
> Alas, "isn't so developed" applies to the US

Wow, I knew that was generally true, didn't know it was true for internet access in the US too, how backwards...

> A common term is "data cap": https://en.wikipedia.org/wiki/Data_cap

I think most are familiar with throttling because most (all?) phone plans have some data cap at one point, but I don't think I've heard of any broadband connections here with data caps, that wouldn't make any sense.

bombcar 5 hours ago|||
Data caps are just documenting the reality that ISPs oversubscribe - if they sell a hundred 1Gb/s connections to a neighborhood, it's highly unlikely they're peering that neighborhood onto the Internet at large at 100Gb/s. I don't know what the current standard is, but in the past it's been 10/100 to 1 - so a hundred 1Gb/s connections might be sharing 1-10Gb/s of uplink; and if usage starts to saturate that they need a way of backing off that is "fair" - data caps are one of the ways they inform the customer of such.

I've seen it with my new fiber rollout - every single customer no matter their purchased speed had 1Gb up and down - as more customers came online and usage became higher, they're not limiting anyone, but you get closer to your advertised rate - but my upload is still faster than my download because most of my neighborhood is downloading, few are uploading.

cestith 8 hours ago||||
I have 5 Gbps symmetric at home. I and my fiancee both work from home, so our backup fiber connection from another provider is 2 Gbps. We can also both tether to cell phones if necessary. We can get 5G home wireless Internet here, too, and we might ditch our 2 Gbps line in favor of that as a backup. We moved from Texas back home to Illinois last year, and one of the biggest considerations was who had service at what tiers due to remote work. Some of the houses we looked at in the same three-county area in the Chicago suburbs didn’t even have 5G home available (not from AT&T, Verizon, or T-Mobile anyway).

My parents have 5G wireless home as their primary connection, and that was only introduced in their area a couple of years ago. Before that, they could get dial-up, 512 kbps wireless with about a $1000 startup cost, ISDN (although the phone company really didn’t want to sell it to them), Starlink, or HughesNet. The folks across the asphalt road from them had 20 Mbps Ethernet over power lines years ago, and that’s now I think 250 Mbps. It’s a different power company, though, so they aren’t eligible.

Around 80% of the US population lives in large urban areas. The other 20% of the population range from smaller towns to living many kilometers from any town at all. There’s a lot of land in the US.

lelandfe 6 hours ago||
Here in dense NYC, most apartments I've lived in have but a single ISP available. It's common to hunt for apartments by searching the address on service maps.

I'm pretty sure one landlord was cut in by his ISP, as he skipped town when I tried to ask about getting fiber, and his office locked their door and drew their shades when I went there with a technician on two occasions. The final time, we got there before they opened and the woman ran into the office and slammed the door on us.

lelandfe 8 hours ago|||
Our ISPs conspire to avoid competition (AKA "overbuilding") and so stuff like this just festers. It's truly a shame.
rsync 7 hours ago|||
"I guess the problem with Backblaze's business model with respect to Backblaze Personal is that it is "unlimited"."

The new and very interesting problem with their business model is that drive prices have doubled - and in some cases, more than doubled - in the last 12 months.

Backblaze has a lot of debt and at some point the numbers don't make sense anymore.

Dylan16807 2 hours ago|||
Is it that bad? When I look at the prices of new drives on amazon I mostly see increases just under 50%. I think used went up more but that's not affecting backblaze as much.
tombert 6 hours ago|||
Yeah, I found that out recently when I had to purchase a new 16TB drive because of them in my RAID died recently. I bought the hard drive used about three years ago for about $130. To replace it I had to shop around and I ended up paying about $270 and I think that was considered a decent deal right now.

Oh well, I guess this is why we're given two kidneys.

ricardobeat 11 hours ago|||
It’s funny that the same person asking for linux support would complain about B2 “not being for home users”. I sync my own backups to B2 and would set that up over installing linux any day of the week! It’s extremely easy.
SomeHacker44 10 hours ago||
What software/workflow do you use for this Linux to B2 backup please?
its-summertime 10 hours ago|||
Restic + rclone personally, with a wrapper script to glue things together nicely
dicytea 6 hours ago||
What's the advantage of additionally using rclone vs. just restic?
Zetaphor 10 hours ago|||
rclone on a cron job
lousken 10 hours ago|||
Yea, that's pretty shady. Either don't call your service unlimited or bump up the prices so you can survive occasional datahoarder, called them out on it many years ago.
matheusmoreira 8 hours ago|||
I actually emailed them years ago about it. Asked them point blank what'd happen if I dumped 20+ TB of encrypted, undeduplicable backups onto their storage servers. They actually replied that there'd be no problem, but I didn't buy it. Not at all surprised to see this now.
monooso 12 hours ago|||
Unlimited means without limits or restrictions.

If a company uses the word unlimited to describe their service, but then attempts to weasel out of it via their T&Cs, that doesn't constitute a disagreement over the meaning of the word unlimited. It just means the company is lying.

swiftcoder 11 hours ago|||
From a philosophical standpoint, I agree, but it terms of service providers "unlimited" has always pretty much always been synonymous with "unmetered" (i.e. we don't charge you for traffic, but we will still throttle you if you are affecting service reliability for other customers)
conductr 8 hours ago|||
Sorry but unlimited has never meant unrestricted. TOCs always have restrictions. If it were unrestricted it would be used for all kinds of illegal stuff they don’t want on their servers, child pr0n and whatnot. They can’t legally offer a service like this without restrictions as they operate within an existing set of laws.

Unlimited however, they can offer. I don’t see how people get into mental block of thinking something is nefarious when a company offers you unlimited hosting or data. Yes, they know it’s impossible if everyone took full advantage of that. They also know most people won’t and so they don’t have to spend time worrying about it. It’s a simple actuarial exercise to work out the pricing that covers the use of your users.

Back in the early 2000s I ran a web hosting service that was predominantly a LAMP stack shared hosting environment. It had several unlimited plans and they were easy to estimate/price. The only times I had an issue of supporting a heavy user, it would turn out they were doing something unrestricted. Back then, it was usually something pron or mp3 related. So the user would get kicked off for that. I didn’t have any issues with supporting the usage load if it was within TOS. The margins were so high it was almost impossible to find a user that could give me any trouble from an economic standpoint.

ape4 11 hours ago|||
Why don't they charge by the Gigabyte
danpalmer 10 hours ago|||
Because approximately no one wants that. Anyone who does already uses S3 etc.
cowboylowrez 9 hours ago||||
I use them for the b2 bucket style storage where this happens. Its expensive per gig compared to the cost of a working personal unlimited desktop account. I like to visit their reddit page occasionally and its a constant stream of desktop client woes and stories of restoring problems and any time b2 is mentioned its like "but muh 50 terabytes" lol
renata 9 hours ago||
It's cheaper if you have multiple computers with normal amounts of data though. My whole family is on my B2 account (Duplicati backing up eight computers each to a separate bucket), and it's $10/month.
ahofmann 10 hours ago|||
They do, it's called B2 and is another product of them.
bitfilped 5 hours ago||
When it comes to storage "unlimited" to me means a promise to be broken at some random point in the future. I'll never use a service that claims unlimited anything over having an actual cost model. Companies that charge by what you use have actually given consideration to the cost of doing business and have priced that in already.
rafabulsing 2 hours ago||
I've long thought that words such as "unlimited", "infinite" and so on should be legally banned from marketing, or at the very least their use should be heavenly regulated.

_Nothing_ is actually infinite. Everything has limits.

"But X terabytes is functionally infinite for 99.99% of users"

Cool, then advertise that you offer Xtb of storage. Infinite means infinite, and if you offer anything less than that - and you do - then you shouldn't be allowed to say otherwise.

nstj 9 hours ago||
As an FYI you can recover from force pushes to GitHub using the GitHub UI[0] or their API[1]. And if you force push to one of your own machines you can use the reflog[2]. [0]: https://stackoverflow.com/a/78872853 [1]: https://stackoverflow.com/a/48110879 [2]: https://stackoverflow.com/a/24236065
bombcar 5 hours ago|
And as a double FYI this means a force push does not permanently delete sensitive data! Beware. Rotate that API key, even if it's a pain in the arse.
noirscape 12 hours ago||
I can understand in theory why they wouldn't want to back up .git folders as-is. Git has a serious object count bloat problem if you have any repository with a good amount of commit history, which causes a lot of unnecessary overhead in just scanning the folder for files alone.

I don't quite understand why it's still like this; it's probably the biggest reason why git tends to play poorly with a lot of filesystem tools (not just backups). If it'd been something like an SQLite database instead (just an example really), you wouldn't get so much unnecessary inode bloat.

At the same time Backblaze is a backup solution. The need to back up everything is sort of baked in there. They promise to be the third backup solution in a three layer strategy (backup directly connected, backup in home, backup external), and that third one is probably the single most important one of them all since it's the one you're going to be touching the least in an ideal scenario. They really can't be excluding any files whatsoever.

The cloud service exclusion is similarly bad, although much worse. Imagine getting hit by a cryptoworm. Your cloud storage tool is dutifully going to sync everything encrypted, junking up your entire storage across devices and because restoring old versions is both ass and near impossible at scale, you need an actual backup solution for that situation. Backblaze excluding files in those folders feels like a complete misunderstanding of what their purpose should be.

stebalien 10 minutes ago||
I've actually spent some time debugging why git causes so many issues with the backup software I use (restic).

Ironically, I believe you have it backwards: pack files, git's solution to the "too many tiny files" problem, are the issue here; not the tiny files themselves.

In my experience, incremental backup software works best with many small files that never change. Scanning is usually just a matter of checking modification times and moving on. This isn't fast, but it's fast enough for backups and can be optimized by monitoring for file changes in a long-running daemon.

However, lots of mostly identical files ARE an issue for filesystems as they tend to waste a lot of space. Git solves this issue by packing these small objects into larger pack files, then compressing them.

Unfortunately, it's those pack files that cause issues for backup software: any time git "garbage collects" and creates new pack files, it ends up deleting and creating a bunch of large files filled with what looks like random data (due to compression). Constantly creating/deleting large files filled with random data wreaks havoc on incremental/deduplicating backup systems.

adithyassekhar 11 hours ago|||
I don’t think this is the right way to see this.

Why should a file backup solution adapt to work with git? Or any application? It should not try to understand what a git object is.

I’m paying to copy files from a folder to their servers just do that. No matter what the file is. Stay at the filesystem level not the application level.

noirscape 11 hours ago|||
I'm not saying Backblaze should adapt to git; the issue isn't application related (besides git being badly configured by default; there's a solution with git gc, it's just that git gc basically never runs).

It's that to back up a folder on a filesystem, you need to traverse that folder and check every file in that folder to see if it's changed. Most filesystem tools usually assume a fairly low file count for these operations.

Git, rather unusually, tends to produce a lot of files in regular use; before packing, every commit/object/branch is simply stored as a file on the filesystem (branches only as pointers). Packing fixes that by compressing commit and object files together, but it's not done by default (only after an initial clone or when the garbage collector runs). Iterating over a .git folder can take a lot of time in a place that's typically not very well optimized (since most "normal" people don't have thousands of tiny files in their folders that contain sprawled out application state.)

The correct solution here is either for git to change, or for Backblaze to implement better iteration logic (which will probably require special handling for git..., so it'd be more "correct" to fix up git, since Backblaze's tools aren't the only ones with this problem.)

masfuerte 10 hours ago|||
7za (the compression app) does blazingly fast iteration over any kind of folder. This doesn't require special code for git. Backblaze's backup app could do the same but rather than fix their code they excluded .git folders.

When I backup my computer the .git folders are among the most important things on there. Most of my personal projects aren't pushed to github or anywhere else.

Fortunately I don't use Backblaze. I guess the moral is don't use a backup solution where the vendor has an incentive to exclude things.

toast0 5 hours ago||
IMHO, you can't do blazingly fast iteration over folders with small files in Windows, because every open is hooked by the anti-virus, and there goes your performance.
noirscape 2 hours ago||
Not just antivirus, there's also file locking.

Windows has a much harsher approach to file locking than Linux and backup software like BackBlaze absolutely should be making use of it (lest they back up files that are being modified while they back them up), but that also means that the software effectively has to ask the OS each time to lock the file, then release the lock when the software is done with it. With a large amount of files, that does stack up.

Linux file locking is to put it mildly, deficient. Most software doesn't even bother acquiring locks in the first place. Piling further onto that, basically nobody actually uses POSIX locks because the API has some very heavy footguns (most notably, every lock on a file is released whenever any close() for that file is called, even if another component of the same process is also having a second lock open). Most Linux file locks instead work on the honor system; you create a file called filename.lock in the same directory as the file you're working on, and then any software that detects the filename.lock file exists should stop reading the file.

Nobody using file locks is probably the bigger reason why Linux chokes less on fast iteration than Windows, given that Windows is slow with loads of files even when you aren't running a virus scanner.

NetMageSCW 8 hours ago|||
Actually once the initial backup is done there is no reason to scan for changes. They can just use a Windows service that tells them when any file is modified or created and add that file to their backup list.
Saris 8 hours ago|||
Backblaze offers 'unlimited' backup space, so they have to do this kind of thing as a result of that poor marketing choice.
conductr 7 hours ago|||
No they don’t. They just have to price the product to reflect changing user patterns. When backblaze started, it was simply “we back up all the files on your drive” they didn’t even have a restore feature that was your job when you needed it. Over time they realized some user behavior changed, these Cloud drives where a huge data source they hadn’t priced in, git gave them some problems that they didn’t factor in, etc. The issue is there solution to dealing with it is to exclude it and that means they’re now a half baked solution to many of their users, they should have just changed the pricing and supported the backup solution people need today.
adithyassekhar 3 hours ago|||
If they must scam, shouldn’t they be deduplicating on the server rather than the client?
Ajedi32 7 hours ago|||
FWIW some other people in this thread are saying the article is wrong about .git folders not being backed up: https://news.ycombinator.com/item?id=47765788

That's a really important fact that's getting buried so I'd like to highlight it here.

rmccue 11 hours ago|||
I think it's understandable for both Backblaze and most users, but surely the solution is to add `.git` to their default exclusion list which the user can manage.
maalhamdan 12 hours ago|||
I think they shouldn't back up git objects individually because git handles the versioning information. Just compress the .git folder itself and back it up as a single unit.
willis936 11 hours ago|||
Better yet, include dedpulication, incremental versioning, verification, and encryption. Wait, that's borg / restic.

This is a joke, but honestly anyone here shouldn't be directly backing up their filesystems and should instead be using the right tool for the job. You'll make the world a more efficient place, have more robust and quicker to recover backups, and save some money along the way.

pkaeding 12 hours ago|||
This is a good point, but you might expect them to back up untracked and modified files in the backup, along with everything else on your filesystem.
pixl97 9 hours ago||
Eh, you really shouldn't do that for any kind of file that acts like a (an impromptu) database. This is how you get corruption. Especially when change information can be split across more than one file.
pkaeding 7 hours ago||
Sorry, what are you saying shouldn't be done? Backing up untracked/modified files in a bit repo? Or compressing the .git folder and backing it up as a unit?
pixl97 6 hours ago||
> Backing up untracked/modified files in a bit repo?

This. It's best to do this in an atomic operation, such as a VSS style snapshot that then is consistent and done with no or paused operations on the files. Something like a zip is generally better because it takes less time on the file system than the upload process typically takes.

ciupicri 12 hours ago|||
> If it'd been something like an SQLite database instead (just an example really)

See Fossil (https://fossil-scm.org/)

P.S. There's also (https://www.sourcegear.com/vault/)

> SourceGear Vault Pro is a version control and bug tracking solution for professional development teams. Vault Standard is for those who only want version control. Vault is based on a client / server architecture using technologies such as Microsoft SQL Server and IIS Web Services for increased performance, scalability, and security.

rcxdude 11 hours ago|||
It's probably primarily because Linus is a kernel and filesystem nerd, not a database nerd, so he preferred to just use the filesystem which he understood the performance characteristics of well (at least on linux).
grumbelbart2 11 hours ago|||
Git packs objects into pack-files on a regular basis. If it doesn't, check your configuration, or do it manually with 'git repack'.
noirscape 11 hours ago||
I decided to look into this (git gc should also be doing this), and I think I figured out why it's such a consistent issue with git in particular. Running git gc does properly pack objects together and reduce inode count to something much more manageable.

It's the same reason why the postgres autovacuum daemon tends to be borderline useless unless you retune it[0]: the defaults are barmy. git gc only runs if there's 6700 loose unpacked objects[1]. Most typical filesystem tools tend to start balking at traversing ~1000 files in a structure (depends a bit on the filesystem/OS as well, Windows tends to get slower a good bit earlier than Linux).

To fix it, running

> git config --global gc.auto 1000

should retune it and any subsequent commit to your repo's will trigger garbage collection properly when there's around 1000 loose files. Pack file management seems to be properly tuned by default; at more than 50 packs, gc will repack into a larger pack.

[0]: For anyone curious, the default postgres autovacuum setting runs only when 10% of the table consists of dead tuples (roughly: deleted+every revision of an updated row). If you're working with a beefy table, you're never hitting 10%. Either tune it down or create an external cronjob to run vacuum analyze more frequently on the tables you need to keep speedy. I'm pretty sure the defaults are tuned solely to ensure that Postgres' internal tables are fast, since those seem to only have active rows to a point where it'd warrant autovacuum.

[1]: https://git-scm.com/docs/git-gc

LetTheSmokeOut 9 hours ago|||
I needed to use

> git config --global gc.auto 1000

with the long option name, and no `=`.

bombcar 5 hours ago||||
I love nothing more than running strange git commands found in HN comments.

Let's ride the lightning and see if it does anything.

Dylan16807 9 hours ago|||
A few thousand files shouldn't be a problem to a program designed to scan entire drives of files. Even in a single folder and considering sloppy programs I wouldn't worry just yet, and git's not putting them in a single folder.
yangm97 11 hours ago|||
You don’t see ZFS/BTRFS block based snapshot replication choking on git or any sort of dataset. Use the right job for the tool or something.
KingMachiavelli 5 hours ago||
They 100% should have communicated this change, absolutely unacceptable to change behavior without an extremely visible warning.

However, backing up these kinds of directories has always been ill-defined. Dropbox/Google Drive/etc. files are not actually present locally - at least not until you access the file or it resides to cache it. Should backup software force you to download all 1TB+ of your cloud storage? What if the local system is low on space? What if the network is too slow? What if the actually data is in an already excluded %AppData% location.

Similar issue with VCS, should you sync changes to .git every minute? Every hour? When is .git in a consistent state?

IMO .git and other VCS should just be synced X times per day and it wait for .git to be unchanged for Y minutes before syncing it. Hell, I bet Claude could write a special Git aware backup script.

But Google Drive and Dropbox mount points are not real. It’s crazy to expect backup software to handle that unless explicitly advertised.

mzmzmzm 4 hours ago|
Dropbox and GDrive desktop clients can be configured to sync files to a local directory. Backing them up with an additional platform would probably need some sort of logic like you described for VCS.
klausa 13 hours ago||
Exclusions are one thing, but I've had Backblaze _fail to restore a file_. I pay for unlimited history.

I contacted the support asking WTF, "oh the file got deleted at some point, sorry for that", and they offered me 3 months of credits.

I do not trust my Backblaze backups anymore.

orr721 9 hours ago||
I had similar experience as well. They upgraded their client and server software something like 5 years ago which put forward different restrictions on character set used for password. I have used a special character which was no longer allowed. When I needed to restore files after disk failure I could not log in either in the app or on the website. The customer service was useless -- we are sorry, your fault. I have lost 1 TB of personal photos due to this as a paying customer. Never trust Backblaze.
nayhel89 11 hours ago|||
I have the same experience with Backblaze. 3 years ago I tried to restore my files from Backblaze, using their desktop client.

First thing I noticed is that if it can't download a file due to network or some other problem then it just skips it. But you can force it to retry by modifying its job file which is just an SQLite DB. Also it stores and downloads files by splitting them into small chunks. It stores checksums of these chunks, but it doesn't store the complete checksum of the file, so judging by how badly the client is written I can't be sure that restored files are not corrupted after the stitching.

Then I found out that it can't download some files even after dozens of retries because it seems they are corrupted on Backblaze side.

But the most jarring issue for me is that it mangled all non-ascii filenames. They are stored as UTF-8 in the DB, but the client saves them as Windows-1252 or something. So I ended up with hundreds of gigabytes of files with names like фикац, and I can't just re-encode these names back, because some characters were dropped during the process.

I wanted to write a script that forces Backblaze Client to redownload files, logs all files that can't be restored, fixes the broken names and splits restored files back into chunks to validate their checksums against the SQLite DB, but it was too big of a task for me, so I just procrastinated for 3 years, while keeping paying monthly Backblaze fees because it's sad to let go of my data.

I wonder if they fixed their client since then.

ValentineC 1 hour ago|||
> But the most jarring issue for me is that it mangled all non-ascii filenames. They are stored as UTF-8 in the DB, but the client saves them as Windows-1252 or something. So I ended up with hundreds of gigabytes of files with names like фикац, and I can't just re-encode these names back, because some characters were dropped during the process.

> I wanted to write a script that forces Backblaze Client to redownload files, logs all files that can't be restored, fixes the broken names and splits restored files back into chunks to validate their checksums against the SQLite DB, but it was too big of a task for me, so I just procrastinated for 3 years, while keeping paying monthly Backblaze fees because it's sad to let go of my data.

Filenames are probably the most valuable of metadata for them to mangle. I value them as much as I do file creation/modification times. A backup program is dead to me if they mess up either of these.

I think it should be trivial for you to pipe your request into Claude now, and get them to write a quick script. Hope that'll free you from Backblaze for good!

toraway 2 hours ago||||

  > I wonder if they fixed their client since then.
They have not. I spent more than a week trying to restore a little less than 2 TB backup because the client would just freeze at the last few % every time. I ended up having to break the restore into 200GB chunks on the web client and download and restore manually which was extremely frustrating and made me despise their (required) Windows client.
PunchyHamster 5 hours ago|||
And they talked so much about how great redundancy they have on backend. I guess they don't count 404's
willis936 11 hours ago|||
Do you have any more details? This is a pretty big deal. The differentiators between Backblaze and Hetzner mostly boil down to this kind of thing supposedly not being possible.
klausa 11 hours ago||
I’m on my phone so forgive the formatting, but here’s my entire support exchange:

- - -

Hey, I tried restoring a file from my backup — downloading it directly didn't work, and creating a restore with it also failed – I got an email telling me contract y'all about it.

Can you explain to me what happened here, and what can I do to get my file(s?) back?

- - -

Hi Jan,

Thanks for writing in!

I've reached out to our engineers regarding your restore, and I will get back to you as soon as I have an update. For now, I will keep the ticket open.

- - -

Hi Jan,

Regarding the file itself - it was deleted back in 2022, but unfortunately, the deletion never got recorded properly, which made it seem like the file still existed.

Thus, when you tried to restore it, the restoration failed, as the file doesn't actually exist anymore. In this case, it shouldn't have been shown in the first place.

For that, I do apologize. As compensation, we've granted you 3 monthly backup credits which will apply on your next renewal. Please let me know if you have any further questions.

- - -

That makes me even more confused to be honest - I’ve been paying for forever history since January 2022 according to my invoices?

Do you know how/when exactly it got deleted?

- - -

Hi Jan,

Unfortunately, we don't have that information available to us. Again, I do apologize.

- - -

I really don’t want to be rude, but that seems like a very serious issue to me and I’m not satisfied with that response.

If I’m paying for a forever backup, I expect it to be forever - and if some file got deleted even despite me paying for the “keep my file history forever” option, “oh whoops sorry our bad but we don’t have any more info” is really not a satisfactory answer.

I don’t hold it against _you_ personally, but I really need to know more about what happened here - if this file got randomly disappeared, how am I supposed to trust the reliability of anything else that’s supposed to be safely backed up?

- - -

Hi Jan,

I'll inquire with our engineers tomorrow when they're back in, and I'll update you as soon as I can. For now, I will keep the ticket open.

- - -

Appreciate that, thank you! It’s fine if the investigation takes longer, but I just want to get to the bottom of what happened here :)

- - -

Hi Jan,

Thanks for your patience.

According to our engineers and my management team:

With the way our program logs information, we don't have the specific information that explains exactly why the file was removed from the backup. Our more recent versions of the client, however, have vastly improved our consistency checks and introduced additional protections and audits to ensure complete reliability from an active backup.

Looking at your account, I do see that your backup is currently not active, so I recommend running the Backblaze installer over your current installation to repair it, and inherit your original backup state so that our updates can check your backup.

I do apologize, and I know it's not an ideal answer, but unfortunately, that is the extent of what we can tell you about what has happened.

- - -

I gave up escalating at this point and just decided these aren’t trusted anymore.

The files in question are four year old at this point so it’s hard for me conclusively state, so I guess there might be a perfect storm of that specific file being deleted because it was due to expire before upgraded to “keep history forever”, but I don’t think it’s super likely, and I absolutely would expect them to have telemetry about that in any case.

If anyone from Backblaze stumbles upon it and wants to escalate/reinvestigate, the support ID is #1181161.

jeremyjh 8 hours ago|||
This reminds me of the Seinfeld riff on car rental reservations. Anyone can make a backup. The important part is holding the backup. If Backblaze doesn’t always do that then it is practically worthless to everyone.
notrealyme123 11 hours ago||||
Thank you for sharing this. A non-persistent backup service is on the same level as a zombie-insurance provider.
rincebrain 8 hours ago||||
This seems absurd from a company offering backups as a service.

Especially if they allow them restoring all your data onto a drive and shipping it to you, they pretty clearly should have enough information available to them to test restorations of data, and the number of times I've heard that failure mode ("oh, we didn't track deletions well enough, so we only found out we deleted it when you tried restoring"), plus them saying they have made improvements to avoid this exact failure mode in newer client versions, makes me think they should have enough reports to investigate it.

...which makes me wonder if they did, and decided they would go bankrupt if they told people how much data they lost, so they decided to bet on people not trying restores on a lot of the lost data.

benjiro3000 10 hours ago|||
[dead]
ValentineC 2 hours ago|||
> Exclusions are one thing, but I've had Backblaze _fail to restore a file_. I pay for unlimited history.

> I contacted the support asking WTF, "oh the file got deleted at some point, sorry for that", and they offered me 3 months of credits.

This happened to me with CrashPlan for Windows many years ago, because of some Volume Shadow Copy Service thing. I noped out of there right after.

christoff12 12 hours ago||
wut
AegirLeet 12 hours ago||
At some point, Backblaze just silently stopped backing up my encrypted (VeraCrypt) drives. Just stopped working without any announcement, warning or notification. After lots of troubleshooting and googling I found out that this was intentional from some random reddit thread. I stopped using their backup service after that.
peteforde 10 hours ago||
Weirdly, reading this had the net impact of me signing up to Backblaze.

I had no idea that it was such a good bargain. I used to be a Crashplan user back in the day, and I always thought Backblaze had tiered limits.

I've been using Duplicati to sync a lot of data to S3's cheapest tape-based long term storage tier. It's a serious pain in the ass because it takes hours to queue up and retrieve a file. It's a heavy enough process that I don't do anything nearly close to enough testing to make sure my backups are restorable, which is a self-inflicted future injury.

Here's the thing: I'm paying about $14/month for that S3 storage, which makes $99/year a total steal. I don't use Dropbox/Box/OneDrive/iCloud so the grievances mentioned by the author are not major hurdles for me. I do find the idea that it is silently ignoring .git folders troubling, primarily because they are indeed not listed in the exclusion list.

I am a bit miffed that we're actively prevented from backing up the various Program Files folders, because I have a large number of VSTi instruments that I'll need to ensure are rcloned or something for this to work.

ValentineC 2 hours ago||
> Here's the thing: I'm paying about $14/month for that S3 storage, which makes $99/year a total steal. I don't use Dropbox/Box/OneDrive/iCloud so the grievances mentioned by the author are not major hurdles for me. I do find the idea that it is silently ignoring .git folders troubling, primarily because they are indeed not listed in the exclusion list.

A big difference here is that Backblaze only keeps deleted/changed files for 30 days. Deleted files can go unnoticed for some time, especially if done by a malicious app or ignorant AI.

I'd pay that extra few dollars for peace of mind.

peteforde 16 minutes ago||
Thanks so much! That is super valuable input.

I do notice in their GUI that they offer a >30 memory for extra $:

30-Day Version History (Current) Included

1-Year Version History

Forever Version History $.006/GB/Month for versions changed or deleted more than 1 year ago

They are not giving much information here at all. The above is not a pasting artifact; their page literally doesn't give any indication of how they price the 1-year history. Presumably it's not just as simple as click to 12x your retention for free.

Meanwhile, it's even more unclear whether that $.006/GB is assessed for change deltas or for the total file size. Indeed, it's not clear if it's assessed against your entire fileset or just files that changed.

I'll have to email them, I guess.

justin66 6 hours ago|||
"Maybe they're only incompetent in the ways that have been enumerated in this blog post" does not seem like much of a sales pitch. Baffling.
peteforde 5 hours ago||
I'm happy to pay an annual fee for a one-size fits all approach that I don't have to think about. I read the post and I'm just saying that his blockers are not blockers for me.

I would ask you: what is the better alternative? That's not a rhetorical question; they don't have my credit card details for another two weeks.

malfist 1 minute ago||
This whole post is saying that you _do_ have to think about it.
tonymet 3 hours ago|||
You lose a bit of control. With S3 you can preprocess (transform, index, filter, downcode, etc) before storing. You can index metadata in place (names, sizes, metadata) for low-cost searching.

As for testing recovery, you can validate file counts, sizes + checksums without performing recovery.

A few shell scripts give you the power of advanced enterprise backup, whereas backblaze only supports GUI restores.

NetMageSCW 8 hours ago||
If you don’t really want backups you can save a lot more money by not signing up for Backblaze.
peteforde 5 hours ago||
Are they known for accidentally erasing your backups?

I get that this is not a restorable image, but for $100 a year I'm not expecting that.

mcherm 12 hours ago|
Some companies are in the business of trust. These companies NEED to understand that trust is somewhat difficult to earn, but easy to lose and nearly IMPOSSIBLE to regain. After reading this article I will almost certainly never use or recommend Backblaze. (And while I don't use them currently, they WERE on the list of companies I would have recommended due to the length of their history.)
baal80spam 4 hours ago|
> trust is somewhat difficult to earn, but easy to lose and nearly IMPOSSIBLE to regain

Eh, I don't agree. Case in point: Microsoft.

Or in other words: a sucker is born every minute.

More comments...