Hacker News new | past | comments | ask | show | jobs | submit login

People who don't use and who don't understand the differences between tape and disk have been saying tape is dying for years.

The most common incorrect assertion is that drives don't fail THAT often, so why not just back up to drives and put them on a shelf? But to be generally safe, all data would have to be on at least two drives, which changes the cost noticeably.

Older, used LTO drives are actually quite affordable, so I use LTO-4 and LTO-5 tapes to back up all sorts of things for both myself and for my clients. It's surprisingly easy and inexpensive, and pulling files back off is as simple as running pax.




One important thing is (at least a few years back) we realised that HDDs fail at an alarming rate (> 80%) if you disconnect them and leave them for a few years. Either that or someone brought a gamma ray source through our lab. Keeping the disks hot was a different story though.


Interesting, I've heard of issues with poor shelf-life for SSDs, but not HDDs. Any data/articles on this topic or is it strictly anecdotal?


I am not aware of any large-scale study, but poor HDD reliability in cold storage corresponds to my experience too.

I have archived data (almost a couple hundred TB in double copies) on more than 60 HDDs, about half from WD and half from Seagate, from several HDD generations, i.e. with capacities of 2 TB, 3 TB, 4 TB, 6 TB and 8 TB, most being from the 4 up to 8 TB generations. All HDDs were more expensive WD and Seagate models (i.e. with longer warranty times), not their cheapest consumer HDDs.

All files were checksummed for error detection. When the HDD content was transferred to tapes after some 3 to 6 years since the initial archival, almost all HDDs had a few errors, which sometimes were not reported by the drives even when they were detected by the file checksums.

Only the fact that all files were stored on at least 2 HDDs has prevented data loss, because even when both HDDs had errors they were in different locations.


I've never used tape on a regular basis but when I have I've had terrible experiences with low end tapes. I got kicked out of the computer center at my Alma Mater because I created a bunch of newsgroups. They (said that they) handed me my files on a SunTape that seemed to be empty when I tried to restore it later.

A few years later I got a QIC tape for my 486 PC and had similar experiences. There was even the time I tried to restore a several kb configuration file from a Tivoli tape robot and was quoted that it would take 18 hours and figured I could recreate it in much less time.

If I did it all the time I'm sure I would do better, but 2 out of 3 times or so I had tape backups fail on me, contrast that to almost always being able to read HDDs, even if there is some bit rot.


> 2 copies, checksummed

Would using something like ZFS done effectively the same thing? Not being critical. I genuinely want to know more.


For SSDs, per JEDEC, Client SSDs have to retain data for 1 year at 30C, and Enterprise SSDs have to retain data for 3 months at 40C:

* https://www.jedec.org/sites/default/files/Alvin_Cox%20[Compa...

I'm sure most last longer, but as a CYA I wouldn't necessarily want to rely on that assumption.


I'd expect the various bearings to seize if left unused for long enough.


Bearings seizing up is the example I usually hear.


Most revive after being left in the fridge for some time. The bits and pieces contract at different rates, so cooling (or, I must assume, heating) is enough to get them unstuck.

Many drives I did that with lived happy afterlives after being revived, some having been brought back from the dead more than once.


With HDDs it's usually that the bearings dry up and seize if they don't spin regularly.


Expand on keeping-them-hot story....

The biggest expense in TAPE storage has always been the tape-changing-robot. (and if you ever worked with SUN HW -- the software was super expensive and the UI/UX of the backup software (logato? I cant recall the name -- was a PoS (not point of sale) and I hated it...

However, with that said, the price/GB/TB for tape is very cheap...

What would be a good idea would be to use tape duplication robots to take one tape to N tapes on a fairly regulated basis (like every 2,3,5 years) where you read and copy to new such that the medium (the cheap tape) is cycled through...


I have bunch of HDD's starting from more than 15+ years ago. I was recently checking and all of them are readable. Due tp the total number of drives I would notice 80% failure rate.

They're not used as a backup, rather I would start with the data disk and as the price goes down and my needs grow the drive would be replaced with the higher capacity one. And there is an extra set of drives (copy of those) I always back them up.


Is the 80% sampling a variety of manufacturers and batches, or a single batch of drives that were all ordered at the same time from the same manufacturer? If the latter, I'd suspect it's more of a realization that you got a flawed batch of drives (which is a real risk to manage, but not universal).


It's a variety of sizes and companies spanning 8 years. And we were talking about 100ish drives. Basically we realised our entire lab backup was useless.


I think a data centre company posted some findings on this recently.


It depends on the scale, doesn't it?

My issue with tape is that unless your operation is big enough to justify a robotic tape library as a budget line item with support contracts and all, then you're down to paying someone (or eating the cost yourself) to physically swap tapes which is much more expensive and boring than deploying a ZFS server with gobs of disk that once configured, just sits in the rack and does its job quietly.


I think you’re dead on. I smaller scales, it’s cheaper just to keep more disks running. But once you get to scale, tape is great for archival work.

The trick I’ve always found is to figure out where you are on that inflection point. And it’s hard. Is 1PB enough to justify tape? (Which seems like a crazy question to me - I remember having megabyte sized tapes).


When I did the calculations a few years ago, break-even was somewhere around 150 TB, and I don’t think it’s changed too much since then. This is just considering the cost of drives and media. Obviously, the real inflection point is going to be different depending on all sorts of factors that may be specific to your situation and your priorities.

Usability isn’t something you can ignore, but it’s not like hard drives are perfect either—do you buy a big NAS / SAN setup and plug drives into it? Will it get full? And tape has the advantage that it’s completely immune, out of the box, to ransomware.

I think there are four cases that really scream for tape.

1. Data hoarders, who just want to store as much data as possible for the cost. There’s an r/DataHoarder subreddit if you’re curious about these people.

2. Archivists, who want to store lots of data long-term. Tape is a lot easier to store. I recommend that archivists standardize on a specific generation of tape for as long as possible and don’t mix generations (don’t mix LTO4 and LTO5, for example, despite the fact that the drives are advertised as working with both).

3. Companies with recordkeeping requirements, like SOX (Sarbanes-Oxley). Tape is just really good for that. It has a way of surviving problems in your IT department.

4. Companies with enough data that they can put a line-item on the budget for backups, and justify the operational cost of tape—support contracts, keeping staff on hand who know how to use tape, that kind of thing.

Of these, the data hoarders are going to use the 150 TB break-even point just because they want as much data as possible. Everybody else is going to make decisions based on other factors, like staffing or compliance. There are a number of gotchas, like problems with mixing tape generations, the prospect of using robotic tape libraries, and support contracts, that make the tradeoff much more nuanced.


Yes, considering the additional cost of a tape drive versus the price difference per stored TB between tapes and HDDs gives an intersection point somewhere between 100 TB and 300 TB.

Taking into account that for long term archiving it is necessary to store 2 or 3 copies of the data reduces proportionally the threshold above which tape is preferable.

The storage cost per TB includes not only the purchase price but also the lifetime of the media, e.g. a HDD model with a warranty of 2 years cannot be trusted to store data for a longer time.

Tape is guaranteed for 30 years, but the storage time is normally reduced to about 10 years by the risk that the corresponding tape drives may become obsolete.


Yeah. It gets even more complicated, because tape drives and tape media has different failure rates.

At the places where I used tape, we used a more efficient encoding for archival tape backups. Rather than storing 2 or 3 copies, we used forward error correction with something like 30% overhead. This, then, gets even more complicated to evaluate, because it multiplies how “hungry” your backup / archive system is for ingesting data in order to remain efficient & still write data out to tape by whatever deadlines you have set. If you store 3 copies of data on LTO-8, then you write data in blocks of 12TB, with one copy on three tapes. If you use forward error correction, you might do something like write out 96TB at once on 11 tapes. You use less than half as many tapes in the long run, but you need to feed the machine faster in order to meet deadlines.


While it is true that more complex error correction methods can reduce a lot the amount of tapes needed for storage, the simple approach of making 2 or 3 copies remains necessary in many cases, because storing the copies in different geographical locations is typically the only method that can protect an archive against a calamity that may destroy an entire storage center.

It is possible to use RAID-5/RAID-6-like encoding schemes that can survive the complete destruction of 1 or even 2 storage centers while using less tapes than with simple copies, but such encoding schemes can be used only by very large organizations, which use more than 3 separate storage centers.


Yes, the needs for archives (sole copy) and backups (additional copies of live data) are different.

> It is possible to use RAID-5/RAID-6-like encoding schemes that can survive the complete destruction of 1 or even 2 storage centers while using less tapes than with simple copies, but such encoding schemes can be used only by very large organizations, which use more than 3 separate storage centers.

Scenarios where two storage centers are destroyed—that’s extreme. The most paranoid scenarios I’m normally willing to entertain are along the lines of one data center burns to the ground in a generator fire, and somebody drives a truck full of backup tapes into a ditch and they’re all covered with mud and sand.

Tapes have a high enough failure rate that you benefit from forward error correction and you benefit from planning to handle individual tape failures. This includes stuff like the tape leader breaking, somebody losing a tape, damage during transport, water damage in storage, etc.

There’s a layered approach here, where you plan for different disasters at different levels of the stack. Each layer exposes some certain failure rate to the layer above it, and deals with some certain failure rate at the layer below it. When I think of backups, I often imagine a top-level data storage system that has a geographically distributed hot backup, and then an offline cold backup. This lets you survive complete destruction of one data center, or lets you survive a catastrophic software bug that destroys data (and a bunch of tapes are damaged on top of that). Pretty good baseline, IMO.


Another big case besides SOX compliance is medical records generally or imagery specifically.

Basically anywhere where you have a lot of data that has to be retained indefinitely for regulatory compliance or practical reasons is a great case for tape. But yeah, the robotic library and the service costs are pretty high until you hit a huge amount of data.

Source: supported a couple of large medical installations for a couple years many years back. Can confirm that you're dealing with a lot of mechanical complexity per GB until you get to an absolutely enormous amount of data. I genuinely can't imagine breaking even with a robotic library you can't climb into.


But presuming this isn't data with low-latency access requirements (since tape is pretty useless for that, so we wouldn't be making the comparison), what's the inflection point where it becomes worth the CapEx to justify even having your own "nearline" + archival storage cluster at all, vs. just using Somebody Else's Computer, i.e. an object-storage or backup service provider?

To me, 1PB is also where I'd draw that line. Which I would interpret as it never really being worth it to go to local drives for these storage modalities: you start on cloud storage, then move to local tapes once you're big enough.

(Heck, AFAIK the origin storage for Netflix is still S3. Possibly not because it's the lowest-OpEx option, though, but rather because their video rendering pipeline is itself on AWS, so that's just where the data naturally ends up at the end of that pipeline — and it'd cost more to ship it all elsewhere than to just serve it from where it is. They do have their self-hosted CDN cache nodes to reduce those serving costs, though.)


On the other hand, with either tape or hard drives, you can leave it on a shelf for 10 years and the data has a decent chance of still being intact. Proper procedure would dictate more frequent maintenance, but if for whatever reason it gets neglected, there's graceful degradation. With AWS, if you don't pay your bills for a few months, your data goes poof. Other companies might have more friendly policies, but they also might go out of business in that span of time.

Particularly relevant for personal data storage.


I think someone else mentioned in this very comments section that hard drives "rot" while spun down — not the platters, but the grease in their spindle motor bearings or something gets un-good, so that when you go to use them again, they die the first time you plug them in. So you don't want to use offlined HDDs for archival storage.

(Offlined SSDs would probably be fine, if those ever became competitively affordable per GB. And https://en.wikipedia.org/wiki/Disk_pack s would work, too, given that they're just the [stable] platters, not the [unstable] mechanism; they would work, if anyone still made these, and if you could still get [or repair] a mechanism to feed them into, come restore time. For archival purposes, these were basically outmoded by LTO tape, as for those, "the mechanism" is at least standardized and you can likely find a working one to read your old tape decades later.)

Even LTO tape is kind of scary to "leave on a shelf" for decades, though, if that shelf isn't itself in some kind of lead-lined bunker, given that stray EM can gradually demagnetize it. If you're keeping your tapes in an attic — or in a basement in an area with radon — then you'd better have encoded the files on there as a parity set!

I think, right now, the long-term archival media of choice is optical, e.g. https://www.verbatim.com/subcat/optical-media/m-disc/. All you need to really guarantee that that'll survive 50 years, is a cool, dry warehouse that won't ever get flooded or burnt down or bombed [or go out of business!] — something like https://www.deepstore.com/.

But if you're dealing with personal data rather than giant gobs of commercial data, and you really want your photo album to survive the next 50 years, then honestly the only cost-efficient archival strategy right now is to keep it onlined, e.g. on a NAS running in RAID5. That way, as disks in the system inevitably begin to die or suffer readback checksum failures, monitoring in the system can alert you of that, and you can reactively replace the "rotting" parts of the physical substrate, while the digital data itself remains intact. (Companies with LTO tape libraries do the same by having a couple redundant copies of each tape; having their system periodically online tapes to checksum them; and if any fail, they erase and overwrite the bad-checksum tape from a good-checksum copy of the same data — as the tape itself hasn't gone bad, just the data on it has.)

Paying an object-storage or backup service provider, is just paying someone else to do that same active bitrot-preventative maintenance for you, that you'd otherwise be doing yourself. (And they have the scale to take advantage of shifting canonical-fallback storage to being optical-disk-in-a-cave-somewhere format — which reduces their long-term "coldline" storage costs.)

Instead, you're just left with the need to do the much rarer "active maintenance" of moving between object-storage providers as they "bit-rot" — i.e. go out of business. As there are programs that auto-sync between cloud storage providers, this is IMHO a lot less work. Especially if you're redundantly archiving to multiple services to begin with; then there's no rush to get things copied over when any one service announces its shutdown.


That’s a really good point. For us (near that 150-300TB inflection point for archival storage), it made more sense to put the data on S3 glacier. First off, the data is originally transferred through S3, but mainly, glacier hits the same archival requirements as tape, at a pretty compelling cost.


Yeah, but that ZFS server is likely in the same rack that's going to get knocked out when the data center gets sucked into a tornado, or whatever. Weather it makes sense to have someone on staff swapping tapes, rotating backups, shipping them offsite, and keeping records of what exactly is where not to mention handling AND TESTING restores really depends on what your data's worth. For most places, it's cheap insurance.


Put an online backup in a another location then.


> Put an online backup in a another location then.

And hope your online backup doesn't get hit by the same group/malware that took out your primary production system.

Offline backups help with certain types of risks.


You can use zfs send which is a append only operation


> You can use zfs send which is a append only operation

It's not about how the data gets there, it's the fact that the system is online and hackable just like any other system.

Good luck wiping offline tapes (Mr. Robot not withstanding).


More realistically you are deploying two of the servers; redundancy is not backup. And now you have two servers to administer.

The whole point of something like tape is having an off-line copy of your data, ideally in a separate physical location. A second server with a bunch of disk in the same location connected to the same network isn't that, and will not save you from ransomware or natural disaster.


> And now you have two servers to administer.

If you use even crude tools such as clusterssh, managing a bunch of machines isn’t linearly harder than managing one.

While tape renders itself easily to be offline storage (it’s offline as soon as you eject it after all), you can have that with remote servers that pull data to back it up instead of receiving pushed data. If no server can push data to any other server, only pull from them, a ransomware attack becomes a lot harder.

Also, while tape in a warehouse (or on the desk) is offline, tapes in the robot are no more difficult to destroy than hard disks. They are just slower.


100%

Its also the software to run the blasted thing as well. As soon as you get into tape shit get's enterprise-y real quick. There are opensource tools to manage tape collections, but its not fun.

LTO tape libraries are fairly cheap to pick up second hand, its the cost of getting the newer drives that hurts.


Frequently there have been problems getting the tapes to go with newer drives because of patent wars too.


There are tape libraries as small as 3U with 25 slots. That’s a capacity of only 300 TB with LTO-8. It’s not hard to justify if you’re working with stuff like video.


> just sits in the rack and does its job quietly

Oh does it? You'll never be forced to update or maintain that configuration due to shifting sands of company policy, infra, or CVEs?


15 years ago I worked for a tape backup company that got bought out by a company making disk backups. Before they bought us out the CEO was constantly saying "tape is dead" in interviews, but once there was money for him to make in tape he stopped saying that. I remember when someone asking him about this and he responded something "amazing how money can make you change your tune"


> It is difficult to get a man to understand something when his salary depends upon his not understanding it.


W transparency


I remember a gmail outage circa 2008 where they said they were restoring from tape, and people found the idea comically retrograde even back then. But I started at a company with large Commvault tape archives in 2011 and they mentioned that it was just more economical at scale for the workload (the backup team had some sweet gear that was apparently purchased with a kickback program on their LTO(4?) tape purchases; perhaps that assisted the engineering decisions) . However, by the end of that decade, they had migrated to mostly using large drive arrays with tape used only for offsite, so my impression was that tape was indeed a shrinking, niche method of data storage.


An interesting Google Tech Talk video from 2013 about Google backing up the internet (and gmail): https://www.youtube.com/watch?v=eNliOm9NtCM


Fantastic talk, and I'm just 12 minutes in on a 1h14m video.

Take the Cartesian product, in fact, of those factors. I want location isolation, isolation from application layer problems, isolation from storage layer problems, isolation from media failure.

Edit:

They do RAID4 on the tape level. 5 tapes, one of them the parity tape.

When that fails: Reconstruction at sub-tape level. "We don't really have data loss."

"We make the backups as complicated and take as long as they need. The restores have to be quick and automatic. Recovery should be stupid, fast and simple."


I'm pretty sure big cloud companies will keep using tape for a long time.

For a home/small business it's still too much effort and expensive.

I tried revisiting this topic so often and HDD was always so much cheaper for just 5-30tb which is quite a lot for quite a lot of companies.


Used older generation LTO drives are cheap.


Counter:

Retrieval (recovery) from disk is easier and faster - ability to randomly read and write. Tapes are sequential - to get to a data which is only at the end, one has to go through all the tapes that is there before that.


The maximum tape seek time is around a couple of minutes. The average seek time is about half of that, i.e. around 1 minute (these example values are valid for LTO-7).

These seek times are seldom a problem. If one would keep on HDDs as much data as it is normally kept on tapes, i.e. at least a few hundred terabytes, one would have the same slow seek times as someone would have to plug and unplug external HDDs, unless the storage would belong to a very large company that could afford the high costs of keeping online many HDDs.

As an individual user, I keep in my computer, on an SSD, an index with the content of all tapes. When I need some file, I get it in at most 5 minutes, which includes starting the tape drive, going to a cabinet and taking the tape from a shelf, inserting it in the drive and waiting for the seek and copy operations.

This is perfectly acceptable, because I do not need files from the tapes every day. Storing the data on external HDDs would not reduce the time wasted with manual operations, it would be much less reliable and it would be much more expensive for great amounts of data.

The sequential transfer speed of tapes is greater than that of HDDs. Therefore, if after a seek you copy large files, e.g. a movie in BluRay format, the whole seek + copy operation takes less than from a HDD.


That detail is abstracted away from the user though, tapes north of LTO 5 (which is pretty old) allow addressing tapes like a normal file system, and if you are doing backup things, the contents of the tape are held separately from the data itself.

As an off-line archival backup medium, access times really are not a concern in the real world.


Tapes are not strictly sequential, they are also a 2D medium, admittedly with one sequential dimension:

https://en.wikipedia.org/wiki/Linear_Tape-Open#Physical_stru...

https://en.wikipedia.org/wiki/Linear_Tape-Open#Positioning_t...


What is the value of that? Is that a common use case for those buying exabytes of tape?


yes, for random IO tapes suck.

but, when you are recovering files, its rarely a random io type deal.

Even then, you generally just dump to the nearline/tape cache and fiddle with the data there.

A decent 25 drive tape library will easily saturate a 100gig network link, and its perfectly possible to add more drives to get more IO.

Tapes as part of tiered storage is something that is really powerful. Yes, flash is cheap, but not cheap enough to be used as long term backup (ie legal hold or general archive.)

Keep your expensive fast storage near your clients, the cheaper less fast, but more voluminous a stage away, and then dump out to tape after n weeks with no access.


If you are using tapes, you aren't going to use it for a use case where you need to recover specific parts of the backup.

Tapes are for archival storage, not for 'accidentally deleted a single file and need it back' type backups.


That’s just not correct. Tapes can be used for either purpose, and with a decent backup software managing them, restoring specific parts of the backup is easy—you just need to wait for it to seek the tape.


Indeed. The only time I have ever seen anyone restore from tape in case of real need was to recover a tiny fraction (< .02%) of files that software had mistakenly deleted. https://gmail.googleblog.com/2011/02/gmail-back-soon-for-eve...


This is incorrect. Last place I worked at with tape, the poor tape guy was constantly going back for some deleted a single file and needed it back situation.

Once, we had to set up a hollow Sharepoint and restore a Sharepoint backup just to get someone's deleted file.

Now, do I think this is a good idea? No. Frankly, people who cause these kinds of things need to see "This cost X hours to recover, stop doing that" as feedback.


The disk also has to move its head to the location of the data. It is faster, but still sequential.


I have wanted to use a tape drive for years, for storing backups. Partly because tapes are cheap and numerous, partly for the aesthetic, honestly. LTO-5 is indeed quite cheap now, you've inspired me to give it a go.


I use LTO-4 and LTO-5 tapes to back up all sorts of things for both myself and for my clients

My archive is getting to the size that optical discs are becoming cumbersome. Do you know of a tape solution that works with macOS?


Every few months I go to check out prices for Ultrium Tape Drives and I've had a super hard time finding affordable ones (everything is almost $1K).

Any tips? Particularly on finding something I can interface with SATA?


> Older, used LTO drives are actually quite affordable

They are, but If I'm not mistaken you need a SAS controller, and those go for ~1k euros in my area, even used.


You can often find compatible SAS controllers used on eBay for quite cheap - often flashing a “raid” controller into a simpler SAS controller.

Fiber channel drives also exist.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: