My wife’s computer is a unibody, late-2012 Mac Mini, model A1347 with Fusion Drive, which we’ve had since then, having ordered it to our specs directly from Apple. We’re happy with it. It’s a lovely little computer with more than enough oomph for my wife’s needs (she is an author).
The HDD on her Mac was silently failing and her computer was getting slower. A quick disk speed test revealed that its write speeds had decreased by about 75%.
Running First Aid on the system volume did not yield any insights into the HDD’s true state. Thankfully, there’s a little app called DriveDx, which I talked about in a previous post. Running that app revealed the HDD’s problems.
The SSD wasn’t doing too well either, but at least its lifespan was at about 50%.
The solution was simple: I needed to replace the HDD. A 1TB SSD would suffice, so I ordered one (an ADATA SU800 1TB SSD). My wife continued to use her computer as usual, since it was still working, although I made doubly sure that it was backing up to Time Machine. I would restore her data from those backups after I replaced the HDD.
Once the SSD arrived, I got to work. I didn’t want Ligia to experience an outage longer than a few hours, so the pressure was on. My plan was to open up her machine, clean the insides thoroughly of dust, replace the thermal paste on the CPU and GPU, then replace the HDD with the SSD. After putting it back together and booting up, I would need to do a data restore.
Here is a gallery of photographs from that process. The insides were indeed full of dust and the thermal paste had dried up. I followed this guide from iFixit, although I have to say it’s not entirely accurate, as detailed below.
I was on my own when it came time to work on the AirPort/Bluetooth board, where the setup differed quite a bit from the guide. There were also a few screws whose location was different in the guide. So I took photos before I disassembled things, just to be safe.
While I love the design of the Mac Mini (inside and out) and I think it’s a fantastic little computer, it’s tricky to work on. Everything has to fit together just right. The things that gave me problems when it came time to re-assemble it were:
the minified SATA cables, which kept popping out of their slots on the motherboard and are really only held in place by the cowling (the little piece of plastic in a semilune shape),
re-seating the top drive, whose side screws have to slide into some holes in the back of the case, but there is little to no tactile feedback when they’re in place, and there’s no way to check things visually; it actually fits asymmetrically over the bottom drive, which is a bit illogical, but that’s how the engineers worked out the hardware design,
and the antenna plate. Oh wow, the antenna plate was a chore to work back in… It has to fit in just right, hugging the inside edge of the case with an indentation made in the wire mesh from which it’s constructed, and for some reason, it just didn’t want to go back in properly. It was off by less than 1 mm, yet it meant that I couldn’t put the screws back on. Be careful with that one!
When it was time to boot it up, the Mac Mini refused to do it. I stared at a black screen for a minute or two, wondering if I’d forgotten to connect some cable inside it, and then it occurred to me to re-seat the AC cable, which is notoriously hard to plug and unplug on this machine, because its slot is too tight. That turned out to be the problem. Whew.
Another wrinkle that I ran into was the Fusion Drive. This machine has an actual SSD inside of it, not a blade SSD, which is what you might find in an iMac or a MacBook. That was a bit of a surprise to me. Anyway, come time to reformat the drives, I figured I could re-enable Fusion Drive and end up with a single volume that used both the Apple SSD and the new ADATA SSD. Nope. While you can run the commands in Terminal to “marry” the two SSDs into a Fusion Drive (see this post for the details), checking the resulting volume with Disk Utility gives an error and Mojave refuses to install on it. So… no Fusion Drive for my wife, I guess. Then I figured I could create a software JBOD in Disk Utility to end up with a single volume once more, and I did that, and it worked, but once again, Mojave refused to install on it. So I had to simply format each SSD as a separate drive and use the 1TB SSD as the system volume, leaving the 128GB Apple SSD as a secondary volume to be used occasionally.
A quick check with DriveDx showed me that the new SSD was doing just fine.
And a disk speed test showed things were humming along nicely.
Here are some Geekbench scores for good measure.
My wife’s pretty happy with it now, she says it is faster than before and it doesn’t crash anymore, which it used to do every now and then. And if my wife’s happy, then I’m happy.
In a recent post, I wrote about upgrading the original (and failing) blade SSD in my iMac to a bigger and faster NVMe module. During that upgrade process, I wondered whether splitting my Mac’s Fusion Drive would result in better performance, but decided against it for simplicity’s sake.
Even though I decided against splitting my Fusion Drive at that time, I read articles that advocated for it and suggested even better performance was to be had by allowing the SSD and HDD to run as separate volumes. The idea is to install the OS and select files and folders on the SSD, with the bulk of the files on the HDD. For the sake of experimentation and learning something new, I decided to tinker with my iMac and see if I could squeeze out some extra speed.
For those who are wondering what I’m talking about, Fusion Drive is an Apple technology built into macOS that creates what is essentially a hybrid drive, by combining an SSD module (NAND flash) with a traditional HDD (platter drive) and presenting the two as a single volume to the user. The protocols that govern the data I/O are called Core Storage. Apple writes: “Presented as a single volume on your Mac, Fusion Drive automatically and dynamically moves frequently used files to flash storage for quicker access, while infrequently used items move to the high-capacity hard disk. As a result, you enjoy shorter startup times and — as the system learns how you work — faster application launches and quicker file access.”
I’ve been using Fusion Drive since it came out, retrofitting my iMac at the time with a new SSD and thus making it run faster than its original specs. I love this technology, because it offers significant performance improvements for a fraction of the cost of buying a large SSD, which used to be be quite expensive a few years ago.
The long and the short of it is that it’s not worth it to split your Mac’s Fusion Drive. If you’re currently running Fusion Drive on your Mac, keep doing that, you won’t see any significant performance improvements if you split it. Actually, some things may run slower than before, and you’ll also have to deal with a few inconveniences, as detailed below.
I’ll present both scenarios here and you can decide what to do for yourself. There are multiple methods to it. These are the methods I’ve chosen. The number of Terminal commands that you have to run for either scenario is minimal, and the time involved has to do mostly with backing up your computer, waiting for the OS to reinstall and for your data to be restored from backup. For example, if you’ve got a 3TB drive and you’re at about 50-60% usage (and you should be at that threshold or lower on any hard drive), then you should figure on 4-5 hours for either of the two scenarios.
How to split your Fusion Drive
First and foremost, did you backup your computer? If you did, go ahead and create a bootable drive using Apple’s instructions, then boot into it by pressing the Option key as soon as your Mac restarts and holding it down until you see the Apple logo. You need to boot into a separate drive because you’ll be deleting your internal drives entirely, including the boot and recovery partitions.
Once you’re in, open Terminal and get a listing of your disks and volumes.
Your Fusion Drive presents itself as a logical volume group that appears as a separate disk with an HFS+ or APFS partition. Say your SSD is disk0 and your HDD is disk1, your Fusion Drive would be disk2 or disk3. In my case, it was disk3 (disk2 being the bootable recovery drive). Now unmount your internal disks.
diskutil unmountDisk disk0
diskutil unmountDisk disk1
You’ll want to delete that entire disk containing Fusion Drive. Be forewarned, this deletes all you data. Did you backup your computer?
diskutil apfs deleteContainer disk3
Now that Fusion Drive has been nuked, you’ll still have your separate drives that you’ll want to make sure are erased. The eraseDisk command requires that you offer a new name for each disk, so I chose to name them SSD and HDD, to keep things simple.
diskutil eraseDisk JHFS+ SSD disk0
diskutil eraseDisk JHFS+ HDD disk1
Now you’ll want to do a fresh install of macOS onto the SSD, and after that’s complete, you’ll boot up into your fresh install and go to Utilities/Migration Assistant, in order to do a selective data restore. Here you’ll have to decide for yourself, based on the total size of your SSD and your data set, how much of it you’ll want to restore onto the SSD. The rest you’ll need to copy manually from the backup drive onto the HDD. In my case, I restored my user settings and the system and libraries folders onto the SSD, and I copied the following folders onto the HDD: Documents, Downloads, Movies, Music, Parallels (in case you’re running some kind of VM software) and Pictures. Each of those folders was too big to keep on the SSD, even though I have a 512GB module (remember the rule about keeping your drive at or below 50-60% usage).
Once you complete all that work, you’ll need to create links to these folders on the HDD in place of your folders on the SSD. Mojave won’t let you do this when you’re logged into your account, so you’ll need to boot up into recovery mode and open Terminal once more.
Go to your homefolder on the SSD.
Delete the folders that are now present on the HDD. You’ll need to do this for each folder that you’ve moved there. Hopefully you’ve written down their names ahead of time.
sudo rm -rf foldername
In your homefolder on the SSD (same location as above), make links to the folders on the HDD. I chose to put mine at the drive’s root level. You may choose to put them in a folder. Just don’t give it the same name as your username, I hear that may cause problems. You’ll need to do this for each folder.
ln -s /Volumes/HDD/foldername
That’s it, restart and use your computer. However, you may find a few inconveniences — these are the ones I experienced:
I noticed no performance improvements. There wasn’t even an improvement in the bootup time. Nothing, nada, zilch.
While Apps may open up faster, if they’re still accessing files on the HDD, editing will still be sluggish. In order for you to see that performance boost talked about with SSDs, both apps and their files need to be on the SSD.
In my case, I had to keep the Photos library on the HDD, because it was too big to keep on the SSD, and while Photos may have opened up fast, loading up the library took forever, until enough of the recent photos were cached on the SSD to allow me to work with my library. So things were a LOT slower with this app.
I kept my mailboxes on the SSD so I was hoping for better performance from Mail, but I didn’t get it. I have a lot of mail stored locally, so in theory, things should have worked faster because everything was on the SSD, but they didn’t. I also experienced odd issues, like when moving messages between mailboxes, it took a lot longer and sometimes didn’t register. I’d drag and drop them, then come back to the app a little while later and find them in the same place, just as if I hadn’t moved them.
iCloud would display an odd notification icon, but when I’d go into it, there was no message. This icon was displayed continually for as long as my Fusion Drive was split. See the screenshot below.
While Time Machine will backup both internal drives, data restores will only restore the files from the SSD. I don’t know why and I don’t know how to fix that, so keep this limitation in mind. You can go into the Time Machine drive manually and copy the files over afterward, but if you run a restore operation on your computer and you wonder where most of your stuff is after it’s completed, don’t freak out, just know you’ll need to get it manually from the drive.
How to enable your Fusion Drive
After about a week of running my Mac with a split Fusion Drive, I’d had enough and decided to re-enable it. Here’s how I did it. Before you proceed with this, I’ll ask you again, did you do a full backup of your computer? This will wipe all your data.
Using the same bootable drive, I booted into it and opened up Terminal. Since you’ll be wiping all your internal drives again, you need to be booted from an external drive.
Apple recommends this single Terminal command that is supposed to do everything in one fell swoop. It didn’t work for me, perhaps because my SSD module was a newer NVMe running off an adapter card, not the Apple-approved blade SSD manufactured specifically for this kind of thing.
I had to do it with a few more commands. First, find out your disk IDs.
Now unmount your internal disks.
diskutil unmountDisk disk0
diskutil unmountDisk disk1
Then create a merged virtual hard drive with Core Storage.
diskutil coreStorage create Macintosh\ HD disk0 disk1
Now get its logical volume group name (the very long alphanumerical name that appears in Terminal after you type this command).
diskutil coreStorage list
Now format and create the JHFS+ volume that will run Fusion Drive.
diskutil coreStorage createVolume yourlogicalvolumegroupname jhfs+ Macintosh\ HD 100%
Don’t worry about formatting the drive to APFS. That’ll happen automatically when you install Mojave. Besides, APFS is not an entirely separate file system, it’s a container running inside HFS+, so like I said, don’t worry about it. That’s it. Now quit Terminal and do a full restore from Time Machine, but prepare yourself for an incomplete data restore (see the reasons given in the previous section). Once the data restore is complete, you’ll need to manually copy the folders that are missing from the Time Machine drive. Or, as I did, you can do a full restore to a backup set that existed before I split my Fusion Drive, which means you’ll get all your old data back in all the right places, but you’ll still need to get your newer files manually from the Time Machine drive.
In my case, I needed to copy the mailboxes, which are located in ~/Library/Mail/V6 from the newest backup set (the one with the split drive) to my computer, and that gave me all my mail, including the interim stuff. I also copied the latest Photos library, and that gave me all my photos, including the interim stuff. Then I went through the Documents and Downloads folders on the Time Machine drive, sorted by date modified and copied the interim files onto my computer. I didn’t need to go through the other folders because I knew I hadn’t worked on other stuff. And once I did this, my data restore was complete. Mail and Photos still needed to rebuild their libraries though, and that took a while.
And because I use Backblaze to backup my computer offsite, I also needed to uninstall and reinstall that, then inherit a previous backup state (don’t worry about this if you’re not using Backblaze).
When that was done, Backblaze told me it had “made” my computer inherit my backup state, as if it had forced it to do this, in a non-consensual way. Kind of a funny way to word things, but their service works well.
Here’s the kicker. I ran another drive performance test after all this, and these were the results.
Everything runs fast now, and it runs as expected, without hiccups.
As I said at the start of this article, if you’re already running Fusion Drive, do yourself a favor and leave it running. You’ll avoid headaches you don’t need, unless you like complications.
The only way I can see to speed up my iMac even more, is to purchase a large 3-4TB SSD and run it as my only internal drive. That might be a little faster. But as you can see from the test screenshot shown above, my iMac is no slouch right now. And 4TB SSDs are still fairly expensive. It might actually be cheaper (and possibly faster) to get a 2TB SSD and a 512GB NVMe module, and run them together with Fusion Drive, although the overall capacity wouldn’t be the same. Food for thought.
I’ve worked on a few upgrades to our family computers in recent months: my 2008 MacBook Pro, my mom’s 2007 iMac and my own late-2013 iMac (A1419, EMC 2639). This is the tale I’d like to recount for you now, because it’s something that I’ve had to deal with since last autumn (october of 2018), and I’ve just recently (I hope) finalized the upgrade/repair. There’s a valuable lesson in here for the people at Apple Support, if they’re interested.
It started with my iMac slowing down to a crawl over the course of a few days, back in late 2018. At first I thought it was spyware or a virus. I checked everything: every app, every file, every process. I removed apps, one by one, to see if it would fix the problem. It didn’t. I ended up removing all of the apps but those that came with macOS, and the problem still persisted. I wiped the drive clean and reinstalled the OS, then reinstalled the apps anew, one by one. It was just as bad. I ran hardware tests using Apple’s software and everything checked out. I scoured the web for solutions. There were some suggestions online that iCloud could cause slowdowns when the syncs weren’t going properly. I checked the Apple Server Status page and indeed they’d been having some problems with iCloud, but they were marked as resolved. I checked Photos and there were major issues: my photos weren’t syncing properly across my devices, and there were image compression/corruption (?) issues going on, with diagonal blue lines appearing all over my recent photos, lines that persisted even when opening the photo at full resolution.
I called Apple Support and began a series of interactions that did not end with any sort of solution. From the start, they agreed that iCloud was causing the slowdowns and had me go through a series of steps such as logging in and out of iCloud, disabling and enabling the various iCloud syncing options, etc. to no avail. Mail started acting up as well, so they suggested I disable Mail syncing, because I had “too many messages in my mailbox”. Documents started acting up, with iCloud Drive showing up empty on my computer, so they suggested disabling that. In spite of the fact that I’d already done it, they tried to convince me to reformat my computer and start fresh. I kept getting a hunch that something else was amiss and asked them if they were sure this wasn’t a hardware issue. They said no. They told me to wait for the photos to finish syncing, then enable the other iCloud features one by one, and things would get back to normal. They didn’t. We kept going back and forth, with Apple Support posting one update per day (or less) to the case, asking me to do this and that, and even though I’d complete their requests and post more updates during the day, they wouldn’t respond till the next. I offered to pay a case fee to expedite issues. I explained to them that this was my main computer and I couldn’t do my work. This went on for weeks, with me getting more desperate and the techs telling me they’d stop helping me because I couldn’t stay calm.
With things going nowhere and Apple Support techs who couldn’t care less, I decided to exercise the “nuclear” option. I found Tim Cook’s email address and wrote him an email. I didn’t expect a response, but I wanted to vent. To my surprise, a short while later I got a response, not from him but from his office, promising me my issue would be forwarded onto someone who would get back to me. I waited a couple of days and… nothing. No one contacted me. I figured I’d try my luck again. I got the same response, but someone finally contacted me and connected me directly with an advanced support technician. There’s apparently a “third tier” of tech support that is only available on a case by case basis; I guess after almost a month of my computer being down for the count, I qualified.
It took a while longer to get to the bottom of the problem: daily communications, screen sharing sessions, uploading log files to Apple servers, trying various steps, etc. It took over a week. iCloud was at first to blame, then Adobe software, then finally, after my case was put in front of a senior technician directly responsible for iCloud connectivity, my issue turned out not be software related at all, but caused by hardware. Hold on to your hat, because as it turns out my HDD was going bad. That was it. It was as prosaic as that! In total, I’d lost over one month of my time and I had to appeal to Tim Cook’s office, all because Apple Technicians couldn’t pinpoint a failing hard drive from the get-go.
Once the problem was found out, it was an easy enough fix. I opened up my iMac and replaced the HDD with a fresh one. I also found and installed a piece of software called DriveDx, which gives detailed stats about hard drives and can let you know of a bad drive before it actually fails. Most drive diagnostic apps rely on the S.M.A.R.T. status flags, but that’s not enough. DriveDx does a whole lot more. The app quickly let me know that the blade SSD (the second half of the Fusion Drive on my iMac) was also close to its lifespan. Since it was still working okay, I decided to hold off on replacing it last year, choosing to monitor it with the app and only replace it when it was close to failing. Here is a gallery of photographs from the time I replaced the HDD. I also chose to take apart the chips and heat sinks and to replace the thermal paste, which had become dry and cracked. Before I put my computer back together, I cleaned the case and the parts thoroughly with a brush and soft cloth, because a lot of dust had accumulated inside and on them. After I put my iMac back together, it worked beautifully once again. I know I could have replaced the HDD with an SSD, and I plan to do it in the future. It’s just that I want to get a 3-4 TB SSD and their prices are still a bit high.
A few weeks ago (about seven months after replacing the HDD), DriveDx told me the blade SSD was fairly close to failing, so I chose to replace it with a Samsung 960 Pro M.2 512 GB NVMe SSD module that I already had, so this upgrade only cost me about $16 for a Sintech NGFF M.2 NVMe SSD Adapter Card, which seems to be the card everyone recommends for MacBooks and iMacs.
I’ve enclosed a separate gallery of photographs of this upgrade below. You’ll laugh when you see one of the photos, so I’ll explain: in what seems to be an iMac design flaw, the cylindrical bracket that secures the screw for the blade SSD is only glued to the motherboard. Unlike every other screw bracket which is secured directly to the motherboard with metal, this one is not designed into the motherboard. Underneath it, on the other side of the motherboard, there are circuits running right across that spot. It seems to me like at the time (2013), the blade SSD and its screw bracket were afterthoughts of sorts for the hardware design team. My bracket came right off the motherboard. I had no glue in the house, only some silicone-based adhesive that takes up to 24 hours to harden, so I put a dab of that under it, tightened the screw and stuck a wood shim in-between the case and the top of the screw, to hold it in place while the adhesive hardened. I know it looks terribly unrefined, but it’s been working fine.
There are two things I should tell you about this upgrade: (1) this particular SSD tends to run hot, so DriveDx will warn you about its temperature, and (2) after booting up my iMac for the first time, it didn’t see the new NVMe SSD, so I powered it down, opened up it up again (thank goodness I hadn’t yet closed it completely) and re-seated the SSD and its adapter in the blade SSD slot. After I did that, it saw it, recognized it, and I was able to boot into recovery, go into Terminal and recreate the Fusion Drive, then reinstall the OS and restore my data.
In spite of the temperature warnings, my iMac has been working great so far. I noticed a bit of a speed boost, but since I’m still using Fusion Drive and I’m tied to a spinning hard drive, a lot of the oomph of the SSD can’t be seen. I suppose I could have chosen to install the OS on the SSD and keep my files on the HDD, but I prefer to work without complications. A single 3.5TB volume works for me. If I could have a single 24TB drive that holds my OS and all my files (that are currently sitting on three external hard drives), I’d be happy with that.
I would like to thank Tim Cook’s office for responding to my messages and getting me out of a real bind. I was at the end of my wits at the time, so the tone of my emails to them was gruff and biting; they could have chosen to ignore me. I don’t know what I’d have done if they hadn’t stepped in. But I do wish I didn’t have to exercise that option. Apple Support should have found out the issue from the start. It wasn’t something arcane, it was a simple drive failure.
I bought a Drobo 5D on the 29th of December, 2012, after experiencing catastrophic data loss with the 1st Gen and 2nd Gen Drobo. During multiple phone conversations with Data Robotics’ CEO at the time, Tom Buiocchi, he convinced me that they were much better-engineered than previous-generation Drobos and they had built-in batteries and circuits that would automatically shut them down safely in case of power loss. I was also told the new firmware running inside them would be checking the data constantly to guard against file corruption or data loss. All of these were problems that I’d experienced with my existing Drobos, so even though I was exhausted after my ordeal and so weary of storage technology, I went ahead and purchased the new model and also agreed not to publish an account of what the Drobos had done to my data at the time. I want to make it clear that I paid for my Drobos, so I didn’t feel that I owed him anything, but I did want their company to do well, because back then they were new and deserved a second chance. Now though, there is no excuse for the multiple times their Drobos have lost my precious data. They’ve been around for 13 years and they’ve had plenty of time to make their technology stable.
What’s probably kept them on the market is the willingness of paying customers like me to take a chance on the uniqueness of their proprietary RAID: as far as I know, they are (unfortunately) the only RAID array that lets you store a large amount of data on a single volume that grows automatically as you add drives and also protects (except when it doesn’t) against hard drive failures.
However, after five years of using my Drobo 5D on a daily basis, I can tell you without any doubt that the Drobo 5D does not keep your data safe. Look elsewhere for safe data storage devices. I certainly cannot trust it with my data anymore, so I’ve elected to publish my account of data loss from 2012, as well as an account of my present data loss. Caveat emptor, lest you also lose your data. I’ve also had multiple problems with my two Drobo 5Ns. Because of these problems, some of which have led to significant data loss and to significant time and effort expended in order to restore my data from on-site and off-site backups, I cannot place my trust the new Drobo models that are available, either: I’m talking about the 5D3, 5N2 and the 8D. I see no reason at all to spend more money on more empty promises from Data Robotics.
I can’t say exactly what happened with my Drobo 5D. Drobo Support could not or did not choose to tell me, even though I sent them multiple diagnostic logs from the Drobo and I asked them to tell me what happened. My best guess is that the 5D kept “healing” a 6TB WD drive with bad sectors instead of asking me to replace it. Then a different drive from the array failed. Once the Drobo told me to replace that drive, I did. But during the process of rebuilding the data set, the Drobo 5D decided it didn’t like the 6TB WD drive it had been healing, and chose to tell me that I needed to replace it now. When I did, it would tell me I shouldn’t have taken it out and I should put it back in. I’d put it back in, and then it’d go through its internal processes, only to find out that it wouldn’t like that drive again, but it didn’t stop there. It’d reboot. At first this reboot cycle would take 10-15 minutes, allowing me to copy some data off it, but then it began doing it every 5 minutes. Since it takes a good 3-4 minutes to boot up, this meant I’d have only 1-2 minutes to copy data off before it’d reboot again. This was not workable. After opening a case with Drobo Support, they told me to put it in Read Only mode. You press Ctrl+Opt+Shift+R while you’re in Drobo Dashboard and this reboots it in that mode, which means it’s not going to try and rebuild any data internally, it’s simply going to present the volume to you as it is. This also turned sour quickly, because after allowing me to copy a small amount of data for a few hours, it began that same 5-minute reboot cycle. So I had no way of getting the data off the damned thing unless I mounted it through Disk Warrior and used the Preview app built into that software, which meant having to put up with USB 1.1 transfer speeds. More on the reduced transfer speeds below.
My take on the situation is that it’s a failure in Drobo’s firmware design. It should have asked me to replace the 6TB WD drive instead of working around its bad sectors. Because it didn’t ask me to replace it in time, it then failed when rebuilding its data set after the second drive went bad. That’s not two drives going bad at the same time, that’s a drive going bad and a few weeks later another drive going bad. The Drobo had plenty of time to fix the ongoing situation if its internals had been programmed correctly, but it didn’t, because of inadequate firmware running the device. That’s bad technology at work, causing me repeated data losses.
Here’s another example of Drobo’s crappy firmware: for the past three years, I have had to force my iMac not to go to sleep, because every single time I’d wake it up, the Drobo 5D would refuse to mount, forcing me to reboot the iMac and/or the Drobo and also disconnect/connect the Thunderbolt cable from my iMac in order to get the computer to see it. Data Robotics tried to fix this horribly annoying problem (which can also cause data loss) through multiple firmware fixes, but I can safely tell you that they still haven’t fixed it. Before I stopped using my Drobo 5D, I was on Drobo Dashboard 3.3.0 and Drobo 5D firmware 4.1.2, while my iMac is on MacOS Mojave 10.14.2, and this problem still very much occurred. Well, it definitely occurred before my 5D crapped itself. Oh, how I’d like to be back to those simpler times when all I had to deal with was keeping my iMac from going to sleep! But no, now I have to deal with massive data loss, once again. For the goddamned umpteenth time, Drobo!
What do you think would happen after enough improper disconnects? That’s right, volume corruption!
Data Robotics marketing speak tells you of super-fast transfer speeds and protection against data loss when you buy their devices. The things you most need to remember are that you will experience data loss (that’s a given) and you won’t be recovering your data at super-fast transfer speeds such as USB 3.0/USB-C or Thunderbolt 1/2/3. You will instead be forced to use Disk Warrior’s Preview application (if you’re on a Mac) and you will be recovering your data at USB 1.1 speeds. That’s right, take a moment to think about that! In order to recover data from my Drobo 5D, I have to use Disk Warrior (a Mac app known for its ability to recover failed volumes), because it won’t mount any other way. It certainly doesn’t mount through the Finder or through Disk Utility.
The size of my Drobo volume is about 12 TB. Thank God some of it was backed up locally with Resilio Sync and some of it I recovered from online backups with Backblaze, so I only needed to recover part of my data from the Drobo 5D itself, but for a single 3 TB Final Cut Pro library, it took roughly 300 hours to transfer it to another drive! 300 HOURS! It might be somewhat tolerable (in some masochistic sort of way) if I knew the damned thing would stay on for all that time, allowing me to copy the data in one go, but you never know when it’s going to restart. It can go at any time.
I’ve posted illustrative screenshots below. Feel free to do the math yourselves as well. Also think about how much worse this problem gets when your Drobo volume is much, much bigger. The Drobo 5D and 5D3 can go to 64 TB, while the Drobo 8D can go to 128 TB. Do you really want to be stuck copying 50-60 TB of data at USB 1.1 speeds? How about 100 TB of data? Think about that before you click on the Buy button and get one of those shiny black boxes.
The hard truth is that you can’t put all your eggs in one basket. Having all your important data on a single volume, which is what the Drobo lets you do, is a dumb idea. If that volume goes, all your data goes along with it, and it can take MONTHS to recover it from backups.
I don’t know about you, but I don’t want to spend anywhere between $5,000-$10,000 with a data recovery firm every few years, when a Drobo unit decides to fail and lose my data. But that’s what Drobo Support advises you to do from the very beginning. They advise against trying to recover the data yourselves, even while working with them, and instead try to convince you to send your Drobo into a drive recovery firm. For a device that’s supposed to protect your data and a company that brags about “protecting what matters”, that’s disgusting. How exactly are they protecting your data when their devices fail miserably every few years? I guess “protecting what matters” really means “protecting their bottom line” by ensuring suckers keep buying their product.
Do you want to know what happens when I try to open a ticket on the Drobo Support website? This:
If a company can’t even fix its support page so that it loads up for its customers, that’s cause for worry. I tried explaining the issue to them, I even sent them screenshots, but the techs I spoke with seemed unable to comprehend why the website wouldn’t load. My guess is they’ve got a geofence that stops visits from Romanian IP addresses.
When it comes to their marketing people, I cannot describe them as anything but a bunch of slime buckets. Twice now they’ve ignored me in dire situations when I reached out to them for help, hoping they would redirect my messages internally and get someone to pay proper attention to my case. Back in 2012, back when my two DAS Drobo units lost over 30,000 files, they began ignoring me after I told them what had happened. Now, in 2018, they pretended to help, just so it looked good at first glance on social media, but they didn’t follow through on their promise and ignored me afterward. See my comments on their tweet here. I also wrote a message to some guy who calls himself the Drobo CTO but gives no real name, asking him to have a look at my case, but he ignored me. If it’s a fictitious account then it’s understandable, but if it isn’t, then he’s a turd as well for ignoring a legitimate and polite request for help from a customer.
I also reached out to Data Robotics’ current CEO, Mihir Shah, to ask for his help, and after an initial reply that said, “I am looking into this case for you. We will get back to you. Thanks.” — his line went dead. I am left to conclude that he’s of the same breed as his marketing people, who did the same thing when I reached out to them.
You might be asking yourselves why I chose to reach out to these people when I had already opened a case with Drobo Support? Because I felt the technician handling my case wasn’t doing a good job. This case wasn’t a freebie, I paid for it, and I wanted to get real help, not bullshit. Here are a few examples of the kind of “support” I received:
He told me to clone a drive using Data Rescue (a Mac app). This was supposed to fix the situation, but it didn’t, because he told me to clone the wrong drive. That’s right, he had me waste an entire day cloning the wrong drive, before I pointed it out and he came back with sheepish excuses, asking me to clone another.
He let me buy two new 8 TB hard drives in order to go through the data recovery process, without saying a damned thing, before another tech stepped in to tell me I should have bought a 6 TB hard drive instead, because the cloning process needed for my scenario requires a drive of the exact same size as the 6 TB WD drive that the Drobo 5D refused to use anymore. Why couldn’t he tell me that from the start? Besides, isn’t the whole point of BeyondRAID, which is Drobo’s proprietary technology, to let people use drives of varying sizes? There goes that advantage when you need it…
Having worked in IT for a long time, and having worked my way up from the help desk, I could tell the technician didn’t give a crap about the case. It was clear to me that he wasn’t interested in helping me or solving the case, he was simply posting daily case updates of 1-2 sentences with incomplete and unclear replies to my questions on how to proceed and dragging the case on, probably hoping I’d give up.
This time around, I’m faring better when it comes to data recovery than I fared in 2012. Unfortunately, I don’t have a full local backup of my data, even though the Resilio Sync software was supposed to mirror it from my Drobo 5D to my Drobo 5N. I do have a full online backup with Backblaze, but getting back about 12 TB of data through online downloads will prove to be a cumbersome and slow affair. They only let you download up to 500 GB at a time, which means that if I want to download all of it, I’d have to create at least 24 restore jobs on their servers, each of which would generate a ZIP file that would need to be downloaded and then unzipped locally. I could also choose their HDD recovery option which ships your data to you on 3.5TB drives, but I’d need 4 of them, which means it would cost $189*4 or $756. It would probably work out to something like $850-1,000 for me in the end, because I would incur a high shipping cost from the US to Romania and I would also be responsible for customs fees. Theoretically, Backblaze offers a money back guarantee if the drives are returned within 30 days, but I doubt I could return them in that time span, given they’d have to make it to Romania and back. They’re supposed to be working on a European data center and it might open this year, and while that’s going to be nice in the future, it’s not going to work for my situation at this time. No, the much more workable solution would have been for me to have a full local backup of the files on the Drobo 5D. But thanks to Data Robotics, that option got shot down and I didn’t even find out until it was too late…
You see, the Resilio Sync software was set to mirror my data from the Drobo 5D to a Drobo 5N. On my gigabit network, that would have worked just fine. The app was running on my iMac and it was also running on the 5N through the DroboApps platform. The software had a few months to do a proper job syncing my data between the two devices and it indicated to me that it had done it. Unfortunately, after my Drobo crashed, I found out that it hadn’t. The reasons offered by Resilio Support were the following:
The Drobo 5N’s processor wasn’t powerful enough
The Drobo 5N’s RAM wasn’t enough
The Drobo 5N ended up using swap memory and somehow it messed up the sync logs, which caused it to think it had finished the sync when it hadn’t
If I’m to take them at their word, the Drobo 5N is too underpowered to handle Resilio Sync. That’s not to say there isn’t plenty of blame to be assigned to the Resilio Sync software as well. After all, it should have kept track of the data sync accurately. What good are they if they can’t perform at their main advertised task?
If I am to look at what Drobo is saying about the successor to the 5N, the 5N2, namely that it “provides up to 2x performance boost with an upgraded processor and port bonding option”, that would certainly mean the 1st gen 5N is underpowered. Finally, who do you think recommended Resilio Sync to me? It was none other than Data Robotics, whom I called to ask for details on DroboDR, their own data sync application. They said it was too barebones of an app for my needs and suggested I look into Resilio Sync, which was much more robust and full-featured. Thanks once again to my trust in the Drobo, my backup plans got sabotaged, forcing me to waste weeks of my time recovering data from a failed Drobo volume.
But enough about the Drobo 5N, I’ll have a separate post where I’ll talk about how that device has also lost all of my data recently… Back to the Drobo 5D.
To recap, the Drobo 5D lost my data in 2015, 2017 and 2018. That’s three separate, significant, data loss events, the last of which wasted more than a month of time till I was back on my feet.
I am back on my feet now. It’s the 26th of January 2019 and I finally got the last of my data recovered. I’ve been down, unable to do my work, since the beginning of December 2018. I can’t even remember what day it was the Drobo 5D failed. It’s a blur now.
I ended up buying a third new hard drive, specifically a 6TB hard drive, to match the size of the drive that the Drobo didn’t want to use any more, and I cloned that drive, which turned out to have 5 clusters of bad sectors, onto the new drive, using Data Rescue software for the Mac. I had to use its “segment” cloning mode, which attempts to get around the bad sectors, and this meant the cloning operation took more than 120 hours!
Around hour 122, my iMac crashed and I had to reboot it, so I don’t know if the cloning operation was completed or not. My screen went black and stayed that way. It could be that the cloning ended and then my iMac crashed, or simply that my iMac had had enough of staying up to work through a lengthy cloning process and went poof. I don’t know, and I wasn’t about to start cloning the damn drive again. I took my Drobo 5D out of Read-Only Mode, stuck the cloned drive into the Drobo 5D and booted it up, expecting some data loss. Besides, even if the cloning operation had completed successfully, I’d have still lost some data. If you look above, you’ll see that there were 298,2 MB of bad, unrecovered data. That wasn’t from one portion of the drive, that was from five different portions, because there were five significant slow-downs in the cloning process.
Once it booted up, I started copying my data off it, not knowing how long it might stay on. For all I knew, it could go into a vicious reboot cycle at any moment. I’d already recovered all I could from Backblaze and from my local backups, and I needed about 6-7 TB of data from the 5D before I was back up to normal… “normal” being a loose term given how unreliable the Drobo is and how much data I’d already lost.
Sure enough, data loss soon reared its ugly head. Thanks, Drobo, you bunch of slime buckets!
So far, two FCPX libraries containing the videos I’d made in 2017 and 2018, are damaged. I don’t know if I’ll be able to get those videos back.
I consoled myself with the idea that at least I was able to recover most of my data at Thunderbolt speeds, not at USB 1.1 speeds, so that saved me about 2-3 more weeks of painful waiting.
All of my data is now off the Drobo 5D, and I don’t plan on using it again. I’ve put it in a storage closet. I’m done with it, and I recommend you also stop using your Drobos, if you are using one, because it’s only a matter of time before they will lose your data. In my opinion, they offer no more protection than using individual drives. They just cost more, force you to buy more hard drives and they make more noise. My office is so much quieter now that the Drobo 5D is gone. I’m using individual drives which I plan on cloning locally. I will continue to use Backblaze for cloud backups, and I may also do some periodic local backups to a NAS of some sort. I have found out during these past few months that I also cannot rely on the Drobo 5N NAS to store my data, because one of my Drobo 5N units has just lost all of the data I’d entrusted to it. I am now in the process of recovering that… It’s non-stop torture, time and data loss with a Drobo!
I have been a Drobo customer since December 2007, when I bought my first Drobo. I was also among the first Drobo “Evangelists”, as they called their enthusiastic customers back then. It’s now January 2019, and I am done with Drobo. I was a loyal customer for 12 years and I stuck with them through an incredibly disheartening amount of data loss, problematic units and buggy firmware. That’s enough of that. Caveat emptor. Drobo no more.
I’ve had multiple Drobo units since 2007. To this day, I still enjoy adding a hard drive to a Drobo. It’s one of those things that can be an ordeal on other tech, but on a Drobo, it’s been made fun through proper planning and design.
It lets you that it’s low on space, you order a drive, and when it comes, you look at the app, which tells you exactly what size-drive is in each bay. Pressing a small lever on the side of the bay releases the drive, which slides out. You put the new one in, the Drobo immediately checks it and formats it, then begins striping the data set across it. By the way, that’s a screen shot showing my Drobo 5D.
I love this process. It’s so simple and so fun! The Drobo doesn’t care what hard drive you buy, as long as it’s larger than what you already had. It allows you to grow the capacity of your Drobo in time, as the prices for newer, bigger hard drives decrease, without any sort of headaches. This is technology done right.
We need to focus our efforts on finding more permanent ways to store data. What we have now is inadequate. Hard drives are susceptible to failure, data corruption and data erasure (see effects of EM pulses for example). CDs and DVDs become unreadable after several years and archival-quality optical media also stops working after 10-15 years, not to mention that the hardware itself that reads and writes to media changes so fast that media written in the past may become unreadable in the future simply because there’s nothing to read it anymore. I don’t think digital bits and codecs are a future-proof solution, but I do think imagery (stills or sequences of stills) and text are the way to go. It’s the way past cultures and civilizations have passed on their knowledge. However, we need to move past pictographs on cave walls and cuneiform writing on stone tablets. Our data storage needs are quite large and we need systems that can accommodate these requirements.
We need to be able to read/write data to permanent media that stores it for hundreds, thousands and even tens of thousands of years, so that we don’t lose our collective knowledge, so that future generations can benefit from all our discoveries, study us, find out what worked and what didn’t.
We need to find ways to store our knowledge permanently in ways that can be easily accessed and read in the future. We need to start thinking long-term when it comes to inventing and marketing data storage devices. I hope this post spurs you on to do some thinking of your own about this topic. Who knows what you might invent?
Yes, you can enable Fusion Drive on older Macs. I’m not sure how this method will work with Macs older than 2011, but I know for sure that it works on mid-2011 iMacs, and quite possibly on other Macs made since then. I have just completed this process for my iMac and I thought it would help you if I detailed it here.
I like Fusion Drive because it’s simple and automated, like Time Machine. Some geekier Mac users will likely prefer to install an SSD and manually separate the system and app files from the user files which take up the most space, which is something that gives them more control over what works faster and what doesn’t, but that’s a more involved process. Fusion Drive works automatically once you set it up, moving the files that are used more often onto the SSD and keeping the ones that are accessed less often on the hard drive. This results in a big performance increase without having to fiddle with bash commands too much.
My machine is a 27″ mid-2011 iMac with a 3.4 GHz processor and 16GB of RAM. I bought it with a 1TB hard drive, which I recently considered upgrading to a 3TB hard drive but decided against, given the fan control issues with the temperature sensor and the special connector used on the factory drive.
I purchased a 128GB Vertex4 SSD from OCZ. It’s a SATA III (6 Gbps) drive and when I look in System Info, my iMac sees it as such and is able to communicate with it at 6 Gbps, which is really nice.
The hardware installation is somewhat involved, as you will need to not only open the iMac but also remove most of the connections and also unseat the motherboard so you can get at the SATA III connector on its back. You will also need a special SATA wire, which is sold as a kit from both OWC and iFixit. The kit includes the suction cups used to remove the screen (held into place with magnets) and a screwdriver set.
You can choose to do the installation yourself if you are so inclined, but realize that you may void the warranty on the original hard drive if something goes wrong, and this is according to Apple Tech Support, with whom I checked prior to ordering the kit. Here are a couple of videos that show you how to do this:
In my case, it just so happened that my iMac needed to go in for service (the video card, SuperDrive and display went bad) and while I had it in there, I asked the technicians to install the SSD behind the optical drive for me. This way, my warranty stayed intact. When I got my iMac back home, all I had to do was to format both the original hard drive and the SSD and proceed with enabling the Fusion Drive (make sure to back up thoroughly first). You can opt to do the same, or you can send your computer into OWC for their Turnkey Program, where you can elect to soup it up even more.
Once I had backed up everything thoroughly through Time Machine, I used the instructions in this Macworld article to proceed. There are other articles that describe the same method, and the first man to realize this was doable and blog about it was Patrick Stein, so he definitely deserves a hat tip. I’ll reproduce the steps I used here; feel free to also consult the original articles.
1. Create a Mountain Lion (10.8.2) bootup disk. Use an 8GB or 16GB stick for this, it will allow you to reformat everything on the computer, just to clean things up. Otherwise you may end up with two recovery partitions when you’re done. I used the instructions in this Cult of Mac post to do so. The process involves re-downloading 10.8.2 from the Apple Store (if you haven’t bought it yet, now is the time to do so) and an app called Lion Diskmaker.
2. Format both the original HD and the SSD, just to make sure they’re clean and ready to go. Use Disk Utility to do this, or if you’re more comfortable with the command line, you can also do that (just be aware you can blow away active partitions with it if you’re not careful).
2. List the drives so you can get their correct names. In my case, they were /dev/disk1 and /dev/disk2.
3. Create the Fusion Drive logical volume group. When this completes, you’ll get something called a Core Storage LGV UUID. Copy that number, you’ll need it for the following step.
5. Quit Terminal and begin a fresh install of Mountain Lion onto the new disk called “Macintosh HD”.
6. Restore your apps, files and system settings from the Time Machine backup using the Migration Assistant once you’ve booted up. Here’s an article that shows you how to do that. When that completes, you’re done!
Was it worth it? Yes. The boot-up time went from 45-60 seconds to 15 seconds, right away. And over time, the apps and files I use most often will be moved onto the SSD, thus decreasing the amount of time it’ll take to open and save them.
At some point, I expect Apple to issue a utility, like Boot Camp, that will allow us to do this more easily and automatically. Until then, that’s how I set up Fusion Drive on my iMac, and I hope it’s been helpful to you!
ioSafe, the company famous for its line of rugged external drives that can withstand disasters such as floods, fires and even crushing weight, has come up with a new product: the N2 NAS (Network Attached Storage) device.
The N2 device comes at the right time. The market for NAS devices is maturing and demand is growing. Western Digital has even come out with a line of hard drives, the WD Red, specifically targeted to NAS enclosures. To my knowledge there is no such other NAS device out there, so ioSafe’s got the lead on this.
The N2 appliance is powered by Synology® DiskStation Manager (DSM) and is aimed at the SOHO, SMB and Remote Office Branch Office (ROBO) markets.
The high performance 2-bay N2 provides up to 8TB of storage capacity and is equipped with a 2GHz Marvel CPU and 512MB of memory. The N2 uses redundant hard drives as well as ioSafe’s patented DataCast, HydroSafe and FloSafe technologies to protect data from loss in fire up to 1550°F and submersion in fresh or salt water up to a 10 foot depth for 3 days.
Local and Remote File Sharing: Between virtually any device from any location online
Cloud Station: File syncing between multiple computers and N2 (like Dropbox)
Surveillance Station: Video surveillance application
Media Server: Stream videos and music
Photo Sharing: Photo sharing with friends and family
Mail Server: Email server
VPN Server: Manage Virtual Private Network
Download Station: Post files for others to download
Audio Station: Stream audio to smartphone (iOS/Android)
FTP Server: Remote file transfers
Multi-platform compatibility with Mac/PC/MS Server/Linux
Dual Redundant Disk, RAID 0/1, Up to 8TB (4TB x 2)
2GHz Marvel CPU and 512MB memory
Gigabit Ethernet Port
Additional ports for USB 3, SD Memory Card
User replaceable drives
Protects Data From Fire: DataCast Technology. 1550°F, 1/2 hr per ASTM E119 with no data loss.
Protects Data From Flood: HydroSafe Technology. Full immersion, 10 ft. 3 days with no data loss.
FloSafe Vent Technology: Active air cooling during normal operation. FloSafe Vents automatically block destructive heat during fire by water vaporization – no moving parts
1 Year No-Hassle Warranty + Data Recovery Service (DRS) Standard (for loaded N2)
DRS included $2500/TB for forensic recovery costs for any reason if required
DRS and Warranty are upgradeable to 5 years ($.99/TB per month)
DRS Pro available includes $5000/TB + coverage of attached server ($2.99/TB per month)
Operating: 0-35° C (95°F)
Non-operating: 0-1550°F, 1/2 hr per ASTM E119
Operating Humidity: 20% – 80% (non-condensing)
Non-operating Humidity: 100%, Full immersion, 10 feet, 3 days, fresh or salt water
Size: 5.9″W x 9.0″H x 11.5″L
Weight: 28 lbs
The N2 appliance is being brought to market with funding obtained through IndieGogo. I know it’s hard to believe it when you look at their products, but ioSafe only has about 20 employees. Sometimes they have to be creative in the ways they fund their R&D.
The ioSafe N2 will begin shipping in January 2013 and will be available in capacities up to 8TB. Introductory pricing for the ioSafe N2 diskless version is available for $499 on Indiegogo ($100 off the retail price of $599.99) if you want to get your own hard drives.
Given my intensive computing needs, I naturally looked toward the Mac Pro, but I was disappointed to find that its specs are lagging behind the times. The differences between it and what other Apple computers offer are enough for me to hold off on making the purchase.
The areas where it needs to improve can (almost) be summarized by this screenshot:
The iMac already has Thunderbolt ports, but the Mac Pro (the top-of-the-line model) doesn’t. Why isn’t it there already?
If Thunderbolt is “the fastest, most versatile I/O in a desktop”, why is it MIA on the Mac Pro? Shouldn’t it be on Apple’s most powerful desktop? If it’s got 10 Gbps data channels, and it’s tens of times faster than FW800 or USB 2.0, why isn’t it on the Mac Pro?
It even made it onto the MBP… Will Thunderbolt make it onto the MacBook before it makes it onto the Mac Pro?
The Mac Pro’s internal architecture can support SATA 6Gb/s speeds. There are already SATA 6Gb/s hard drives on the market, at very affordable prices. Why doesn’t the Mac Pro, where storage bandwidth really matters, include SATA 6Gb/s technology? Why is it still stuck at SATA 3Gb/s?
Where is USB 3.0 on the Mac Pro? It still only has USB 2.0 ports. I already have USB 3.0 peripherals, which I wouldn’t be able to use at their native speed if I got a Mac Pro.
Furthermore, where is USB 3.0 on any Mac? USB 3.0 is here to stay, it’s fast, and it’s on a lot of peripherals. It goes without saying that Apple needs to include it on the Mac Pro and on the rest of its computers.
What’s worrying is that Mac OS X Lion may not ship with support for USB 3.0, as this article suggests. No, no, no, Apple, please don’t do that…
This is more of an annoyance, but it’s worth mentioning. Ever since it came out, I always dreamed of owning a 30-inch Cinema Display. Given some of the stuff I’m doing these days, I could really use it, too. But Apple no longer makes it. The only large display they make these days is the 27-inch Cinema Display, which is very nice, and it uses LED instead of LCD technology, but it’s not a 30-inch, is it?
What’s the motivation behind this post? It’s simple. For Ligia’s Kitchen, it costs me about 10.5 GB for 5 minutes of final, edited footage of show, with a one-camera setup. What goes into the 10.5GB? There’s the raw footage (and sound files, if I use a standalone mic), the edits, and the final, published footage. When I use two cameras, the space needed can easily go up by 1.5-2.5x, depending on the shots I need to get. I shoot and edit in 1080p, and output to 720p.
My storage needs are okay for now. I’ve got plenty of space, and if I keep going at this rate, I should be fine. But… and there’s always a but, isn’t there… I have more show ideas in mind. And there’s the hypothetical possibility of shooting with a RED camera at some point in the future, if certain factors come together to allow it. So I’m thinking ahead.
Current hard drive technology (bits of data on disks) has certainly come a long way. Those of us who’ve been in the business long enough know what prices used to be like for capacities that are laughable by today’s standards. Back in 1999, I paid $275 for a 27GB hard drive. My laptop’s drive in college could store a grand total of 120MB. And when I began to learn programming, I’d load the code into memory from tape…
I remember being really excited about Hitachi’s new Perpendicular Magnetic Recording Technology, which came out in early 2006. They even had an animation on their website, which they’ve taken down since. That technology is behind all of the new hard drives that are on the market today, by the way. Hitachi came up with a way to get the bits of data to stand up (hence the term perpendicular) instead of lying down on hard drive platters, thus doubling the amount of data that could be stored onto them.
There are two roads ahead when it comes to data storage, of which one is more likely to succeed:
Optical storage (this is probably the future of storage)
Let’s first look at biological storage. One particular article made the rounds lately: researchers at the Chinese University in Hong Kong have managed to store 90GB of data in 1g of bacteria. While it sounds exciting, the idea of storing my data in petri dishes on my desk doesn’t readily appeal to me, and certain complications come up:
1g of bacteria is about 10 million cells (that’s a LOT); one must start thinking about the potential for bio hazards when you work with bacteria.
The data is stored in a bacteria’s DNA, which means it’s encrypted (a good thing), but it’s also subject to significant mutation (a bad thing) and it takes a long time to retrieve it because you need a gene sequencer, which is tedious and expensive (a bad thing).
I’m not against this. Hey, if they can make it safe and fast, okay. But I believe this is going to be relegated to special applications. The article suggests the technique is currently used to store copyright information for newly created organisms (I wonder how many new bacteria researchers as a whole have created, and is it any wonder antibiotics have such a hard time working against them when we keep playing God). I also see this sort of data storage as a way for spies to operate, or for governments to keep certain secrets.
Okay, onto more cheery stuff, like optical storage. I’ve always thought there was massive potential here, and am glad to see significant work has already been done to make this a reality. There are two technologies which are feasible, according to research that’s already been done:
HDSS (Holographic Data Storage Systems), which so far can store up to 1TB of data in a crystal the size of a sugar cube, but doesn’t yet allow rewrites
These developments are very encouraging. Optical storage is safe, and its potential capacities are huge, possibly endless. And when you think about computer hardware, and how manufacturers are looking at using optical technology in the bridges and buses and wires inside the hardware, because it’s incredibly fast, you start to see how optical makes sense. Let’s also not forget fiber optic cabling, and its incredible capacity to carry data. It certainly looks like optical is the future!
So what’s going to happen to the standard 3.5″ form factor of today’s hard drives? Well, it’s likely that it will stay the same, even though it the storage technology inside it might change. We’ll have crystals and lasers instead of platters and heads, but they’ll likely be able to fit them in there somehow. I don’t think we’ll need to start keeping crystal libraries on our desks, like in Superman’s Crystal Cave, and sticking various-sized crystals into our computers any time soon, although it did look pretty cool when Christopher Reeve did it in the movie.
It really all depends on how soon this new technology will come to market. Right now, there’s clearly enough vested interest in the 3.5″ and 2.5″ form factors to motivate drive manufacturers to shoehorn the new technologies into those shapes, but if optical hard drives won’t be here for the next 5-10 years, then it’s possible that the form factor will change as well. We are after all moving to smaller, sleeker shapes for most computers, notebooks and desktops alike.