Storage Soup


January 27, 2009  5:59 PM

WD launches 2 TB desktop drive

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

Like their 1 TB brethren before them, 2 TB drives are showing up first at the desktop — beginning with the shipment this week of a 2 TB version of Western Digital’s Caviar Green 3.5-inch SATA drive.

The drive follows two generations of high-capacity desktop drives in the Green line. The first was a four-platter 1 TB drive with a 16 MB cache, followed by a 3-platter 1 TB drive with a 32 MB cache. The new 2 TB version uses four 500 GB platters and eight drive heads and also includes a 32 MB cache.

In addition to cramming more data into the same drive footprint, the WD Green line, as its name indicates, targets power-conscious users with drive firmware features that manage energy efficiency. This includes “IntelliPower,” described by WD as “a finely tuned balance” of spin speed, transfer rates, and caching algorithms designed to keep the drive’s power draw just over 10 watts while reading and writing and at 10 watts or under while idle.

Other 1 TB 3.5-inch drives, such as Seagate’s Barracuda, draw about 8 watts while idle. WD’s goal with this drive was to double the capacity without doubling the power draw, according to Caviar Green product manager Mojgan Pessian. Because energy efficiency and capacity are the focus of this product, she said, the drive’s exact RPM–somewhere between 5400 and the typical desktop 7200–is not being disclosed. “This is a low-RPM drive,” she said. “That’s how it’s [able to be] low-power.”

According to IDC analyst John Rydning, these drives will probably find their way into what IDC calls personal storage devices, boxes from Iomega, Buffalo and others sold to consumers and prosumers for backup. “Last year, with one terabyte hard drives, you saw two terabyte solutions in the market,” he said. “This year, we should see those devices offer 4 TB in the same form factor.”

Enterprise users are growing wary of ever-increasing drive sizes, as big drives can make failures more devastating, and double-drive failures more likely. For consumers, though, Rydning pointed out that drives like this in personal storage devices are most often used for less critical copies of data rather than for ‘primary’ storage.

January 27, 2009  3:37 AM

Welcome to Storage Soup’s new home on IT Knowledge Exchange

Dave Raffo Dave Raffo Profile: Dave Raffo

I’d like to take a moment to introduce you to some of our new blog features and also some of the features on IT Knowledge Exchange.

Instead of a long list of categories, we now have a Tag Cloud. Click any topic in the Tag Cloud and you’ll see only posts on that topic. The Tag Cloud is dynamic, so the more a tag is used, the larger and darker it will appear. This helps you quickly see the most popular topics.

You’ll also notice we’ve integrated more of our related editorial content in the right sidebar. If you’re on a post about a specific topic and wish to know more after reading the post, be sure to browse the links in the right sidebar.

We always appreciate your sharing our content on social networking sites and we’ve increased the number of bookmarking tools from four to forty-three. If you enjoy a post, please be sure to share.

Look near the top of the page and you’ll see a row of tabs. You can click the IT Blogs tab to find dozens of technology blogs, both user-generated and TechTarget editorial blogs. You can even request your own blog.

There is also a tab labeled IT Answers. This is where you can ask your own IT question and have it seen by thousands of IT Knowledge Exchange members. So be sure to pose your storage question, browse thousands of storage answers or help out a fellow IT pro by answering a question.

Thank you for stopping by and be sure to bookmark our new blog location and visit the storage section on IT Knowledge Exchange.


January 26, 2009  9:27 PM

Storage vendors put together ESX iSCSI cookbook

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

Just came across a pretty interesting resource on EMC’er Chad Sakac’s Virtual Geek blog (first brought to my attention by Stephen Foskett). It’s a guide to ESX and iSCSI co-developed by, among others, Andy Banta of VMware, Vaughn Stewart of NetApp, Eric Schott of Dell/EqualLogic, Adam Carter of HP/Lefthand, and David Black of EMC.

The post gets into nitty-gritty details and even includes what look like scanned-in napkin drawings to illustrate some of the complexities of performance management using ESX 3.x server with iSCSI. There are multiple links to futher resources on everything from the fundamentals of link aggregation to the full iSCSI spec.

But the bottom line for storage users is that “the ESX 3.x software initiator only supports a single iSCSI session with a single TCP connection for each iSCSI target…So, no matter what MPIO setup you have in ESX, it doesn’t matter how many paths show up in the storage multipathing GUI for multipathing to a single iSCSI Target, because there’s only one iSCSI initiator port.”

There are ways around it–in short, the post states, “Use the ESX iSCSI software initiator. Use multiple iSCSI targets. Use MPIO at the ESX layer. Add Ethernet links and iSCSI targets to increase overall throughput. Ser your expectation for no more than ~160MBps for a single iSCSI target.”

There’s also a workaround for single LUNs needing more than 160 MBps, using an iSCSI initiator in the guest along with MPIO, though the post acknowledges, “It has a big downside…you need to manually configure the storage inside each guest, which doesn’t scale particularly well from a configuration standpoint – so for most customers [say] they stick with the ‘keep it simple’ method.”

The best news out of this post for VMware and iSCSI users, though, is probably the pre-announcement that this behavior will be changing in future ESX releases.


January 23, 2009  10:51 PM

Storage Headlines Podcast for 01-23-09

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

Stories referenced:

Last week’s headlines


January 23, 2009  5:44 PM

EMC eyes primary dedupe for NAS

Dave Raffo Dave Raffo Profile: Dave Raffo

EMC is preparing an upgrade to its Celerra NAS platform, including built-in deduplication for file systems. Internally, at least, EMC refers to this as primary deduplication although it is obviously limited to files and best suited for shared folders and home directories.

EMC hasn’t yet disclosed the upgrade and refused comment, but industry sources and EMC briefing materials obtained by SeachStorage indicate four new Celerra NS models are coming soon. Two of the new models will support solid state drives (SSDs). The dedupe is based on EMC’s Avamar host-based software with RecoverPoint compression.

The only major vendor offering dedupe for primary data is NetApp, which is also EMC’s main NAS competitor. NetApp Deduplication handles block-based data as well as files, although NetApp executives indicate it is used frequently for home directories. EMC documents claim a 30% to 40% primary storage savings is possible for “typical unstructured file share datasets” in primary and archive storage with the dedupe.

The Celerra dedupe will compress files with low usage activity, and single instance files to remove duplicates. EMC is also adding a compliance option to its Celerra File-Level Retention WORM software as a competitor to NetApp SnapLock. FLR-C, as EMC calls its new option, locks files to prevent file system deletions and has a non-spoofable clock to honor file retention time. To avoid competing with its Centera archiving system, EMC will recommend Celerra for archiving and locking files on its NAS systems and Centera for application data and fixed content.

The new models include two enterprise systems – the NS G8 gateway and NS-960 – and two midrange systems – the NS-480 and NS-120. The NS G8 replaces the NSX and NS80G, and the NS-960 replaces the NS80. The NS-480 will replace the NS40 and the NS-240 will replace the NS20. All of the systems support higher capacities than their predecessors, and the NS-480 and NS-960 will support flash SSDs.


January 15, 2009  5:20 PM

Seagate 1 TB Barracuda users report failures

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

More than 20 users have posted to a thread on Seagate’s official Community Forums reporting that their 1 TB 7200 RPM Seagate Barracuda desktop drives have failed. The drives failed soon after purchase, from as little as 2 weeks to as long as seven months; several users reported the problem to be the drive becoming undetectable to the BIOS. Users reporting failed drives on the thread also said the drives had been manufactured in Thailand.

Freezing problems were previously reported for the 1.5 TB version of the 7200.11. Seagate issued a firmware fix for those drives in late November.

Seagate officials did not comment when contacted by Storage Soup today.

UPDATE–Please see the comments section below for a response about a firmware fix from Seagate.


January 15, 2009  2:09 PM

SSDs: Your mileage almost certainly will vary

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

Maybe this is what happens with any brand-new technology, but so far there’s been such wide variability in the solid-state drives that have been announced for the enterprise since EMC added STEC drives to Symmetrix last January that I can’t help but be curious about it. Here are the specs available on some of the latest SSDs rolled out:

  • STEC (as used by EMC):  73 GB and 146 GB, FC interface, 52,000 sustained random read IOPS; 17,000 sustained random write IOPS; 250 MBps sustained, sequential reads; and 200 MBps sustained, sequential writes.
  • Intel X-25 E: 32 and 64 GB, 170 MBps sequential write; 4 KB random-read IOPS 35,000; random writes at 3,300. 10 I/O channels on controller. Claims smaller write amplification than other drives.
  • pureSilicon: SLC capacities of 256 and 512 GB. MLC capacity of 1 TB. Up to 50,000 random read IOPS. 32 channels on controller.
  • Samsung: 100 GB; 25,000 random read IOPS; 6,000 random write IOPS; read data sequentially at 230 MBps and write sequentially at 180 MBps. Also claims to have a new ‘recipe’ for NAND Flash material itself that could boost SSD durability, but hasn’t decided when, where, or how to release it to OEMs.

This week I’ve also come across Toshiba, which announced a new SAS interface SLC drive in a 100 GB capacity this week. But it’s the darndest thing–another drive, another very different claim about random read and write IOPS–for Toshiba, the numbers are much closer together than they are for others with these two specs: 25,000 random read IOPS, and 20,000 random write IOPS.

The problem is trying to get any of these vendors to drill down further. Most of them won’t because it’s getting into proprietary information, which I understand, but I wish it was somehow otherwise. At most, I’m told it has to do with the number of channels in the controller, the Flash recipe, how many channels are dedicated to reads vs. writes, how data streams are interleaved and/or cached with DRAM on their way into or out of the Flash capacity. It gives me some idea, but I can’t wait until users start testing these in volume, in real world environments.

I also wonder whether the “biodiversity” of creatures in the SSD habitat will continue, or whether eventually enterprise SSDs will commoditize like any other storage medium. There’s some debate about this in the industry so far.


January 15, 2009  11:57 AM

Storage Headlines for 01-14-2009

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

We’re starting up a new weekly podcast, hosted by yours truly, to review the top stories of the last week you may have missed. Here’s our first installment.

Stories referenced:


January 13, 2009  4:34 PM

Pillar claims SPC-1 supremacy

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

Pillar Data Systems has published its first system test benchmarks for the Axiom 600 disk array via the Storage Performance Council (SPC). It tested a system with 42 total TB, including 10 TB used and mirrored to allocate a total of about 20 TB. The results were 64,992 total IOPS.

These numbers come in ahead of competitive systems from IBM, NetApp and EMC. The EMC CX 3-40 was tested last January by archrival NetApp, making its SPC benchmarks controversial, but as listed a 22 TB system with 8.5 TB used and mirrored produced 24,997 SPC-1 IOPS. A NetApp 3170 with 32 TB and 19.6 used and mirrored resulted in 60,515 June 10. A 37.5 TB IBM 5300 with 13.7 TB used and mirrored produced 58,158 SPC-1 IOPS Sept. 25.

I found it interesting that with one 2,000-IOPS deviation between the IBM 5300 and NetApp 3170, the systems generally performed better according to which had been most recently tested. Note also how much of an outlier EMC is, both in terms of capacity used and total capacity. It was also an outlier in its free space, with just under 1 TB unused. IBM and NetApp both left approximately 5 TB of free space in their configurations, and Pillar had 16 TB of free space.

I do have to wonder how much weight users give to these industry benchmarks when selecting a product.  NetApp’s submitting EMC systems to SPC, a flap last summer over server virtualization benchmark testing, and  continued inconsistency among vendors as to who submits systems for benchmarking leaves a lot of potential reasons to take benchmarks with a grain of salt.


January 12, 2009  5:05 PM

Startup claims 50,000 random read IOPS on 1 TB SSD

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

A company called pureSilicon came out of stealth last week at CES with new solid state drives in 256 GB, 512 GB and 1 TB capacities. The drives, expected to ship this summer, also include a proprietary 32-channel controller architecture that company founder and president Jason Breakstone said has been clocked at 50,000 random read IOPS.

By comparison, the drives EMC Corp. ships today from STEC are 73 GB and 146 GB. Intel’s X-25 E SSDs have 10-channel controllers and are offered in 32 GB and 64 GB capacities, and recently announced enterprise SSDs from Samsung have 8-channel controllers and are offered at 100 GB capacity. This is an already crowded market, but if pureSilicon can do what it says it’s going to do, it’s found some differentiators already.

A 1 TB SSDs might grab enterprise customers’ attention, but the drive is manufactured using multi-level cell (MLC) technology. The 256 GB and 512 GB sizes are single-level cell (SLC) drives, which generally have a longer lifecycle, and are viewed as more reliable than MLC drives because only one layer of data is stored in each Flash cell at a time. For now,  pureSilicon offers the 1 TB MLC drives for enterprise applications with a three-year warranty. Breakstone maintains the the bigger the SSD, the more compelling its value proposition for consolidating large numbers of short-stroked hard drives.

On the other side, however, is the expense of SSDs, despite the fact that Flash pricing has declined over the last year. This is part of the reason for smaller capacities for SSDs so far. Breakstone said pricing won’t be set for the new 1 TB behemoth until closer to its release date, but with high capacity and high I/O, “our product can perform at the level of a larger array. If you can achieve the same results using a factor of 10 or 100 fewer drives, it’s a win-win.”


Forgot Password

No problem! Submit your e-mail address below. We'll send you an e-mail containing your password.

Your password has been sent to: