It took him awhile, but Eric Kelly is running the Snap Appliance business again.
Kelly today became CEO of Overland Storage, which acquired the Snap business from Adaptec last June. He replaces Vern LoForti, who remains at Overland as president.
Kelly has a long history with Snap. He put together a group of investors to buy Snap from Quantum for $10 million in 2002 and served as its CEO until selling it to Adaptec for $100.4 million in 2004. Kelly worked as GM of Adaptec’s storage business for two years and when Adaptec put the Snap business up for sale, there were rumblings in the storage industry that Kelly tried to put together a group to buy it back. That didn’t work out, but he joined Overland’s board in late 2007 and strongly recommended Overland buy the Snap business.
When that deal was completed, LoForti said Kelly told the Overland board, ‘”If you don’t want [Snap], I’ll buy it myself.”
Kelly told SearchStorage.com today he’s looking forward to working again with other members of the Snap team who remain from his CEO days.
“Some of the same team is here,” he said. “I think we do have a little advantage in terms of understanding what our customers are looking for, and how we position the product and grow the business.”
So with Overland trying to go from a tape vendor to a storage systems company built around its new NAS products, Kelly seems like the right guy for the job. LoForti has done his best to revitalize Overland after moving CFO to CEO of the troubled company in 2007, but had to spend much of his time looking for financing to keep it afloat.
Overland raised $9 million in financing last Decmeber to keep the doors open but will need to reverse its long history of losing money if it is to survive. The vendor last week said it would reduce its workforce by 17% by cutting 53 employees and slashed the salaries of executives and other salaried employees by 10%. That followed previous layoffs totaling 64 employees since last August.
Kelly said Overland will seek more funding and will obviously strive to become profitable as soon as possible. Overland’s strategy will be to provide end to end data protection across disk and tape, which involves leveraging Snap’s software with Overland’s other platforms.
LoForti says because the Snap Guardian OS and the OS for Overland’s REO disk appliances are based on the same Linux kernel, it won’t be difficult to integrate the product lines.
“When you sell and appliance, people look at the hardware, but our value is in the software,” Kelly said.
Like their 1 TB brethren before them, 2 TB drives are showing up first at the desktop — beginning with the shipment this week of a 2 TB version of Western Digital’s Caviar Green 3.5-inch SATA drive.
The drive follows two generations of high-capacity desktop drives in the Green line. The first was a four-platter 1 TB drive with a 16 MB cache, followed by a 3-platter 1 TB drive with a 32 MB cache. The new 2 TB version uses four 500 GB platters and eight drive heads and also includes a 32 MB cache.
In addition to cramming more data into the same drive footprint, the WD Green line, as its name indicates, targets power-conscious users with drive firmware features that manage energy efficiency. This includes “IntelliPower,” described by WD as “a finely tuned balance” of spin speed, transfer rates, and caching algorithms designed to keep the drive’s power draw just over 10 watts while reading and writing and at 10 watts or under while idle.
Other 1 TB 3.5-inch drives, such as Seagate’s Barracuda, draw about 8 watts while idle. WD’s goal with this drive was to double the capacity without doubling the power draw, according to Caviar Green product manager Mojgan Pessian. Because energy efficiency and capacity are the focus of this product, she said, the drive’s exact RPM–somewhere between 5400 and the typical desktop 7200–is not being disclosed. “This is a low-RPM drive,” she said. “That’s how it’s [able to be] low-power.”
According to IDC analyst John Rydning, these drives will probably find their way into what IDC calls personal storage devices, boxes from Iomega, Buffalo and others sold to consumers and prosumers for backup. “Last year, with one terabyte hard drives, you saw two terabyte solutions in the market,” he said. “This year, we should see those devices offer 4 TB in the same form factor.”
Enterprise users are growing wary of ever-increasing drive sizes, as big drives can make failures more devastating, and double-drive failures more likely. For consumers, though, Rydning pointed out that drives like this in personal storage devices are most often used for less critical copies of data rather than for ‘primary’ storage.
I’d like to take a moment to introduce you to some of our new blog features and also some of the features on IT Knowledge Exchange.
Instead of a long list of categories, we now have a Tag Cloud. Click any topic in the Tag Cloud and you’ll see only posts on that topic. The Tag Cloud is dynamic, so the more a tag is used, the larger and darker it will appear. This helps you quickly see the most popular topics.
You’ll also notice we’ve integrated more of our related editorial content in the right sidebar. If you’re on a post about a specific topic and wish to know more after reading the post, be sure to browse the links in the right sidebar.
We always appreciate your sharing our content on social networking sites and we’ve increased the number of bookmarking tools from four to forty-three. If you enjoy a post, please be sure to share.
Look near the top of the page and you’ll see a row of tabs. You can click the IT Blogs tab to find dozens of technology blogs, both user-generated and TechTarget editorial blogs. You can even request your own blog.
There is also a tab labeled IT Answers. This is where you can ask your own IT question and have it seen by thousands of IT Knowledge Exchange members. So be sure to pose your storage question, browse thousands of storage answers or help out a fellow IT pro by answering a question.
Thank you for stopping by and be sure to bookmark our new blog location and visit the storage section on IT Knowledge Exchange.
Just came across a pretty interesting resource on EMC’er Chad Sakac’s Virtual Geek blog (first brought to my attention by Stephen Foskett). It’s a guide to ESX and iSCSI co-developed by, among others, Andy Banta of VMware, Vaughn Stewart of NetApp, Eric Schott of Dell/EqualLogic, Adam Carter of HP/Lefthand, and David Black of EMC.
The post gets into nitty-gritty details and even includes what look like scanned-in napkin drawings to illustrate some of the complexities of performance management using ESX 3.x server with iSCSI. There are multiple links to futher resources on everything from the fundamentals of link aggregation to the full iSCSI spec.
But the bottom line for storage users is that “the ESX 3.x software initiator only supports a single iSCSI session with a single TCP connection for each iSCSI target…So, no matter what MPIO setup you have in ESX, it doesn’t matter how many paths show up in the storage multipathing GUI for multipathing to a single iSCSI Target, because there’s only one iSCSI initiator port.”
There are ways around it–in short, the post states, “Use the ESX iSCSI software initiator. Use multiple iSCSI targets. Use MPIO at the ESX layer. Add Ethernet links and iSCSI targets to increase overall throughput. Ser your expectation for no more than ~160MBps for a single iSCSI target.”
There’s also a workaround for single LUNs needing more than 160 MBps, using an iSCSI initiator in the guest along with MPIO, though the post acknowledges, “It has a big downside…you need to manually configure the storage inside each guest, which doesn’t scale particularly well from a configuration standpoint – so for most customers [say] they stick with the ‘keep it simple’ method.”
The best news out of this post for VMware and iSCSI users, though, is probably the pre-announcement that this behavior will be changing in future ESX releases.
- Seagate’s new CEO faces declining sales, faulty drives
- Nexsan offers SAS, SATA, spin-down in one box
- Dot Hill rolls out 2.5-inch SAS, SSD systems
- Symantec to integrate Enterprise Vault with the cloud
- Enterprise Vault supports data deduplication
- Survey finds data centers understaffed
EMC is preparing an upgrade to its Celerra NAS platform, including built-in deduplication for file systems. Internally, at least, EMC refers to this as primary deduplication although it is obviously limited to files and best suited for shared folders and home directories.
EMC hasn’t yet disclosed the upgrade and refused comment, but industry sources and EMC briefing materials obtained by SeachStorage indicate four new Celerra NS models are coming soon. Two of the new models will support solid state drives (SSDs). The dedupe is based on EMC’s Avamar host-based software with RecoverPoint compression.
The only major vendor offering dedupe for primary data is NetApp, which is also EMC’s main NAS competitor. NetApp Deduplication handles block-based data as well as files, although NetApp executives indicate it is used frequently for home directories. EMC documents claim a 30% to 40% primary storage savings is possible for “typical unstructured file share datasets” in primary and archive storage with the dedupe.
The Celerra dedupe will compress files with low usage activity, and single instance files to remove duplicates. EMC is also adding a compliance option to its Celerra File-Level Retention WORM software as a competitor to NetApp SnapLock. FLR-C, as EMC calls its new option, locks files to prevent file system deletions and has a non-spoofable clock to honor file retention time. To avoid competing with its Centera archiving system, EMC will recommend Celerra for archiving and locking files on its NAS systems and Centera for application data and fixed content.
The new models include two enterprise systems – the NS G8 gateway and NS-960 – and two midrange systems – the NS-480 and NS-120. The NS G8 replaces the NSX and NS80G, and the NS-960 replaces the NS80. The NS-480 will replace the NS40 and the NS-240 will replace the NS20. All of the systems support higher capacities than their predecessors, and the NS-480 and NS-960 will support flash SSDs.
More than 20 users have posted to a thread on Seagate’s official Community Forums reporting that their 1 TB 7200 RPM Seagate Barracuda desktop drives have failed. The drives failed soon after purchase, from as little as 2 weeks to as long as seven months; several users reported the problem to be the drive becoming undetectable to the BIOS. Users reporting failed drives on the thread also said the drives had been manufactured in Thailand.
Freezing problems were previously reported for the 1.5 TB version of the 7200.11. Seagate issued a firmware fix for those drives in late November.
Seagate officials did not comment when contacted by Storage Soup today.
UPDATE–Please see the comments section below for a response about a firmware fix from Seagate.
Maybe this is what happens with any brand-new technology, but so far there’s been such wide variability in the solid-state drives that have been announced for the enterprise since EMC added STEC drives to Symmetrix last January that I can’t help but be curious about it. Here are the specs available on some of the latest SSDs rolled out:
- STEC (as used by EMC): 73 GB and 146 GB, FC interface, 52,000 sustained random read IOPS; 17,000 sustained random write IOPS; 250 MBps sustained, sequential reads; and 200 MBps sustained, sequential writes.
- Intel X-25 E: 32 and 64 GB, 170 MBps sequential write; 4 KB random-read IOPS 35,000; random writes at 3,300. 10 I/O channels on controller. Claims smaller write amplification than other drives.
- pureSilicon: SLC capacities of 256 and 512 GB. MLC capacity of 1 TB. Up to 50,000 random read IOPS. 32 channels on controller.
- Samsung: 100 GB; 25,000 random read IOPS; 6,000 random write IOPS; read data sequentially at 230 MBps and write sequentially at 180 MBps. Also claims to have a new ‘recipe’ for NAND Flash material itself that could boost SSD durability, but hasn’t decided when, where, or how to release it to OEMs.
This week I’ve also come across Toshiba, which announced a new SAS interface SLC drive in a 100 GB capacity this week. But it’s the darndest thing–another drive, another very different claim about random read and write IOPS–for Toshiba, the numbers are much closer together than they are for others with these two specs: 25,000 random read IOPS, and 20,000 random write IOPS.
The problem is trying to get any of these vendors to drill down further. Most of them won’t because it’s getting into proprietary information, which I understand, but I wish it was somehow otherwise. At most, I’m told it has to do with the number of channels in the controller, the Flash recipe, how many channels are dedicated to reads vs. writes, how data streams are interleaved and/or cached with DRAM on their way into or out of the Flash capacity. It gives me some idea, but I can’t wait until users start testing these in volume, in real world environments.
I also wonder whether the “biodiversity” of creatures in the SSD habitat will continue, or whether eventually enterprise SSDs will commoditize like any other storage medium. There’s some debate about this in the industry so far.
We’re starting up a new weekly podcast, hosted by yours truly, to review the top stories of the last week you may have missed. Here’s our first installment.
- Seagate drives out its CEO
- IBM changes storage chiefs, but says it’s routine
- Microsoft adds cloud backup to Data Protection Manager
- EMC overhauls Retrospect backup software for Mac
- Samsung launches bigger SSDs to challenge Intel
Pillar Data Systems has published its first system test benchmarks for the Axiom 600 disk array via the Storage Performance Council (SPC). It tested a system with 42 total TB, including 10 TB used and mirrored to allocate a total of about 20 TB. The results were 64,992 total IOPS.
These numbers come in ahead of competitive systems from IBM, NetApp and EMC. The EMC CX 3-40 was tested last January by archrival NetApp, making its SPC benchmarks controversial, but as listed a 22 TB system with 8.5 TB used and mirrored produced 24,997 SPC-1 IOPS. A NetApp 3170 with 32 TB and 19.6 used and mirrored resulted in 60,515 June 10. A 37.5 TB IBM 5300 with 13.7 TB used and mirrored produced 58,158 SPC-1 IOPS Sept. 25.
I found it interesting that with one 2,000-IOPS deviation between the IBM 5300 and NetApp 3170, the systems generally performed better according to which had been most recently tested. Note also how much of an outlier EMC is, both in terms of capacity used and total capacity. It was also an outlier in its free space, with just under 1 TB unused. IBM and NetApp both left approximately 5 TB of free space in their configurations, and Pillar had 16 TB of free space.
I do have to wonder how much weight users give to these industry benchmarks when selecting a product. NetApp’s submitting EMC systems to SPC, a flap last summer over server virtualization benchmark testing, and continued inconsistency among vendors as to who submits systems for benchmarking leaves a lot of potential reasons to take benchmarks with a grain of salt.