Storage Soup

June 11, 2008  4:51 PM

Symantec’s SwapDrive acquisition draws glares from EMC

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

According to a statement issued by Symantec today,

Symantec has acquired SwapDrive, a privately-held online storage company to strengthen the services offerings in the Norton consumer portfolio and to help consumers manage data across their devices. This was a small, targeted acquisition and is a very natural move for us because of our close two-year OEM relationship and existing product partnership on Norton 360.  

If reports elsewhere are to be believed, however, “small” and “targeted” are relative terms–the deal is reportedly worth $123 million.

EMC, which made a similar acquisition of Mozy for $76 million last year, is already firing off counter punches, via an emailed statement to press pointing out competitive differences. SwapDrive doesn’t backup open files, for example, doesn’t have a Mac client, and charges $50 per year for 2 GB of backup while Mozy gives that away for free.  On the other hand, Symantec already had a field-tested infrastructure for SaaS in its Symantec Protection Network (originally used for hosting accounts in the security division); began as a security company while EMC has had to assimilate RSA; and arleady  has an established brand in the consumer/SMB space that EMC is trying to penetrate. Symantec previously partnered with SwapDrive for its Norton 360 backup SaaS, while EMC has had to integrate Mozy. 

Aside from these positioning differences, I haven’t been able to help notice that EMC and Symantec are looking a lot alike these days. While at EMC World and Symantec Vision, both in Vegas just a few weeks apart, at times it’s been eerie just how similar the company lines have begun to sound from these rivals. Both CEOs are keen to talk about the “consumerization of IT.” Both are interested in supporting access to data from mobile devices as part of that shift (though Symantec appears a bit ahead there with application streaming and data loss protection software that’s shipping today to offer that kind of service, while EMC is still cultivating Pi Corp.’s IP). Both have large, multifaceted backup portfolios they say they plan to integrate at the management layer (EMC also says it plans to integrate repositories while Symantec argues that they should remain separate), and both have used the same “one size doesn’t fit all” line to describe their backup portfolios and strategies.

It’s clear from the way Symantec execs react to questions about these overlaps that they still see EMC as a newcomer in the software space. When I pointed out that both companies are talking about the consumerization of IT in similar terms, Symantec CEO John Thompson interrupted me to retort, “and what consumer experience do they have?” He made a similar comment about backup and archiving integration, pointing out Enterpise Vault archives certain data types already while EMC is still working on integrating file, database and email archiving through Documentum. It’s clear these companies are in each other’s heads, and that their competition is growing fiercer than even the turf wars between EMC and NetApp. With this acquisition, SaaS will be added to the list of their increasingly contentious battlegrounds.

June 11, 2008  11:22 AM

Incipient offers data migration without storage virtualization

Dave Raffo Dave Raffo Profile: Dave Raffo

While IBM is upgrading its SAN Volume Controller (SVC) virtualization platform, Incipient is taking one of the major features of its storage virtualization product — data migration — and spinning it off into a separate application.

Incipient today rolled out Incipient Automated Data Migration (iADM), which is designed to automate and manage large data migration projects.

Migration is supported in Incipient Network Storage Platform (iNSP), switch-based virtualization software that competes with EMC Invista and LSI’s StoreAge SVM. It’s no secret that switch-based block virtualization has been a bust compared to IBM’s network-based SVC and Hitachi Data Systems’ array-based virtualization. Incipient isn’t waiting for the market to come around.

[iADM] “is not storage virtualization, it’s process automation for storage,” Incipient marketing VP Rob Infantino said. “This is used with or without storage virtualization. Some customers think storage virtualization is still risky because we’re adding an abstraction layer between the host and storage, but they need data migration capability.”

iADM doesn’t permanently take control of LUNs. While iNSP requires a Fibre Channel switch, iADM sits on a Windows server and connects to the storage fabric via Ethernet. It performs the discovery of devices on a SAN, maps hosts to source LUNs, and does the LUN masking, auto zoning and host reconfiguration. iADM works with array-based data movers, such as EMC’s SRDF and SANCopy. The idea is to make large migration projects go faster and save companies from having to pay huge services fees.  Incipient is aiming iADM at shops with more than a petbyte of storage and looking to migrate hundreds of terabytes. Licenses typically run from $2,000 to $2,500 per TB, depending on how many TBs are migrated.

Since the migration requires a few hours of downtime for hosts to recognize the new targets, it’s usually done on weekends. After the migration, customers can pull out the iADM software.

Infantino says the new product doesn’t mean that Incipient is backing off switch-based virtualization. He won’t reveal how many customers Incipient has, but says it does have some $1 million-plus deals with major financial institutions using iNSP. Now it’s working on upgrading iNSP to add remote replication, support for software from VMware, NetApp’s Onaro and Akorri, and greater scalability.

June 10, 2008  8:52 PM

SSDs vs. 20,000 rpm drives: Who wins? I do!

Tskyers Tory Skyers Profile: Tskyers

It’s been a while since I’ve been able to spend some quality time behind the keyboard, I’ve been suffering from a gigantic “honey-do” list and it was difficult for me to use some huge work emergency to weasel out of it! In the time I haven’t been blogging there’s been some major storage news: Sun says that SSDs are ready for arrays while Western Digital is reportedly developing 20,000 rpm drives.

I’ve been chomping at the bit to get my hands on an SSD for my desktop. After going SAS, I’m open to the prospect of even higher performance for my desktop disk subsystem, and it’s something I think I’m going to be chasing from now on.

We’re currently rolling out SSDs in a limited deployment for highly available single hard disk bay blades (say that three times fast). IBM has managed to fit a RAID 1 setup in a single drive bay for their line of blades and we like the performance numbers as well as the idea of no moving parts at all in a blade with onboard storage. Not only will we have the higher MTBF of the SSDs but the read performance is crazy!

Three months ago I was talking about how SAS would spell doom for SATA. Well, now I’m ready to eat some crow because in no way did I expect SSDs to become this close to affordable this quickly. Take a look at the non-server market: Lenovo and Apple are already offering laptop models with SSD exclusively.

SSDs have a lower power and heat footprint and have great read speeds. Write speeds aren’t as good as the read speeds, but slap a couple together in RAID 0 and that issue becomes moot. SSD looks like a shoe-in to be the next big thing. Or does it. . . ?

My take on the possibility of a 20,000 rpm drive is that Western Digital might not like the idea of the next big thing being something that isn’t, well, theirs. They also just released the 10,000 rpm Velociraptor SATA drive, which in itself something spectacular, since it brings the performance of higher-rpm SAS down to the cheap, ubiquitous SATA controller.

Details are sketchy when it comes to the potential heat and power of the alleged 20,000 rpm drive. It may not even make it to market, and there might not be much place for it with solid state drives delivering even faster performance with less in the way of power and cooling requirements. But me, I’m interested to see a knife fight between traditional disks (and maybe hybrids) and SSDs, since it can only result in me getting a faster storage subsystem, and it may lower prices even more.

In fact, I’m a little annoyed that it’s taken the disk industry this bleeding long to come up with an additional 5,000 rpm. I’m sure some of you out there are in the hard drive industry and have a list of reasons why it’s a hard thing to do. To which I, the jaded technologist/consumer, say, “So?” We live in an age where we have teraflop chips on video cards, where chips in mp3 players have more computing power than the first Space Shuttle, where cars park themselves and where we can see full color photos beamed back from Mars. MARS!! If you ask me, 15,000 rpm has been the ceiling for waaaaaaaay too long.

June 6, 2008  11:50 AM

Dell moves to top of growing iSCSI SAN market

Dave Raffo Dave Raffo Profile: Dave Raffo

IDC confirmed Dell’s claim that it has moved into the No. 1 spot in iSCSI SAN market share following its $1.4 billion acquisition of EqualLogic. According to the IDC quarterly storage numbers for the first quarter released today, Dell passed EMC and NetApp to take the lead with 27.7 percent of the iSCSI market. Overall, IDC pegged iSCSI as accounting for 6.1 percent of the overall external disk storage revenue — up from 4.1 percent a year ago and 5.1 percent in the fourth quarter of 2007.

It appears that a good chunk of the market share Dell gained came at the expense of its storage partner EMC. EMC, which co-markets Clariion systems with Dell, slipped from 18.2 percent to 15 percent of iSCSI market share in one quarter. Does that mean Dell customers are buying EqualLogic systems instead of Clariion AX iSCSI boxes? Maybe, although Dell still accounts for about one-third of overall Clariion sales.

NetApp increased its iSCSI share from 18 percent to 20.5 percent to move ahead of EMC while slipping behind Dell. No other major vendor has more than 4.7 percent of the iSCSI market, with the “other” (all vendors except for Dell, EMC, NetApp, HP, IBM, Hitachi, and Sun) category at 29.3 percent – nearly twice as much as “others” have of the Fibre Channel market. The others’ iSCSI share actually came down – it was 40.1 percent in the fourth quarter — reflecting  the shift of EqualLogic’s revenue from others to Dell.

It’s a good bet that LeftHand Networks sits in fourth place overall with a large piece of the others’ share. LeftHand was considered a close second to EqualLogic among private iSCSI vendors before Dell scooped up EqualLogic. LeftHand remains private and we don’t know its financials, but marketing VP Larry Cormier said it is picking up 200 to 300 customers a quarter. . .some of those because EqualLogic is no longer independent. “Some shops just won’t buy Dell,” he said.

In any case, with iSCSI drivers such as server virtualization and the eventual emergence of 10-Gig Ethernet fueling interest and the vendor landscape changing, the iSCSI space will be interesting to watch over the next year. . .just as converged Fibre Channel/Ethernet networks may blunt iSCSI’s encroachment into the enterprise.

June 6, 2008  8:05 AM

Storage Decisions: Questioning the FCoE hype

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

“Fibre Channel over Ethernet is like a fast car,” said consultant Howard Goldstein of Howard Goldstein Accociates Thursday on his session about FCoE at Storage Decisions Toronto. “It looks great, but it probably won’t run as well as you thought or be as cheap as they say it’s going to be.”

Goldstein’s point basically boiled down to: if Ethernet’s good enough to be the transport layer, why bother layering an FC protocol on top of it? He dismissed the common answer to that question, which is that mixing FC and Ethernet will allow users to maintain existing investments in FC systems, saying it’s a myth. “You’re going to have to buy brand new HBAs and Fibre Channel switches to support FCoE,” he said. “Is this really the time to reinvest in Fibre Channel infrastructure?”

Instead, Goldstein pointed out that FC services such as log-in, address assignment and name server, to name a few, could be done in software. “Those services don’t have to be in the switch–Fibre Channel allows them in the server,” he said. He also questioned the need for a revamping of the Ethernet specification for “Data Center Ethernet” capabilities. “Is converged Ethernet a real requirement or a theoretical requirement?” he said. He also questioned whether or not storage traffic was really fundamentally different from network traffic.

However, users at the show said FCoE is still so new they weren’t sure whether or not to agree with Goldstein. “It’s too immature to say right now,” said Maple Leaf Foods enterprise support analyst Ricki Biala. He also pointed out an all-too-true fact: in the end, such technology decisions will be based on equal parts politics to technology. “It’s easier to convince management to buy in if you’re going the way the rest of the market’s going,” he said.

Your thoughts and quibbles on FCoE are welcome as always in the comments.

June 4, 2008  4:09 PM

Disk vendors driving toward SAS in enterprise storage

Dave Raffo Dave Raffo Profile: Dave Raffo

Solid state isn’t the only thing looming on the horizon in the enterprise storage drive space. Drive makers say small-factor (2.5-inch) SAS is poised to encroach on 3.5-inch Fibre Channel’s turf in storage arrays.

Seagate is eyeing enterprise storage arrays with drives such as the Savvio 10k.3 that it launched this week. At 300GB, the self-encrypting drive offers more than twice the capacity of Seagate’s previous SAS drives. It also supports the SAS 2 interface. SAS 2 includes 6 Gbit/s speed and other enterprise features likely to show up in storage systems by next year.

“300-gig drives will be more attractive to storage vendors, and they’re starting to find the small form factor drives more compelling,” said Henry Fabian, executive director of marketing for Seagate’s enterprise business. “You’ll start to see the small form factor ship in the second half of the year in storage arrays because of higher capacity and lower power requirements.”

Joel Hagberg, VP of business development for Seagate rival Fujitsu Computer Products of America, also sees small form factor SAS coming on strong in enterprise storage. “The storage vendors all recognize there is a shift coming as we get to 300 gigs or 600 gigs in the next couple of years in the 2.5-inch package,” he said. “We’re cutting power in half and the green initiative in storage is increasing.”

As for Fibre Channel, the drive makers agree you won’t see hard drives going above the current 4-Gbit/s bandwidth level.

“Four-gig is the end of the road for Fibre Channel on the device level,” Hagberg said. “All the external storage vendors are looking to migrate to SAS.”

By the way, Hagberg says Fujitsu isn’t buying into the solid state hype for enterprise storage yet. He considers solid state to be a few years away from taking off in storage arrays.

“There’s a lot of industry buzz on solid state, and I have to chuckle,” he said. “I meet with engineers of all storage vendors and talk about the hype versus reality on solid state drives. Every notebook vendor released solid state in the last year. Are any of those vendors happy with those products? The answer is no. The specs of solid state performance look tremendous on paper, but a lot less is delivered in operation.”

June 4, 2008  2:52 PM

EMC doles out fewer refunds now that Insignia website is Google-improved

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

Google has revamped its business search site, and rechristened it Google Site Search (it was previously called Custom Search Business Edition). It’s the SaaS version of the Google Search Appliance, but it’s limited to website data because the hosted software can only see public data.

So essentially it’s custom search for e-commerce websites. Almost completely unrelated to storage. . .except when it came to one of Google’s customer references for Site Search: EMC’s Insignia website, which sells some of EMC’s lower-end products online.  Prior to implementing the site search,  apparently the Insignia site had no search functionality. Visitors had to page through the site manually–including when it came time to look for support documents or troubleshooting tips.

EMC’s webmaster Layla Rudy was quoted in Google’s collateral as saying that sales have gone up 20% since they added search to the site. Moreover, according to her statement, there has been an 85% decrease in customer-requested refunds now that they can find the correct product in the first place as well as its associated support documents. What’s especially amazing about this to me is that Insignia is a relatively new business unit, rolled out by EMC within the last three years–it’s not like it was the 90’s when Google was relatively unknown or site search a “nice to have” feature of most websites.

Of course, I don’t know how long the Insignia site was up without search, or what the absolute numbers are when it comes to the refund decrease–85% can be 8.5 out of 10 or 85 out of 100. (EMC hasn’t returned my calls on this). 

Meanwhile, with EMC getting into cloud computing, I wonder what kind of search — if any — it makes available on backup/recovery or archiving SaaS websites. Right now Google claims its Search Appliance can index onsite and offsite repositories and, unlike the SaaS version, can search protected private data. While there are no plans to make this a feature of the hosted version, service providers can offer hosted search by managing their own appliance in the cloud. Whatever they chose, hopefully it was a Day One feature.

June 3, 2008  3:11 PM

Linux file system ‘firestorm’ fizzles

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

I was intrigued when a colleague sent me a link to an article by Henry Newman referring to a “firestorm” touched off by some remarks he recently made in another article he wrote. The first article addressed the scalability of standalone Linux file systems vs. standalone symmetric multiprocessing (SMP) file systems such as IBM’s GPFS or Sun’s ZFS. His point was that in high-performance computing environments requiring a single file system to handle large files and provide streaming performance, an SMP file system that pools CPU and memory components yields the best performance.

Newman begins his followup piece by writing, “My article three weeks ago on Linux file systems set off a firestorm unlike any other I’ve written in the decade I’ve been writing on storage and technology issues.” He refers later on to “emotional responses and personal attacks.” I’m no stranger to such responses myself, so it’s not that I doubt they occurred, but in poking around on message boards and the various places Newman’s article was syndicated I haven’t been able to uncover any of that controversy in a public place. And I’m not sure why there would be any firestorm.

I hit up StorageMojo blogger and Data Mobility Group analyst Robin Harris yesterday for an opinion on whether what Newman wrote was really that incendiary. Harris answered that while he disagreed with Newman’s contention that Linux was invented as a desktop replacement for Windows, he didn’t see what was so crazy about Newman’s ultimate point: a single, standalone Linux file system (Newman is explicit in the article that he is not referring to file systems clustered by another application) does not offer the characteristics ideal for a high-performance computing environment. “It seems he made a reasonable statement about a particular use case,” was Harris’s take. “I’m kind of surprised at the response that he says he got.”

That said, how do you define the use case Newman is referring to–what exactly is HPC, and how do you draw the line between HPC and high-end OLTP environments in the enterprise? Harris conceded that those lines are blurring, and that moreover, image processing in general is something more and more companies are discovering in various fields that didn’t consider such applications 15 years ago, like medicine. So isn’t the problem Newman is describing headed for the enterprise anyway?

“Not necessarily,” Harris said, because Newman is also referring to applications requiring a single standalone large file system. “The business of aggregating individual bricks with individual file systems is a fine way to build reliable systems,” he said.

But what about another point Newman raised–that general-purpose Linux file systems often have difficulty with large numbers of file requests? Just a little while ago I was speaking with a user who was looking for streaming performance from a file system, and an overload of small random requests brought an XFS system down. “Well, someone worried about small files does have a problem,” Harris said, though it’s a tangential point to the original point Newman raised. “But everybody has this problem–there is no single instance file system that does everything for everybody.” He added, “this may be an earea where Flash drives have a particular impact going forward.”

May 29, 2008  4:39 PM

EMC shifts Cisco WAAS to services group

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

Despite rumors that pop up from time to time, Cisco has not taken its Wide Area Application Services (WAAS) product off the market. Cisco is still developing and selling WAAS, and its storage partner EMC sells it too, although there has been a change there. Instead of selling WAAS directly, EMC now sells it through its professional services group.

So, essentially, WAAS becomes another tool in the toolbelt for EMC’s Professional Services. The services folks decide when it fits in an environment and they do the deployment, rather than the customer themselves. According to an EMC spokesman in an email to

In April, EMC modified its go-to-market approach around Cisco WAAS based on customer feedback and to provide differentiated value. EMC was originally selling it without professional services. We received feedback from customers asking us to pair it with professional services so that they could take full advantage of the technology — which they realized had a lot of potential for transforming application delivery, infrastructure through consolidation, etc. Before customers made the investment, they often asked for an assesment to see if it was right for them. Once they purchased it, customers were asking for help implementing it. As a result, we developed a specialized practice within EMC Global Services called EMC Data Center Networking Practice — which is where Cisco WAAS is now offered — and it includes comprehensive professional services.

That same spokesman stressed that it was just a shift in the delivery of the product, rather than a reflection on the product itself. But why are customers ask for help deploying and understanding the use cases for it? There are plenty of products, especially in EMC’s portfolio, that require professional services engagements to get the best results, and companies shift the delivery of products all the time in an effort to boost sales. Still, a shift in delivery method mid-stream, which is directly attributed to customers’ difficulties with understanding and deploying a product, doesn’t sound like good news for WAAS.

May 28, 2008  11:47 AM

Storage experts pan report on tape archiving TCO

Beth Pariseau Beth Pariseau Profile: Beth Pariseau

The disk vs. tape debate that has been going on for years is heating up again, given technologies like data deduplication that are bringing disk costs into line with tape.

Or, at least, so some people believe.

The Clipper Group released a report today sponsored by the LTO Program which compared five-year total cost of ownership (TCO) for data in tiered disk-to-disk-to-tape versus disk-to-disk-to-disk configurations. The conclusion?

“After factoring in acquisition costs of equipment and media, as well as electricity and data center floor space, Clipper found that the total cost of SATA disk archiving solutions were up to 23 times more expensive than tape solutions for archiving. When calculating energy costs for the competing approaches, the costs for disk were up to 290 times that of tape.”

Let’s see. . .sponsored by the LTO trade group. . .conclusion is that tape is superior to disk. In Boston, we would say, “SHOCKA.”

This didn’t get by “Mr. Backup,” Curtis Preston, either, who gave the whitepaper a thorough fisking on his blog today. His point-by-point criticism should be read in its entirety, but he seems primarily outraged by the omission of data deduplication and compression from the equation on the disk side.

How can you release a white paper today that talks about the relative TCO of disk and tape, and not talk about deduplication?  Here’s the really hilarious part: one of the assumptions that the paper makes is both disk and tape solutions will have the first 13 weeks on disk, and the TCO analysis only looks at the additional disk and/or tape needed for long term backup storage.  If you do that AND you include deduplication, dedupe has a major advantage, as the additional storage needed to store the quarterly fulls will be barely incremental.  The only additional storage each quarterly full backup will require is the amount needed to store the unique new blocks in that backup.  So, instead of needing enough disk for 20 full backups, we’ll probably need about 2-20% of that, depending on how much new data is in each full.

TCO also can’t be done so generally, as pricing is all over the board.  I’d say there’s a 1000% difference from the least to the most expensive systems I look at.  That’s why you have to compare the cost of system A to system B to system C, not use numbers like “disk cost $10/GB.” 

Jon Toigo isn’t exactly impressed, either:

Perhaps the LTO guys thought we needed some handy stats to reference.  I guess the tape industry will be all over this one and referencing the report to bolster their white papers and other leave behinds just as the replace-disk-with-tape have been leveraging the counter white papers from Gartner and Forrester that give stats on tape failures that are bought and paid for by their sponsors.

Neither Preston nor Toigo disagrees with the conclusion that tape has a lower TCO than disk. But for Preston, it’s a matter of how much. “Tape is still winning — by a much smaller margin than it used to — but it’s not 23x or 250x cheaper,” he writes.

For Toigo, the study doesn’t overlook what he sees as a bigger issue when it comes to tape adoption:

The problem with tape is that it has become the whipping boy in many IT shops.  Mostly, that’s because it is used incorrectly – LTO should not be applied when 24 X 7 duty cycles are required, for example…Sanity is needed in this discussion… 

Even when analysts agree in general, they argue.

Forgot Password

No problem! Submit your e-mail address below. We'll send you an e-mail containing your password.

Your password has been sent to: