As certification announcements go, this one is more interesting, I think, than most others, if only because it harkens back to one of the most interesting product announcements/demonstrations I saw last year.
At last year’s VMWorld in San Francisco, Cisco and VMware demonstrated distance VMotion, a technology that will be key to VMware’s vision of data center federation and fluidity between public and private clouds. However, distance VMotion as of that conference had several limitations, the most significant of which from a storage perspective is the need to migrate potentially large volumes of data over distance very quickly in order to support VMotion between data centers.
VMware said last year it will support customers if they deploy distance VMotion using the Cisco network, but its support statement included extensive fine print, including a minimum network bandwidth of 622 Mbps, or an OC12 connection.
Partners were scrambling at that time to step in to solve the data migration problem (including EMC, which was developing “active-active” storage to support distance VMotion), and some of the exhibitors on the show floor, including NetEx and F5 Networks, claimed to be able to solve the problem today. At the time, however, no WAN optimization products were certified for distance VMotion with VMware.
Today, NetEx announced certification of its HyperIP software as VMware Ready, which according to a press release means “HyperIP integrates consistently with VMware technology and is ready for deployment in customer environments.” The press release doesn’t mention distance VMotion specifically, but a NetEx spokesperson said a large oil and gas company has deployed the software for distance VMotion. That customer is not open to taking questions from press, the spokesperson said.
According to IDC’s 2010 Digital Universe report, digital data grew 62% last year as 800,000 PB were added. IDC says 1.2 million PB (1.2 zettabytes) will be added this year, and that will increase to 35 ZB in 10 years.
While those numbers may look staggering on a page, they probably don’t shock anybody charged with managing data storage. The real shocking – and frightening – number is that IDC says the amount of IT staff to manage all this data will only grow by a factor of 1.4 by 2020. If IDC is correct, than the dreaded “do more with less” mantra will become a long-term way of life.
So how will this all change the way we manage data? Chuck Hollis, global marketing CTO of EMC – which sponsored the IDC study – says the data growth will push a lot more of it to the cloud this year. Hollis says the IT staffs at large enterprises that he talks to are ready to set up private clouds to manage data.
“For tech guys, this is the year of putting your cloud strategy together,” Hollis says. “We’re way beyond the ‘What is the cloud?’ discussion, and it’s a very mature discussion with the IT guys I talk to.
“The larger enterprises say, ‘We’re big, we can do this ourselves. We can build a private cloud behind the firewall and get comfortable with it.’ They’re saying, ‘We pay the same price for this stuff – the processors, server, storage – there’s no reason I can’t do what Amazon does.’”
Hollis says as long as organizations feel they can control their data in the cloud, they’re willing to move it there.
“The cloud works when enterprise guys can be in control,” he said. “Ask them to give up control, and it’s not that attractive a proposition for them. You can’t outsource responsibility and accountability. In financial services, a trillion dollars a day floats around the global economy over the cloud. Most days we’re OK with that. Clouds, schmouds, it doesn’t matter as long as enterprise guys feel they’re in control.”
Other emerging methods of managing growth aren’t quite as mature, Hollis says. That includes data deduplication for primary data. While EMC is now the leader in backup dedupe, Hollis says the success of primary deduplication “has a lot to do with processors being fast enough to do it without impacting performance. If you have a SAP application with 10,000 demanding users, maybe it [deduplication]’s a false savings. The concern is, at what cost? The technology gets better year over year, but some are of the opinion this is just a temporary fix, you’re just buying yourself some time. A lot of information is not compressible, like JPEGs. You can’t compress something that’s already compressed.”
Flash solid state storage is another area where EMC has been out front, but it’s another technology where the greatest benefits are still down the road. “If you take what processors have done in the last 10 years as far as density, price and performance, then start with flash in 2010 and forecast it out in 10 years, it could actually get cheaper than disk,” Hollis said. “That would be an interesting world.”
FalconStor Software officially reported revenue Thursday, confirming what if first said in a preliminary report April 19 – it had a lousy quarter.
FalconStor’s $17.1 million in revenue was down from $21 million a year ago, and it lost $5.5 million compared to a loss of $900,000 in the same quarter last year.
FalconStor’s problem is it sells mainly through OEM partners, and its largest partnerships were disrupted last year. Its biggest OEM partner, EMC, bought Data Domain and now sells more Data Domain data deduplication boxes and fewer of its VTLs that use FalconStor software. So FalconStor revenue from EMC declined $300,000.
Sun is another partner, but Sun was in the process of getting acquired by Oracle for most of 2009 and its been unclear which of its products would survive the acquisition. FalconStor revenue from Sun dropped $1.1 million last quarter. FalconStor also took a hit when Hewlett-Packard acquired FalconStor partner 3Com, although FalconStor executives say they expect a rebound now that 3Com is integrated into HP. Another FalconStor partner, Copan, effectively went out of business last year before SGI acquired its assets and resurrected its archiving product.
FalconStor says it will cut spending and has imposed a hiring freeze until it becomes profitable again. More importantly, it is finding new OEM partners. As VP of business development Bernie Wu put it, “We had an unusually high level of disruption with our OEM partners last year, and we’re forming a new foundation of partnerships.”
FalconStor executives say they expect to launch two new Tier 1 OEM deals late this year. One will be for a cloud services offering. They didn’t such much about the other, but one possibility is a deal with Hitachi Data Systems for FalconStor’s File-interface Deduplication System (FDS) software.
HDS so far has a piecemeal approach to backup data deduplication. It resells IBM Diligent ProtecTier, but doesn’t push a product owned by its rival IBM. HDS salespeople have financial incentive to sell the new Sepaton VTLs built on HDS disk, but there’s no formal reseller deal. HDS OEMs CommVault’s Simpana that includes deduplication and certifies FalconStor’s dedupe, but lacks one main dedupe product.
During the earnings call Thursday, Wu said FalconStor had a “significant pipeline” with HDS for the FalconStor software it resells and “we expect that partnership to deepen.”
Ever since Oracle said it would end its OEM deal with Hitachi Data Systems for its enterprise storage systems, people in the industry have wondered if Oracle would also sever its midrange storage OEM deal with LSI.
Oracle executives say they killed their HDS deal because they don’t make enough money selling other vendors’ storage, which doesn’t bode well for LSI.
But LSI CEO Abhi Talwalkar says he’s optimistic about continuing with Oracle. During LSI’s earnings call Wednesday evening, Talwaker even talked about expanding the partnership.
“We are pleased with our competitive position at Oracle,” he said. “Oracle recently posted a pdf on [its] website to address the partnership with Hitachi Data Systems. We believe there have been positive developments for LSI, including the termination of the HDS relationship. This will give LSI more room to grow, and Oracle also mentions support for technology partners associated with the [Sun StorageTek] 6780 system and 6000 series. which is all leveraging LSI system technology.”
Oracle hasn’t said anything publicly either way about LSI. During Oracle’s earnings call last month, CEO Larry Ellison says OEM relationships with HDS and Symantec Veritas backup software have ended but did not mention LSI. In explaining the HDS and Symantec decisions he said “we add no value so we are out of that business.” He did mention expanding the SunStorageTek 7000 midrange storage as well as high performance and high end servers platforms, but not the 6000. “Where Sun was specifically a distributor of someone else’s intellectual property and lost money doing it, we are out of that business,” Ellison said.
But what if Oracle/Sun is making money on LSI’s storage? LSI reported its second straight strong quarter ifor storage system sales with revenue of $221 million, up 40% from a year ago. Besides Oracle, LSI’s OEM deals include midrange storage systems for IBM and entry-level enclosures for Dell and other smaller vendors.
Talwaker also said LSI will launch a new 6 Gbps SAS entry level platform with up to four times performance and twice the capacity of its current platform in the second half of this year.
Earlier this week, we ran a story about email hosting provider Intermedia attributing a recent outage to a failure in its EMC SAN. After the story ran, we received feedback from Bob Adams, a storage systems engineer at a leading Boston teaching hospital, on the case:
“I can’t see how Intermedia can truly blame this on EMC,” Adams wrote in an email.
First of all, the EMC SAN referred to here is clearly an EMC CLARiiON based on the information provided. The fact that one of the storage processor’s had a failure, probably a bugcheck panic (like a windows BSOD…CX’s run Windows OS on the SP’s) due to a bug in the firmware aka FLARE code is a case that their SAN Admin hadn’t been patching/updating the FLARE code on a regular basis as he/she should be doing.
Then with the failure and having to run on one storage processors is something the CLARiiON is designed to be able to do for fault tolerance as well as load balancing, again the SAN admin was at fault for this CLARiiON was clearly over utilized. The utilization on the storage processors has to be within a CPU percentage range so that if an SP had a failure the second SP could handle its own load plus the load of the other. Meaning if the utilization of say SPA was 75% and the utilization of SPB was 75%, there is no way if SPA failed SPB will be able to handle the load. Which sounds what happened here. I see this as more of Intermedia’s own fault over EMC.
What do you think? Comments operators are standing by…
Unitrends Inc. has put its own spin on data reduction for small and midsized businesses (SMBs) that use its backup appliances.
Previously, Unitrends has offered file-level compression and post-process subfile data deduplication with its products, but said the CPU overhead of doing subfile level deduplication on its customers’ relatively small data sets resulted in the need for beefier processors and appliance hardware. This in turn might be more expensive for some small customers than just buying more disk, according to Unitrends COO Mark Campbell.
Unitrends today announced what it calls Adaptive Deduplication, and Campbell says the goal is to offer users the best storage utilization possible without compromising performance. Adaptive Deduplication adds a light content-aware algorithm that evaluates the type of data (structured or unstructured) as well as its size as it comes into the system, and determines how the best data reduction ratio can be achieved. All files are compressed as they come into the system, but only larger data objects will be pulled apart for sub-file dedupe later.
“Typically structured data is better served by the compression ratio — files almost always dedupe pretty quickly,” he said. Now, if a user is making small incremental changes to a database, the system won’t have to pull apart every small block to look for additional data reduction beyond compression — it can just compress the data and move on.
While the high performance overhead of doing data deduplication has been a major issue with the technology since its inception, users at midsized and larger companies have been willing to pay the price for processors in order to contain unmanageable backup capacity growth.
But Campbell brings up an interesting challenge to dedupe-as-panacea: Unitrends customers are often in small shops that require as little as seven days data retention, and “they don’t get great ratios with traditional block-level deduplication. When disk drives are so cheap, it’s not necessarily a no-brainer to purchase next generation hardware to push subfile dedupe.”
The compression and file-level dedupe will be included with the software that comes on all Unitrends appliances, and current customers will be able to download it beginning next month. For customers still looking for subfile dedupe, Unitrends will also come out with a new appliance heavier on processors than capacity later this year, which will make subfile deduplication more likely under Adaptive Deduplication. “It sounds funky and weird and why not just put [a new appliance] out there, but it’s a price-performance issue,” Campbell said.
Virtual Instruments, a spinoff of network monitoring company Finisar, rolled out a new hardware configuration for its Traffic Access Port (TAP) devices, which watch bi-directional Fibre Channel traffic on SAN switch links to diagnose problems.
TAPs are an option with Virtual Instruments’ VirtualWisdom and NetWisdom FC SAN monitoring products, which also involve software components and can be deployed with or without adding TAPs to the SAN. Previously, according to Virtual Instruments VP of marketing Len Rosenthal, users wishing to use TAPs would have to attach two fiber optic splitters to each connection they wanted to monitor. That was time consuming and disruptive if it involved critical connections. It could also result in more “light loss” from the fiber optic connections depending on how many TAPs were in place.
Today, Virtual Instruments introduced a new patch panel hardware form factor for TAPs, which allows multiple fiber optic connections to be connected to one 10-Gigabit panel that can be dropped into a user’s existing core switches. This makes connecting to the TAP once it’s in place simpler and uses patch cables, which Rosenthal said users should also be familiar with.
However, even with the new patch panel there will still be some disruption when retrofitting already-running SANs with TAPs. Rosenthal said Virtual Instruments advises users to TAP new SAN connections as a part of installation. “Life is so much easier if the TAP is in there to begin with,” he said.
NAS vendor Isilon made it two straight profitable quarters when it beat expectations for last quarter to report 46% year-over-year revenue growth.
Isilon’s revenue of $39.3 million was even up 5% over the previous quarter, and it’s rare for storage companies to increase revenue from the fourth quarter of one year to the first quarter of the next. Isilon had $1.1 million of net income, up from $100,000 in the previous quarter. That compares to a $10.4 million loss in the first quarter of 2009. Isilon executives even raised their forecast for 2010 revenue from growth in the low-to-mid 20% range to an increase in the mid-30% range.
Isilon CEO Sujal Patel said the revenue growth came partly from pent-up demand of customers who has budget constraints in 2009, as well as better international sales and an increased channel focus. He also said Isilon is moving into more mainstream enterprise NAS accounts to go with its previous success in verticals such as media and entertainment.
Patel says Isilon is also benefiting from an industry focus on scale out NAS because its main NAS competitors don’t have clustered products yet.
“We have very little competition in the scale out space,” Patel said today on the Isilon earnings conference call. “There are lots of announcements out there from large companies, but most of the products we compete with are still two-controller products.”
After adding support for solid state drives (SSDs) and other hardware enhancements over the past year, Patel said Isilon’s product focus will be on software this year.
He says Isilon plans an operating system upgrade this year with and new software applications in 2010 and early 2011. He didn’t offer details, but Isilon could benefit from greater optimization with virtual servers and mainstream storage applications such as Oracle databases as well as native data deduplication.