Storage Soup


June 27, 2014  1:09 PM

Dell-Nutanix deal can be hyper-contagious

Dave Raffo Dave Raffo Profile: Dave Raffo
Dell, Nutanix, SimpliVity, Storage

Dell and Nutanix took a lot of people by surprise this week when they disclosed an OEM deal that will result in Dell selling Nutanix hyper-converged software on PowerEdge servers. If this plays out like other emerging technologies have, you can expect more established storage vendors to partner with – or acquire – hyper-converged technology that bundles storage, networking, and virtual servers.

There was an erroneous report a few weeks ago that Hewlett-Packard was set to acquire SimpliVity, a rival of Nutanix. That was likely the result of rumblings in the industry about large vendors looking to add this type of technology. It’s no secret that EMC is working on its own hyper-converged system – Project Mystic – and most of the major server hardware vendors will have pre-packaged hyper-converged bundles running VMware’s Virtual SAN (VSAN) software.

‘This speaks volumes about the validation of our technology and our product,” Nutanix VP of marketing Howard Ting said of the Dell deal.art

When a hot storage technology starts gaining momentum, the major vendors always look to jump on the bandwagon. They sometimes develop their own, but it is easier to buy it from a startup that has spent years and hundreds of millions of dollars developing it. Big vendors looking to take a shortcut into hyper-convergence have a few options but not too many. Software-only hyper-convergence startup Maxta would make a good target. SimpliVity CEO Doron Kempel says his company is not for sale, but can be an OEM partner despite requiring a PCIe card to carry out its inline deduplication without impacting performance.

“We have the technological capability to do that,” he said of an OEM deal. “We can give a large server vendor functionality of our stack, which includes the software and card.”

Kempel insists that he has never discussed a sale, OEM deal or any other partnership with HP, however.

“I’ve never spoken with HP and never had any M&A discussions,” he said. “We have no commercial relationships with them. We’re not being acquired by anyone, not even for $2 billion.”

Taneja Group consulting analyst Arun Taneja said the Dell-Nutanix deal “puts hyper-convergence in hyper-drive” and predicts more deals. “Hyper-convergence is a totally tomorrow type of product,” he said. “It’s taking all these technologies that at one time were best of breed but now you can integrate them.”

Sanbolic CEO Momchil “Memo” Michailov predicted the Dell-Nutanix deal will spark a feeding frenzy of partnerships involving other types of server-attached storage software as well.

“This is classic Dell: take a high-value product, smash it into the channel and roll out a commodity offering,” Michailov wrote in an e-mail statement. “As an entry-level platform, this might be the right move for Nutanix and lead to a successful IPO down the line. From a broader industry perspective, the move is the first in what I expect to be a line of strategic plays to move billions of dollars in market value from storage back to the server layer. HP, IBM/Lenovo and others won’t be far behind, but the real enterprise market opportunity relies on software-defined storage players that can offer tier-one, scale-out capabilities.”

June 26, 2014  1:58 PM

HP brings Helion services to the cloud

Sonia Lelii Sonia Lelii Profile: Sonia Lelii
Storage

HP has renamed an extended version of its managed storage and backup services as Helion.

Formerly called Storage Management Services and Backup Restore services, Helion now is positioned as a managed cloud offering that uses HP 3PAR StoreServ and HP StoreOnce Backup systems as the underlying hardware. The offering can be customized for private, virtual private and public cloud. It leverages block, file and object storage along with backup services.

“We collapsed a set of servers and provisioned gigabytes of storage into a single offering,” said Mike Meeker, HP’s global offering manager for HP enterprise services.”The focus is extending the existing offering. We have a core set of services and structures. With that structure comes the ability of multiple features.”

Meeker said the offering can be customized. For instance, multiple-tiers in a SAN or block array can be configured for high performance applications that require low latency. The service also has tiers of files services that can be used for numerous use case based on the type of application and databases across private, virtual private and public clouds.

Its customized configurations leverage block, file and object storage as well as backup services powered by HP 3PAR StorServ and HP StoreOnce Backup systems. The different tiers and features are front-ended with a new portal.

The block-based service targets application and databases that need high-performance via Fiber Channel communications. The file storage is for NAS to provide file-level storage accessed via the Internet protocol. The object storage provides backup applications the ability to add, delete and modify files via REST or HTTP protocols. Server backup creates copies of data for services in the data center and remote offices.

“They now have the flexibility to pick and choose different tiers, at different price points and different application needs,” Meeker said.


June 25, 2014  4:24 PM

Hitachi Data Systems tiers up its cloud platform

Dave Raffo Dave Raffo Profile: Dave Raffo
Cloud file-sharing, Cloud storage, Hitachi Data Systems, Public Cloud, Storage

Hitachi Data Systems (HDS) has enhanced its cloud storage platform for people who want to use the public cloud as a tier.

The vendor made changes to three cloud products – the Hitachi Content Platform (HCP), HCPAnywhere and Hitachi Data Ingestor (HDI). HCP is object-based software that serves as HDS’s main cloud platform and can run on any of its hardware products. HCP Anywhere is the file sync and share software introduced a year ago that lets mobile devices access HCP. HDI is what HDS calls its “cloud on-ramp” that lets companies connect to the cloud from remote offices. All three can run as virtual appliances.

With HCP 7 launched this week HDS added cloud adaptive tiering, which lets customers move data on and off Amazon, Google and Microsoft public clouds. The policy-based tiering controls what data is kept on-premise and what data goes to a public cloud. New synchronization capabilities let customers sync data across active sites. The new tiering and synch features let companies store data local and in the cloud and access them through mobile devices.

“We’ve become a cloud broker,” HDS CTO Peter Sjoberg said. “Anything brought into HCP can be sent out to any cloud. The information flows through HCP, but the metadata is retained in the data center. The information can go to your cloud of choice.”

Because HDI caches data that is used locally, there is less of a need for WAN optimization for remote sites. HDS also says the new features alleviate the need to do distributed backups across site. “You don’t have to pop a tape in,” said Tanya Loughlin,HDS director of file, object and cloud product marketing at HDS.

Making it easier for end users to share information could solve a BYOD headache caused by people using their own devices in their own way, or as Sjoberg put it, “going rogue, and going around IT.”


June 23, 2014  1:46 PM

Cloud storage does not eliminate IT’s responsibility for data

Randy Kerns Randy Kerns Profile: Randy Kerns
Storage

I have been told by a storage administrator that if he moved data to the cloud he was no longer responsible for it.  He made the wash-my-hands of it sign during that conversation to illustrate it was “not my problem.” That’s because of guarantees offered by the cloud service provider storing the information.  I did not get any of the details about those guarantees but I would also question how sound they were and what recourse the company would have if anything did get lost.

An event happened this past week that illustrates this problem and set off alarm bells for all storage administrators and IT management.  U.K.-based code hosting company Code Spaces lost every client’s data and ceased operation.  The data loss was due to a malicious hacker that deleted the data and was so proficient at it that Code Spaces’ protection mechanisms were incapable of protecting or recovering data.  Russ Fellows wrote a  blog with more detail on the Evaluator Group web site.  More can be found at SearchAWS.com.

This was a major fail for Code Spaces and a major loss of valuable development source code that was stored there by many companies. And there were guarantees about how the data was protected.  A conversation with an IT guy who decided to put data on Code Spaces would be completely different now.  Would a company executive believe that the IT guy was no longer responsible when the data was moved there?  There is no realistic defense for the IT guy here.

This means the responsibility for protecting information and ensuring its availability for use remains with IT, specifically the storage team, irrespective of where it is physically located.  Protecting data from disaster or some type of alteration or destruction is one of the earliest and most basic jobs in IT.  Making the data available to the point of business continuity is a responsibility. The consideration that IT would be absolved of those responsibilities by moving to a cloud provider is wishful thinking.

Moving the data to the cloud may have economic benefit.  But it still requires the operational effort and expense of ensuring the data is protected and available. The protection and availability must be proven with periodic exercising of recovery and availability switchover.  Without that, the liability is there but without evidence that responsible actions have been taken.

In advising our IT clients, this is a great example to use.  From this point on, when someone says they are moving data to the cloud and no longer have responsibility, I’ll just ask if they have it in writing that they will not be held liable. We certainly have an example now.

(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).


June 20, 2014  3:38 PM

One-click cloud DR? Actifio says, ‘copy that’

Dave Raffo Dave Raffo Profile: Dave Raffo
Storage

More than three years after coming out with its copy data storage software, Actifio is expanding the use cases for its data protection technology. Actifio launched Actifio Sky platform in May as a remote office companion to its Copy Data Storage (CDS) core data center product, and this week built on those platforms with a Resiliency Director service for disaster recovery.

The goal for Resiliency Director is to provide one-click failover of heavily virtualized environments, using the cloud or a co-location as a DR site.

Resiliency Director consists of virtual appliance data collectors on the primary site that discovers VMware virtual machines and vApps, and CDS systems at the primary and DR sites. The CDS software performs deduplication and asynchronous replication of VMs, keeping them in a ready state for DR. When there is a failure (or DR test), Resiliency Director software at the DR site orchestrates storage, compute, network and data, and enables full application recovery.

Actifio claims it can enable application recovery faster than anybody else because of its granular understanding of VMs and vApp reduces the size of the data sets and it only has to rehydrate changed deduped blocks when making restores. Also, a read cache in CDS reduces latency and improves IOPS in the recovery stack.

Customers can organize VMs and vApps in application groups and prioritize the order in which they are recovered.

“Our customers said they wanted DR to operate in a much more granular VM level,” Actifio VP of product management Brian Reagan said. “And they want to orchestrate recoveries without relying on array-based replication that adds cost.”

Reagan said Actifio plans to add support for more hypervisors – Microsoft Hyper-V is likely the next – and physical machines to Resiliency Director.

Sungard Availability Services is already using Actifio Resiliency Director for its Recover2Cloud service. This expands the vendors’ partnership, which already includes Sungard’s Managed vaulting Backup for Actifio service. Sungard Availability VP of product management Souvik Choudhury said Recover2Cloud will be available as a standalone service and integrated with other Sungard services.


June 20, 2014  1:40 PM

Veeam offers secure SSL connection for offsite backup

Sonia Lelii Sonia Lelii Profile: Sonia Lelii
Storage

Veeam Software is offering its service provider partners a way to give more customers an integrated and secure way to move backups to an offsite repository.

The company recently introduced VeeamCloudConnect as part of its Veeam Availability Suite v8. The service requires a single server and takes less than 10 minutes to implement, while providing all the infrastructure management needed for an offsite repository service.

Doug Hazelman, Veeam’s vice president of product strategy, said VeeamCloudConnect offers the service to provide its partners a Secure Sockets Layer (SSL) connection for secure communications on the Internet for data transfers.

“Normally the service provider has to set up a  VPN and a separate repository for customers,” Hazelman said. “(With this) they can make one repository available and create multiple tenants in that repository, which is at the cloud provider’s location.”

Hazelman said they are not disclosing pricing, but Veeam’s marketing data says they are offering pricing that is tailored to meet service providers’ specific needs, so instead of an up-front perpetual license Veeam will offer service providers monthly per virtual machine licensing.

The service providers will most likely charge monthly on capacity and not based on the number of virtual machines. Veeam customers can search for the nearest service providers through an integrated Web portal in the Veeam Availablity Suite v8.


June 20, 2014  8:46 AM

Carbonite goes after SMBs with cloud backup appliance

Dave Raffo Dave Raffo Profile: Dave Raffo
backup appliance, Carbonite, Cloud Backup, Storage

Carbonite, a cloud backup pioneer going back to the days when it was something done only by consumers, is moving deeper into the SMB market.  Carbonite is also changing its delivery method from software that customers download off its web site to appliances sold by channel partners.

Carbonite launched its first appliance this week. The Carbonite Appliance HT10 includes Carbonite software and 1 TB of local storage with the ability to move 500 GB into the Amazon cloud.

Dave Maffei, Carbonite’s VP of Global Channels, said the appliance is designed for SMBs with up to 500 employees. Customers will pay a monthly fee for the appliance and Amazon capacity. He said channel partners will set the pricing, but Carbonite CEO David Friend said the appliance would cost $99 per month during the vendor’s latest earnings call in April.

Maffei said the software on the appliance is different than what Carbonite sells to consumer. The appliance software likely includes technology that Carbonite acquired when it bought SMB cloud vendor Zmanda in 2012. HT10 connects to servers via Ethernet, takes local bare metal images of the servers and replicates them to the cloud.

The software for businesses must be more reliable than that for consumers. “A couple of days without data is potentially the end of the road for a business,” Maffei said. “It’s different than losing pictures of the kids.”

The Carbonite sells its Zmanda software as Carbonite Server for SMBs, but the majority of its revenue still comes from consumers.  In the first quarter of this year, consumers accounted for $23.3 million in revenue compared to $9.2 million from SMBs. The appliance should close that gap.

Carbonite’s appliance and new SMB focus also shows how cloud backup is gaining acceptance. Carbonite, Mozy and the other early cloud backup vendors focused on consumers at the start because businesses wouldn’t think of trusting data protection to the cloud. Now Mozy is part of EMC, and even the largest corporations are backing up to the cloud. That means you can expect larger appliances from Carbonite in the not-so-distant future.


June 19, 2014  12:52 PM

Asigra can commoditize backup hardware for cloud providers, but what about users?

Dave Raffo Dave Raffo Profile: Dave Raffo
Asigra, Storage

Asigra unveiled interesting technology this week with its new software stack that can reduce costs for its cloud backup service provider partners.

Calling it software-defined data protection, Asigra is offering a free download that runs on commodity hardware and replaces the need for expensive disk arrays as backup targets. Asigra’s open-source stack includes FreeBSD software and ZFS. ZFS serves as a mount point on the hardware. The providers still need to install Asigra Cloud Backup on customers’ site.

The software adds value for service providers, but doesn’t change things for users. It could, though. Asigra executive vice president Eran Farajun said the company is talking to commodity hardware vendors about bundling the software on low-cost appliances.

“It’s absolutely possible,” he said. “We’re talking with a couple of OEM vendors that sell commoditized hardware who are looking for applications like this. Software sells hardware, not the other way around.”

In keeping with Asgira’s business model, these appliances would be sold to service providers. But that wouldn’t stop other backup software vendors from doing something similar with their software and bringing it market for end users on appliances that cost far less than the specialized targets already on the market.


June 19, 2014  12:16 PM

Cloud-to-cloud backup picks up steam with acquisition, new Salesforce.com features

Dave Raffo Dave Raffo Profile: Dave Raffo
Salesforce.com', Storage

The cloud-to-cloud backup universe expanded over the past week when cloud file sync vendor eFolder acquired Cloudfinder and Spanning Cloud Apps added features for its Salesforce.com backup.

eFoder acquired Sweden-based Cloudfinder last week for an undisclosed price. Cloudfinder backs up data stored in Microsoft Office 365, Google Apps and Salesforce. Cloudfinder joins an eFolder lineup of services that includes file synch and share, DR/business continuity, and backup.

eFolder operates its cloud out of data centers in Atlanta, Salt Lake City and Colonna, British Columbia. Until now, CloudFinder used the Amazon cloud to store customer data. Ted Hulsy, efolder VP of marketing, said efolder will migrate CloudFinder customer data from Amazon to the eFolder cloud, in similar fashion as it did after acquiring sync and share vendor Anchor and moved its customers’ data from the SoftLayer cloud.

Unlike Cloudfinder, eFolder does not sell directly to users. It will also turn Cloudfinder into a 100-percent channel play, according to eFolder VP of marketing Ted Hulsy.

Hulsy said eFolder will continue to sell its services separately, although its service provider customers will eventually be able to manage all from the eFolder partner portal.

Cloudfinder launched in early 2013 with backup for Office 365, which remains its most popular product.

“We’re all studying how the landscape is shifting to a cloud-centric future with the proliferation of mobile devices, and we see the writing on the wall that traditional offerings have to transform [to the cloud],” Hulsy said. “Cloudfinder was an ideal company to acquire because of that.”

While Cloudfinder is smaller than cloud-to-cloud backup vendors Backupify and Spanning, Hulsy said the service will benefit from eFolder’s bulk and channel. He said eFolder has more than 120 employees and 2,200 partners worldwide, and has raised $26 million in funding.

“Our product vision is to build out really robust infrastructure of cloud services,” he said. “We’re talking about hundreds of applications over time.”

Spanning this week launched Spanning Backup for Salesforce1 Mobile, allowing Salesforce admins to monitor backups and restores remotely. The Spanning application runs on the Salesforce 1 app that is available on the Apple iOS and Google Android operating systems.

Last week Spanning added the capability for  Spanning Backup for Salesforce customers to restore data objects directly from the Salesforce user interface. This lets users restore their own records without IT intervention. They can go back to previous versions of Salesforce data objects such as Accounts, Opportunities and Contacts.  This allows them to restore similarly to how they do it in Google Apps , with each user responsible for his or her own data.

“Companies want to put data recovery into the hands of people who own their data, and they want granular point in time recovery,” Spanning CEO Jeff Erramouspe said. “Salesforce is one big database carved up into individual pieces, where in Google Apps each user owns his data – email, Drive documents, calendar – on an individual basis.”


June 16, 2014  8:50 AM

SanDisk gobbles up Fusion-io for PCIe flash; fate of arrays not clear

Dave Raffo Dave Raffo Profile: Dave Raffo
Fusion-io, PCIe flash, SanDisk, Storage

SanDisk expanded its enterprise flash product line today by acquiring PCIe flash leader Fusion-io for $1.1 billion. However, SanDisk provided few specifics about its plans for its newly acquired products.

SanDisk is known primarily as a PC solid state drive (SSD) vendor although its enterprise revenue shot up in 2013 and this is its second acquisition for server-side enterprise flash in 10 months.

Fusion-io was an early flash success story, turning its dominance in the PCIe market to an initial public offering in 2011 but fell on hard times last year.

“This will position us for a leadership position in enterprise storage,” SanDisk CEO Sanjay Mehrotra said on a conference call this morning to discuss the deal.

SanDisk last August acquired SSD and flash controller company Smart Storage for around $307 million. Smart Storage sells ULLtraDIMM flash memory cards that sit in the server’s motherboard and have lower latency than PCIe cards. Mehrotra said the Fusion-io technology complements SmartS torage’s UltraDIMM product, which is seen as competitive with PCIe flash. He claimed there is a market for both types of products as well as SAS and SATA SSDs.

Fusion-io expanded its product line in April 2013 by acquiring storage array startup NexGen Storage for $114 million. Fusion-io sells the NexGen hybrid flash arrays under the ioControl brand. Those arrays use Fusion-io flash instead of solid state drives (SSDs).

However, Mehrotra would not commit to the ioControl product. He did not mention the arrays when discussing Fusion-io technology during the call. When asked twice if SanDisk would continue the hybrid SAN array product, Mehrotra said he would talk about that about it after the deal closes. He expects that to happen in Setember.

Mehrotra also declined to say how the deal would affect SanDisk’s current early development of PCIe flash products.

SanDisk moved into the enterprise SSD market when it acquired Pliant Technology for $327 million in 2011.

The combined revenues from SanDisk and Fusion-io would have made SanDisk the No. 2 overall SSD vendor last year behind Samsung, according to Gartner’s SSD and solid-state array market share report published last week. SanDisk and Fusion-io combined for more than $1.6 billion in revenue in 2013.

According to Gartner, SanDisk jumped from fifth in SSD shipments in 2012 to third in 2013, passing Fusion-io and Micron. SanDisk’s SSD revenue increased 263 percent over the year (compared to 53 percent growth for the entire market), from $355 million to $1.3 billion and its market share grew from five percent to 11.7 percent. It grew more than any other SSD vendor in 2013.

While most of SanDisk’s 2013 revenue came from PC SSDs, it did jump from eighth to fourth in revenue from enterprise SSDs. SanDisk enterprise revenue increased 181 percent (compared to 47 percent for the entire market) to $375 million and its market share went from 4.4 percent in 2012 to 8.5 percent last year.

Meanwhile, Fusion-io’s revenues declined sharply in 2013 as its early large customers Facebook and Apple greatly decreased their spending while competition grew after storage giant EMC and others entered the PCIe market. Fusion-io switched CEOs in May 2013, pushing out David Flynn and replacing him with Shane Robison.

Robison is out now, too. Mehrotra said Fusion-io president/COO Lance Smith will become senior VP and president at SanDisk, leading the Fusion-io development and marketing teams.

Fusion-io fell from fourth in 2012 in total SSD revenue to eighth in 2013. Fusion-io’s revenue declined 18 percent to $339 million last year. It stood eighth overall and fifth in enterprise SSD revenue. Its market share of 7.7 percent of the enterprise market was down from 13.6 percent.


Forgot Password

No problem! Submit your e-mail address below. We'll send you an e-mail containing your password.

Your password has been sent to: