While companies may be confident in their disaster recovery strategy, DR planning and testing still has a ways to go, according to results of a survey by data protection vendor Zetta.
Among 403 IT professionals, 88% said they were somewhat or very confident in their disaster recovery. But while 96% said they have some type of DR, 40% responded that their organization lacks a formally documented DR plan. For those with a plan, 40% said they test it once a year, while 28% rarely or never test it.
Similar to Zetta’s findings, a recent TechTarget survey found that companies are generally confident in their business disaster recovery plans. Also similarly, 65% test their DR plan just once a year or less, according to the TechTarget survey.
“Companies need to be more rigorous around how they develop their DR plans,” Zetta CEO Mike Grossman said. That’s especially important given that more than half of the companies in the survey experienced a downtime event in the last five years.
It’s not enough for business disaster recovery plans to ask, “What happens if a hurricane hits?” According to the survey, the most common type of downtime event for an organization in the last five years was a power outage. For those that had a downtime event in that time period, nearly 75% said their organization suffered a power outage. Only 20% experienced a natural disaster in the last five years. A hardware error was the second most common response, with 53%. Both a human error and a virus/malware attack registered at close to 35%.
When people think about disaster recovery, they often think of catastrophic events. “In reality, that’s not what causes the biggest impacts day to day,” Grossman said.
Grossman recommends testing business disaster recovery plans a minimum of once per quarter and preferably once per month.
“Unless you have a continuous process, including testing, you’re not really protected,” Grossman said.
But “not all testing is created equal,” he warned. The more rigorous and real-life, the better.
A lot of companies like to think they’re protected but they’re not, Grossman said.
According to the “State of Disaster Recovery” survey, 55% changed their DR strategy after a downtime event. It’s a positive that companies are paying more attention to the issue but a negative that they underestimate the risk, Grossman said.
For those that made changes following their last downtime event, 55% of organizations changed their DR approach, 55% added DR technology and 39% increased their DR investment. Almost 1 in 4 respondents said they increased DR testing.
As IT gets more complex, companies like Zetta need to make DR easier to manage, Grossman said. But how do you take the complexity out of something that’s complicated?
The cloud helps. Zetta is “cloud-first,” Grossman said, with backup in the cloud and failover to the cloud. And security, which has traditionally been a challenge in the cloud, is getting better.
Ninety percent of IT professionals who are using the cloud in their disaster recovery strategy said they are confident in their DR, according to the survey. Seventy-four percent of organizations using only on-premises DR said they are confident in their plans.
“It’s simpler,” Grossman said of the cloud. “It provides better protection.”
Companies spend a lot of money on disaster recovery solutions but that doesn’t translate into faster data recovery, according to a survey conducted by Quorum titled the “State of Disaster Recovery.”
The report, which surveyed 250 CIOs, CTO and IT vice presidents, found that 80% of the companies surveyed claimed it takes more than an hour to recover from a server failure and 26% said it takes more than two hours for data recovery. Only 19% said it took less than an hour to recover. Seventy-two percent consider the speed of backup and data recovery as “critical.”
“All those backup and disaster recovery (BDR) products aren’t making their recovery any faster,” the report claimed. “While speed is essential for continuity and security… a staggering 80 percent of respondents need more than an hour to recover from a server failure. And its gets worse: more than a quarter need more than two hours.”
Sixty-four percent use more than three different disaster recovery solutions, with 26% using more than five and less than 40% using between one and three different disaster recovery products.
Moreover, a majority of the respondents said they wanted a method to simplify the management of all the BDR products they are using. Ninety percent of the respondents want to consolidate their disaster recovery solutions into one dashboard.
The report shows that the movement to the cloud has grown. Seventy-five percent of the survey respondents are using cloud-based disaster recovery solutions, while 36% use a hybrid model mixing on-premises and cloud DR. Thirty-nine percent use on Disaster Recovery as a Service (DraaS).
Eighty-nine percent have plans for more cloud-based disaster recovery solutions, with five percent stating they have no further plans and six percent stating they “don’t know.”
Disaster recovery products are growing in importance as as concerns about security increase. Seventy-seven percent said they have used their disaster recovery solutions after a security threat event occurred. Fifty-three percent respondents are worried about security threats compared to concerns about hardware failure, backup disk corruption or a natural disaster.
“Natural disasters crashing in on a data center, an employee error or a hardware failure can all pose immense problems for an organization,” the report stated. “But a skilled and willful attack can cripple a brand for years and could cost a literal fortune. Ransomware attacks particularly depend on a team’s inability to recover quickly.”
Companies are diligent about testing the production level with their current disaster recovery products. Eighty-eight percent of the respondents said they can achieve production-level testing with their current DR.
How much data do you actually recover?
That’s a question that Asigra users answered in a data recovery report.
Featuring statistics gathered from nearly 1,100 organizations across eight sectors, between Jan. 1, 2014 and August 1, 2016, the backup and recovery software vendor’s report found that those users only recover about 5% of their data on average.
“People really don’t recover a lot of data,” said Eran Farajun, executive vice president at Asigra. “Ultimately they’re paying like they recover all their data.”
Farajun compared the situation to what many experience with cable bills – customers often pay for hundreds of stations but don’t watch all of them.
Broken up by industry, manufacturing and energy recovered the most, averaging about 6%, according to the data recovery report. Public sector and health care recovered the least, at about 2%.
Users picked file-level systems as the most common data type restored.
The most common reason cited for a data restoration request was to access a previous generation of data, selected by 52% of users. Ransomware was a major cause of that need, Farajun said.
The second most common reason for data recovery was user error or accidental deletion, with 13%. A lost or stolen device was third with 10%. Interestingly, disaster was only picked by 6% of respondents, according to the data recovery report.
Asigra is working on improving cybersecurity and how it can best combine with data protection, Farajun said. In the face of the growing threat of ransomware, Farajun also suggested organizations educate their employees, have strong anti-virus protection and back up their data.
The average size of a recovery across all sectors was 13 GB.
Farajun described cost as the bane of a company’s relationship with its backup vendor.
“Mostly [companies] don’t feel they can do anything about it,” Farajun said. “You can do something about it.”
In 2013, Asigra launched its Recovery License Model and now almost all of the vendor’s customers use it. Pricing is based on the percentage of data recovered over the course of a contractual term, with a ceiling of 25%.
Asigra did a healthy amount of research before launching the model. It looked into other markets, such as the music and telecommunications industries, and assorted “fair-pay” cases. Music customers, for example, can now buy one-song downloads vs. an entire album that they may not listen to in its entirety.
“What happened?” Farajun said. “People bought boatloads and boatloads of songs.”
Asigra had been nervous when undertaking the new model. It anticipated a three-year dip but revenue started to go up after 12 months, Farajun said.
So why hasn’t this model caught on more in the backup market?
“There’s no incentive for software vendors to reduce their prices,” Farajun said. “We’re trying to price based on fairness.”
Farajun said the data recovery report vindicates the vendor’s underlying premise.
“People don’t recover nearly as much as they think they do and they overpay for their backup software.”
Data protection provider Druva has launched platform-as-a-service capabilities to support indexed search queries of data across local and public cloud storage.
The Druva Cloud Platform is designed to help enterprises better manage and use information related to analytics, compliance, e-discovery and records retention. More than 30 RESTful APIs are included to allow third-party vendors to access data sets in Druva storage.
The API calls allow disparate information management applications to pull data directly from Druva InSync and Phoenix storage platforms. Druva cloud storage uses Amazon Web Services or Microsoft Azure as a target destination for inactive data that companies need to keep for legal regulations.
Global source-side data deduplication creates a single gold copy in the cloud. The Druva cloud technology takes point-in-time snapshots of queried data and applies advanced encryption. Changed data blocks are synchronized to deduplicated data sets in Amazon Web Services or Microsoft Azure.
The APIs allow disparate information management systems to communicate directly with Druva to improve data hygiene, said Dave Packer, a Druva vice president of product and corporate marketing.
“We designed Druva Cloud Platform so your data doesn’t have to traverse across corporate networks. We take care behind the scenes to ensure handoffs occur accordingly, without taxing internal systems,” Packer said.
Druva’s SaaS pricing is based on deduplicated data and starts at $6 per user per month.
Cloud NAS vendor Panzura is expanding into archiving.
The vendor today made available Freedom Archive software that moves infrequently accessed data to public clouds or low-cost on-premises storage.
Panzura CEO Patrick Harr describes Freedom Archive as storage for “long-term unstructured archived data that now sits on-premises on traditional NAS or tape libraries. The key thing is, it’s for active data.”
Harr said target markets include healthcare, video surveillance, gas and seismic exploration and media and entertainment.
Freedom Archive is a separate application from Panzura’s flagship Cloud NAS platform, which caches frequently used primary data on-site and moves the rest to the cloud. Freedom Archive is available on a physical appliance or as software only. It uses caching algorithms and smart policy manager to identify cooler data and move it from on-premises storage to the cloud. Freedom Archive compresses, deduplicates and encrypts data at-rest and in-flight.
Freedom Archive supports Amazon Web Services, Microsoft Azure, Google, and IBM Cloud public clouds, and private object storage from IBM, Hitachi Data systems and Dell EMC. Customers can download the software from Panzura’s web site. Pricing begins at less than two cents per GB per month, which does not include public cloud subscriptions. There is a 30-day free trial period for the software. IBM Cloud is offering 10 TB of free storage for 30 days and AWS will give 10 TB of free storage for 14 days to Freedom Archive customers.
Harr said Chevron, American College of Radiology, NBC Universal, Time Warner Cable, and law enforcement agencies already use Freedom Archive. The product became generally available today.
An expansion into archiving was among the goals Harr laid out when he became Panzura CEO earlier this year.
Harr emphasized Freedom Archive is for active data rather than cold data that rarely if ever needs to be accessed. That means Panzura is not competing with public cloud services such as Amazon Glacier, Microsoft Azure Cool Blob and Google Coldline storage.
“This is complementary to what Google, Azure and AWS do, not competitive,” he said. “Glacier is not for active data, and it’s extremely expensive to pull data back from Glacier. Ours is a hybrid cloud where you still have a performant nature to your data.
“Chevron has to access data in real-time instead of waiting for a slow response that doesn’t meet the business need. In the medical space, you don’t want to have to wait when you pull back an MRI.”
Cash-bleeding all-flash array vendor Violin Memory today filed for Chapter 11 bankruptcy, and will seek to sell off its assets at an auction next month.
After years of financial problems, Violin petitioned the U.S. Bankruptcy Court for the District of Delaware for Chapter 11 relief. In a prepared statement, Violin CEO Kevin DeNuccio said the vendor will continue operations during the bankruptcy period while hoping to sell its assets.
“We are taking this action, which should conclude by the end of January 2017, to bolster Violin’s ability to serve the needs of its customers,” DeNuccio said. “Violin intends to continue to sell solutions to customers and prospects as well as service and support customers during this restructuring.”
Violin’s problem is it doesn’t have enough customers, and has been unable to come close to profitability since becoming a public company in Sept. 2013. Violin has twice conducted extensive searches for a buyer without success, and its 2016 sales slowed to a trickle.
Failing to find a buyer, Violin has cut expenses through layoffs and pay reductions. The vendor reduced headcount from 437 employees in Jan. 2014 to 82 through several staff reductions.
DeNuccio took a pay cut last week, dropping his salary from $750,000 to $150,000. The bankruptcy filing placed Violin’s average monthly payroll at approximately $758,000 with another $109,000 per month in health benefits. The vendor said it also owes approximately $244,000 in sales commissions to employees.
According to the petition, Violin will have $3.62 million in cash at the end of this week and that total will drop to around $1.6 million by Jan. 20, 2017. Most of the operational expenses over those six weeks will be payroll-related. Violin has lost $25.5 million, $22.2 million and $20.1 million over the last three quarters.
Violin’s executive team and directors have tried in vain to find a buyer since late 2015. Violin hired Jefferies Group as its financial adviser in Nov. 2015. Jefferies contacted 39 strategic and eight financial sponsors, according to Violin’s court filing. Those contacts resulted in nine parties signing confidentiality agreements and 10 parties conducting management meetings. But none made offers and the search for a buyer ended in March, 2016 when Violin instead put restructuring plans in place.
Violin’s sales decreased, as it reported $10.9 million, $9.7 million and $7.5 million in revenue over the past three quarters. Those totals were especially disappointing considering the market for all-flash arrays is one of the hottest in the storage industry.
Violin’s last-ditch effort to increase sales with its launch of a new Flash Storage Platform array in September came too late to stave off bankruptcy.
Violin hired another financial adviser, Houlihan Lokey Capital, in September to seek a buyer for the company. Houlihan contacted 202 potential strategic buyers and 78 financial sponsors, according to court filings. A total of 26 parties signed confidentiality agreements and 13 conducted management interviews but none submitted letters of interest.
In recent weeks, Violin sought debtor-in-possession financing that would allow Violin to remain in control of its assets through Chapter 11. Again, it found no interest. Those failed proposals “have necessitated the filing of this case,” according to the filing.
Violin asked the court to approve a Jan. 13 bid deadline for the auction, and is looking to finalize any sale by Jan. 20.
One of the first all-flash vendors on the market, Violin claims 58 U.S. patents and 64 foreign patents. It has another 22 U.S. and 38 foreign patents pending.
Cloud file storage vendor Nasuni picked up $25 million in funding today, bringing its total to $80.5 million. The vendor’s executives expect the latest funding to bring it to cash-flow positive status in 2018.
Nasuni was one of the original cloud gateway startups, launching its Nasuni Filer in 2010. Nasuni software caches active data on-premises and moves other files off to public clouds, mainly Microsoft Azure, Amazon Web Services and IBM Cloud Services.
Nasuni’s software uses the UniFS cloud-native file system. It ships on Dell servers or runs as a virtual appliance to provide an edge connector. Customers can then expand capacity without adding hardware by sending data to a public cloud.
“We solve a storage problem, although we don’t actually store a gigabyte of data,” said Scott Dussault, Nasuni COO and CFO. “Our software enables customers to run a file system so they can have unstructured data management in the cloud.”
Dussault said the vendor will expand its sales footprint in North America and Europe to go with its 2016 push into the U.K. He said the expansion will help Nasuni attract larger customers.
“We started out in the SMB space, moved to the mid-market in 2014 and now we’re also selling to the uber-enterprise,” he said. “Companies are creating a strategy around the cloud, using Nasuni as their file system and mostly private cloud vendors for object storage.”
Other cloud gateway vendors include Panzura and Ctera. Microsoft acquired one-time Nasuni rival StorSimple and EMC bought TwinStrata. Dussault said Nasuni still competes mainly with traditional NAS products from NetApp and Dell EMC.
Dussault said Nasuni grew more than 75% in bookings and revenue in 2016. He said the company has 115 employees and he expects the funding to fuel greater than 25% headcount expansion in 2017.
“This round keeps us on the path of cash flow/break even in 2018,” he said.
The E funding round included $17.5 million in equity funding, led by new investor Sigma Prime Ventures, plus $7.5 million in venture debt financing from Eastward Capital.
Software-defined storage vendor Primary Data’s open standards parallel NFS contributions made it into the NFS 4.2 standard, which could help the startup make inroads with scale-out storage customers.
Primary Data’s contributions to NFS 4.2 include enhancements to the pNFS Flex File layout that allows clients to provide statistics on how data is used and the performance of the underlying storage.
NFS 4.2 enables clients and application servers to natively support data virtualization and mobility features. That plays well with Primary Data’s DataSphere software that virtualizes different types of storage into tiers using a single global data space.
“Data virtualization sits between hardware and storage arrays below us and the virtual machine space above us, virtualizing compute resources,” Primary Data CEO Lance Smith said. “We separate the logical view of data from where it’s physically stored. Now we can put data on the right type of storage without bothering or interrupting the application. To do that we need a touch point on the client, and that’s what this is about. When you put our metadata software into the infrastructure, that’s where virtualization comes alive.”
DataSphere supports SAN and DAS as well as NAS, but the integration of Primary Data technology into NFS 4.2 fits with scale-out NAS customers. The NFS 4.2 spec was completed in November.
“We are heavily engaged with media and entertainment companies,” Smith said. “They can now do clustering of their storage, even if it’s from different vendors. Oil and gas is right behind, looking for performance and scale-out. Financial service firms have about 20 percent of their data that’s super hot and needs to be on the highest performance tier, they want stuff migrated to a cheaper tier and use cloud and object storage. But that migration has to be seamless and not disrupt the application.”
Western Digital told financial analysts its business will be better than expected this quarter, and added solid-state drives (SSDs), hard-disk drives (HDDs) and an all-flash array platform to its portfolio.
WD updated its quarterly forecast to $4.75 billion from $4.7 billion at its analyst day this week. It also gave details on NVMe products from its SanDisk and HGST acquisitions and expanded its helium HDD line.
New NVMe all-flash array coming
WD previewed a 2U all-flash platform that the vendor claims will deliver 18 million IOPS using NVMe over PCIe fabric. The first system is due to ship in the first half of 2017. WD pledged to contribute software supporting the platform to the open source community.
WD will position the NVMe array for real-time and streaming analytics applications such as credit card fraud detection, video stream analysis, location-based services, advertising servers, automated systems, and solutions built on artificial intelligence or machine learning (ML).
The new system is part of WD’s InfiniFlash brand it gained as part of its SanDisk acquisition. Dave Tang, GM of WD’s data center systems, said WD will eventually expand the platform to include NVMe over Ethernet, which supports longer distances than PCIe but has more latency. Scalability of NVMe over PCIe is limited to either a single rack or an adjacent rack.
“We think customers interested in ultimate performance will go to NVMe over PCIe, and those looking for scalability may opt for NVMe over Ethernet,” Tang said. “They will co-exist and serve different purposes in the data center.”
Tang said he suspects NVMe over Ethernet support remains a few years away. Widespread adoption will require more expensive 100-Gigabit Ethernet. Also, NVMe over Ethernet standards are still evolving.
SSDs, HDDs expand capacities
WD added two SSDs, the Ultrastar SN200 and Ultrastar SS200 Series. The SN200 is an NVMe PCIe SSD and the SS200 is a SAS SSD drive. Both are available in 2.5-inch and Half-Height, Half-Length form factors and in capacities up to 7.68 TB.
The NVMe SSDs are built for cloud and hyperscale storage and big data analytics.The SAS SSDs are aimed at hyper-converged and other server-based storage that use a dual-port design and SAS interface. The UltraStar line comes from WD’s 2012 HGST acquisition.
WD also said it would ship new higher capacity helium and Singled Magnetic Recording (SMR) HDDs. The Ultrastar He12 is a 12 TB 3.5-inch SAS/SATA drive. It will be the highest capacity helium drive on the market. WD and rival Seagate currently ship 10 TB Helium drives. WD will also add a 14 TB SMR He12 drive, which surpasses its current 10 TB SMR capacity drives.
The Ultrastar SN200 and SS200 SSDs are expected to be generally available in the first quarter of 2017 with the He12 Helium drive expected in the first half of 2017 and the 14 TB SMR drive around the middle of next year.
Object storage provider SwiftStack Inc. has added hybrid cloud synchronization to its OpenStack-based software controller.
SwiftStack Cloud Sync allows data or subsets of data to exist simultaneously behind a firewall and in multiple public clouds. Customers with private SwiftStack cloud storage could create data copies in Amazon Simple Storage Service (S3) and Google Compute Platform and replicate between the two.
Cloud Sync replicates native objects between physical nodes and the public cloud. SwiftStack does not charge extra to support multiple copies of objects in different locations. The hybrid cloud topology places data in a single namespace.
Mario Blandini, SwiftStack’s vice president of marketing, said Cloud Sync policy management extends the storage of a physical enterprise data center.
“Cloud Sync moves your data close to where your application is running. We synchronize your data to the public cloud. It’s not a one-time copy; it’s a continuous sync based on policies that are applied to S3 or Google (storage),” Blandini said.
Distributed users can collaborate securely by accessing the same data bucket via Cloud Sync. Other use cases include active archiving, cloud bursting, offsite disaster recovery and cold archiving with Amazon Glacier and Google Cloud Storage Nearline.
Blandini said SwiftStack customers requested additional cloud support beyond Amazon Web Services.
“Customers buy servers from multiple vendors,” he said. “They want the same consumption experience with the public cloud. They want to have data in Amazon and Google and be able to balance between them.”
Existing SwiftStack customers can obtain Cloud Sync at no charge. SwiftStack’s capacity-based licensing starts at $375 per terabyte annually for a minimum of 50 TB.