While IDC probably won’t release its hyper-converged market share numbers for the first quarter until late June, it appears Dell will stay No. 1 on the HCI revenue hardware and software charts.
IDC recognized hardware and software HCI revenue separately for the first time in the fourth quarter of 2017. Dell EMC led hardware with 27.8% share, ahead of second-place Nutanix at 19.5%. IDC recognized Dell-owned VMware as the software HCI market leader with 32.4% of the market, with No. 2 Nutanix at 29.5%.
Nutanix, which is switching to a software-centric business model, reported 41% revenue growth for the first quarter of 2018. Dell said its Dell EMC VxRail HCI appliance and VMware’s vSAN HCI software each grew more than that.
VMware reported vSAN licenses grew 70% year-over-year, including vSAN sold on Dell EMC vXRail appliances.
On VMware’s earnings call, last week CEO Pat Gelsinger said the company finished the quarter with over 14,000 vSAN accounts, and its 10 largest deals included vSAN.
Like Nutanix CEO Dheeraj Pandey, Gelsinger said he still sees plenty of room for growth in the HCI market.
“I tease my team saying, ‘how many vSpheres have storage attached to it?’” The answer is 100%,” Gelsinger said. “And I’d say ‘that’s your market opportunity for vSAN, 100% of vSphere attach rate.’ So I am pushing them to continue to make it that simple and easy that we are attaching vSAN to every one of our compute nodes as part of this larger migration to the complete solution, either software or software plus hardware.”
Gelsinger said that migration is well under way as VMware customers have gone from licensing vSAN separately to using it as part of complete HCI bundles such as VxRail. Those HCI packages often include NSX software-defined networking.
“So that’s the shift in the marketplace we are seeing, where I will say people are no longer putting the Tinker Toys together themselves,” he said. “So, the storage piece by itself is a large market that we have lots of growth to go yet.”
Qumulo closed a $93 million funding round today, giving the scale-out file storage vendor “fuel” to compete with the established players in a rapidly expanding market.
BlackRock Private Equity Partners led the Series D round, which brings Qumulo’s total funding to over $220 million. Goldman Sachs and Western Digital also participated in the funding, along with previous Qumulo investors Highland Capital Partners, Kleiner Perkins Caufield & Byers, the Madrona Venture Group, and Valhalla Partners.
Qumulo CEO Bill Richter hinted that this may be the last venture funding the company will need.
“This will fuel the company’s growth for many years to come,” Richter said. “This is enough fuel that we need for the foreseeable future, which is important in this industry.”
He said Qumulo will use the funding to add sales and marketing people, and “especially code-writing engineers” to enhance the Qumulo File Fabric (QF2) software. He sees the market as ripe for software-defined and cloud-based NAS storage in the wake of fast-growing file data driven by technologies such as artificial intelligence, machine learning and 4K video.
Richter said Qumulo is innovating with QF2’s analytics and the ability to run in the cloud, while the established NAS products from NetApp and Dell EMC Isilon are stagnating. Qumulo’s founders were original Isilon developers, and Richter was president of EMC’s Isilon division for three years. He became Qumulo CEO in November, 2016.
“The last generation of storage was all about the box,” Richter said. “The next generation is about the data and the place where data lives. How can we allow customers to interoperate between the public and their private cloud for a true hybrid experience? And how can they really understand their data and not just their storage? Legacy guys, which I was part of for many, many years, were just churning out more boxes.”
Richter said Qumulo has hundreds of customers, including nine leading movie studios and Fortune 500 companies. He said Qumulo will use the funding to grow aggressively from its current 180 employees, although he declined to share specific hiring goals.
Richter also said the funding from Black Rock and Golden Sachs does not mean an initial public stock offering is a short-term goal.
“Eventually that might be in the cards, but now we’re focused on growing Qumulo and creating a long-term sustainable business,” he said. “Don’t expect us to try and put ourselves out to the public market at the first opportunity.”
Despite the cash infusion, Qumulo still faces large challenges. The vendor is far from profitable and is battling giants NetApp and Dell at the top of the market as well as other cloud file system startups such as Elastifile, WekaIO and Quobyte. Object storage vendors are increasingly adding file systems to their products as well.
And despite citing multi-cloud as a big selling point, Qumulo only supports AWS now. Richter said Microsoft Azure and Google Cloud Platform support is on the way. Qumulo is also exploring more hardware partnerships, such as its deal to sell its QF2 on HPE Apollo servers.
“Customers want a better, more modern choice,” said Richter, who now has the funding to fuel those code-writing engineers to help develop that modern file system.
Hewlett Packard Enterprise expanded its M-series Ethernet storage switching with a smaller switch and orchestration software designed to improve management features for iSCSI SANs.
HPE first launched the M-series in 2017, using Mellanox’s chipset. The M-series is one of three HPE storage switching families. The others are the B-series from Broadcom/Brocade and the C-series from Cisco. While the B and C series are Fibre Channel platforms, the M-series is Ethernet-only.
The new SN2010M is a smaller version of the first three M-series switches released in last year. The new switch is a half-width top of rack switch with 18 ports that support up to 25 Gigabit per second Ethernet and four 40/100-GigE ports. Like the larger M-Series switches, it supports zero packet loss and even bandwidth allocation for predictable performance.
The new HPE Smart Fabric Orchestrator (SFO) software handles management and configuration features such as zoning and provisioning that have long been valuable to Fibre Channel but difficult to set up and manage. The design goal is to eliminate the need for SAN administrators to run storage networks.
Marty Lans, HPE’s GM of storage connectivity, said 80% of data is on Ethernet-based secondary storage platforms. That includes hyper-converged storage, virtual SANs and backup data.
“What Fibre Channel does for primary storage, we’ve done for everything else,” he said of SFO’s management capabilities.
That makes Ethernet-based iSCSI storage important for the emerging NVMe-over Fabrics wave, Lans said. HPE has been building out its secondary storage ecosystem through partnerships with vendors such as Qumulo, Cohesity, Rubrik, Hedvig, Cloudian and Scality. It also has its own SimpliVity hyper-converged platform.
HPE’s major primary SAN platforms – 3PAR and Nimble – support Fibre Channel and Ethernet, although 3PAR’s customer base is mainly Fibre Channel and Nimble’s is mostly iSCSI.
“We realized a big part of the opportunity around NVMe is iSCSI-based,” Lans said. “If we can take the features of Fibre Channel and move it to Ethernet, customers have a choice. They can pick Ethernet instead of Fibre Channel. We’re not suggesting that, we just wanted to build another protocol for applications.”
Lans said industry-wide attempts to make Ethernet better for storage – including Converged Enhanced Ethernet and Data Center Ethernet – overlooked the management capabilities.
He said SFO eliminates the need for interoperability testing between different vendors’ switches. “Wouldn’t it be great if you didn’t have to go to a Website and look it up?” he said of interoperability lists. “We wanted to settle the politics for this. We layered on orchestration software and did all the interoperability testing.”
Dell Technologies on Monday projected positive share growth in enterprise storage, the first quarter that has happened since the Dell and EMC merger closed in 2016. The next question is whether the growth in storage is sustainable, attributable to a seasonal anomaly, or requires a winnowing of redundant storage platforms.
Dell EMC storage revenue climbed 10% during the April quarter, generating $4.1 billion, which the vendor attributed largely to a triple-digit increase in sales of Dell EMC VxRack and VxRail hyper-converged appliances.
The vendor also claimed progress in midrange storage sales, a chronic trouble spot, although company executives declined to provide specific revenue figures. In a related move, Dell said it hired several hundred sales reps last quarter specifically for storage.
Only a passing reference was made to sales of the flagship all-flash Dell EMC PowerMax (formerly VMAX), which exited fiscal year 2018 strongly with a $5 billion run rate. Dell executives also made off limits for discussion the status of internal talks regarding a potential merger with VMware, the virtualization subsidiary it picked up through EMC.
Dell EMC rides momentum of cloud, software-defined storage
Dell’s combined revenue last quarter surged 19% to $21.4 billion. That’s on a net loss of $538 million. Non-GAAP operating income was $2 billion, up 42%.
After taking on sizable debt to acquire the EMC storage business, Dell has paid down about $13 billion of the financial obligations, including a $2.5 billion payment last week to satisfy investment-grade notes. All told, Dell’s debt service is roughly $53 billion, excluding debt related to the Dell Financial Services arm.
The Dell EMC storage portfolio is part of the vendor’s Infrastructure Solutions Group (ISG), which also includes PCs, servers and networking. Overall ISG product revenue of $8.7 billion was up 25%, getting a bounce largely from customers using Dell PowerEdge servers to run software-defined storage.
A strong quarter by VMware generated license revenue of $2 billion last quarter, including double-digit bookings for compute management, end user computing, NSX and VSAN/VxRail. Deferred revenue of $21 billion is up more than $3 billion from the year-ago period, due to recurring maintenance contracts and increased customer adoption of Dell’s consumption-based services and software.
Research firm IDC projects spending on storage systems will grow nearly 10 percent in the first quarter. The Dell EMC storage business is positioned to equal if not outpace the market average, said Jeff Clarke, a Dell vice chairman of products and operations.
Clarke said data centers increasingly want to run software-defined storage to support analytics and similar workloads in a multicloud environment. That helped Dell post its sixth consecutive quarter of increased server revenues, although rising DRAM prices pose potential headwinds.
“Storage had a solid quarter as a portion of the Q4 demand (in March). Following positive demand growth in our fiscal fourth quarter, we expect to gain share year over year in storage when the first-quarter industry share numbers are final. This will be our first quarter (with) storage share gains since we closed the EMC transaction,” Clarke said.
So has the flagging Dell EMC storage business finally turned a corner? Clarke sounded bullish on the idea, but acknowledged the recent revenue figures are only a snapshot in time.
“This quarter’s results are an indicator that our actions are having a positive impact. We continue to be focused on improving storage velocity and acknowledge that results maybe lumpy quarter-to-quarter. That said, our FY 19 storage plan is built on revenue growth and to deliver share gain over the full year.”
Despite several all-flash storage moves at Dell Technologies World in May, the only acknowledgement of flash was the new Dell PowerEdge models that integrate NVMe flash drives and support NVMe over Fabrics. Dell reported in March it exited the fiscal year with a $5 billion run rate in all-flash, and that was prior to launching a rebranded VMAX.
All-flash pioneer Pure Storage in April posted year-over-year revenue of 40% and NetApp did even better, closing its year at $2.4 billion and a 43% year-over-year gain. As is its custom, Dell EMC does not provide revenue guidance to Wall Street.
If you read between the lines, it’s possible to imagine that Dell EMC may start streamlining its storage gear. Clarke said the Dell EMC storage road map will aim for “simplification,” which he said started with the reorganization that resulted in creation of the ISG last year.
“Since then we have started simplifying the portfolio into a road map squarely directed at meeting customer needs and winning in the market. We will offer a single industry-leading solution for every segment in which we compete, including entry level, midrange, high-end and unstructured. Those solutions will put forward the best features and innovation from our current portfolio, plus new innovation to ensure we are offering the industry’s best solutions.”
“We are not exiting any market. All the products on our current road map will continue to be supported over their lifecycle,” as part of Dell’s customer loyalty program.
The flurry of activity surrounding on-premise object storage specialist Cloudian suggests how rapidly the market has been moving during the last year.
Cloudian CEO Michael Tso said 2017 was the startup’s best year, as revenue grew by a factor of 3x and the customer count soared past 200. He said 89% of Cloudian object storage sales came through value-added resellers by the fourth quarter, and the trend continued into 2018 – a significant jump over the 25% in third-party sales in 2015 and 36% in 2016.
“That really is a signal to me that the product is ready for a broader channel,” Tso said.
Cloudian launched in late 2011 with a focus on prominent Japanese telcommunications/service provider customers NTT East, NTT Communications and Nifty that needed multi-tenant, geographically distributed storage after the Fukushima earthquake.
Tso said he expects Cloudian to become profitable over the next few years, with a possible IPO down the road. But he said growth is more important than profitability right now.
“Our board and our investors are telling us to grow just as fast as we can and don’t worry about profitability,” he said. “I think if we stopped growing as aggressively as we are, we either would be able to be profitable this year, or if I look at the numbers carefully, maybe we could have been profitable even earlier. We are looking at a potential IPO probably three to four years down the road. We are not really in a hurry. We expect to be profitable before that.”
Cloudian is concentrating on partnerships with channel and OEM partners to grab a significant share of the storage market. The object vendor followed its 2016 OEM deal with Lenovo with an EMEA-based joint reseller agreement with Hewlett Packard Enterprise in late 2017. In 2018, Cloudian partnered with Machine Box on a machine-learning option and made available a Cloudian object storage “HyperStore Test Drive” for Google Cloud Platform.
Also, late 2017 conversations with Cisco Systems led to a significant investment earlier this year from Digital Alpha, a private equity firm started by former Cisco executives. Digital Alpha made a $25 million equity commitment to Cloudian and set up a utility financing facility of up to $100 million.
“The goal for the $100 million is to set up a separate company that would purchase appliances and solutions from Cloudian and be able to provide those to the end user through a paper drink consumption model,” Tso said. “They will add more gear when you need it, and they’ll remove gear when you’re trying to take it away. It’s just like the way cloud works except it’s cloud being put into your own data center, because our product is only sold into on-prem environments.”
Cloudian expanded in March with the acquisition of Infinity Storage, an Italian file-based software-defined storage vendor. Cloudian already used Infinity’s technology in its HyperFile appliance that combines file and object storage.
“They make an NFS/CIFS front end that can move data into object storage or into the cloud,” Tso said of Infinity Storage. “We partner with every one of the gateway companies out there, but we weren’t really happy with any of their solutions. The problem with a lot of products out there is that they’re not in the kernel space. File systems have traditionally always been done inside the kernel. It’s really the only way to do it that’s really robust, but it’s very hard. We spent a year testing pretty much every vendor in the market, and we eventually came on this small company based out of Milan. They’ve been doing it for over 10 years.”
Crossbar recently chalked up another milestone in its quest to get on-chip non-volatile resistive RAM (ReRAM) technology to market.
The Santa Clara, California-based startup licensed its core ReRAM intellectual property to Microsemi, a semiconductor supplier to the military and aerospace industry. The companies plan to collaborate on the research, development and application of the Crossbar ReRAM into Microsemi products designed for 16, 14, and 12 nanometer (nm) process nodes.
ReRAM is a type of non-volatile memory that consumes less power and offers faster reads and writes, higher endurance and greater storage density than NAND flash. But Crossbar does not position its ReRAM as an alternative to technologies such as flash-based solid-state drives (SSDs). The initial use cases for ReRAM will more likely be under the covers in CPUs, field programmable gate arrays (FPGAs), and system-on-a-chip (SoC) architectures, possibly as an alternative to slower static RAM (SRAM) or less energy-efficient dynamic RAM (DRAM), according to Sylvain Dubois, vice president of business development and marketing at Crossbar.
Dubois expects the embedded Crossbar ReRAM technology to reach products that businesses or consumers might use in 2019.
ReRAM use case example
In the meantime, Crossbar demonstrated potential use cases in “artificial intelligence applications at the edge” – designed for devices such as surveillance cameras and mobile phones – at the recent Embedded Vision Summit in San Jose, California. The startup showed off ReRAM test chips with applications designed to recognize faces and license plates. Dubois said the test chips integrate the algorithms and the database, enabling classifications to be done efficiently at low latency and low power, with no need to communicate with a distant cloud-based database.
Alternative technologies that a designer might have chosen for such applications include static RAM (SRAM) and dynamic RAM (DRAM), according to Dubois. But he claimed SRAM would have been slower and DRAM would have consumed more energy than the Crossbar ReRAM.
Microsemi did not disclose the types of products that might incorporate the Crossbar ReRAM technology nor the foundry with which it is working. The company’s product line includes PGAs, controllers, communications chips and artificial intelligence computing chips.
“What is unique is that this Microsemi deal is targeted at 1x nanometer,” Dubois said. “Getting to the point where you can prove the scalability is a major milestone. Now we have not only a customer but we also have a foundry that is getting access to embedded ReRAM at 1x nanometer.”
Dubois said Crossbar started its non-volatile ReRAM commercialization phase on 40 nm process nodes with Semiconductor Manufacturing Corp. (SMIC). He said Crossbar also works with two other foundries and licenses its ReRAM to about a dozen chip designers.
Crossbar ReRAM nanofilament technology is built on standard complementary metal oxide semiconductor (CMOS) processes, and the company claims it will scale to below 10 nm without impact to performance.
“The big trick is to get into high volume production because the most important thing in semiconductors is to get the volume high enough so you can drive the costs out,” said Jim Handy, general director and semiconductor analyst at Objective Analysis.
Handy said the Microsemi agreement with Crossbar shows the semiconductor supplier took a hard look at the ReRAM, decided it’s a good technology, and expects other companies to sign on. He said the 1x nanometer process achievement indicates Crossbar could be looking at use cases where NOR flash won’t work any longer due to scaling issues.
“Microsemi supplies stuff for aerospace, and in space, there’s a lot of radiation. Radiation tends to cause NOR flash, or any flash, to lose its content. The radioactive particles go through the chip and drag the electrons out of the floating gate,” Handy said. “The Crossbar metal filament technology doesn’t use electrons. And nuclear particles can go zipping through the chip and not destroy bits.”
Handy said enterprise users aren’t likely to notice ReRAM in their servers or storage systems because it would be buried in places they won’t see. But he said Crossbar ReRAM could lower the cost of SSD controllers marginally by allowing them to scale better.
“Now 80% or more of the cost of the drive is the NAND flash. So this is going to make a small difference, but it will make a difference,” Handy said.
Komprise struck its first major reseller agreement, partnering with IBM to offer Komprise Intelligent Data Management software with the vendor’s storage portfolio.
Customers can use Komprise software to move file-based data from network-attached storage (NAS) systems to IBM cloud-based object storage, according to Krishna Subramanian, co-founder and chief operating officer at Komprise.
Users can download the Komprise Observer virtual appliance and point it at their NAS systems. The Komprise software analyzes how much hot and cold data they have and the file growth rate to help them set policies to move data to on-premise or off-site IBM Cloud Object Storage for archival or disaster recovery (DR) purposes. The Komprise Director management console, which can run at the customer’s data center or in the cloud, displays the potential savings considering NAS, backup, DR and new storage target costs.
The Komprise software also handles the file-to-object mapping and non-disruptively migrates the data from the NAS system to the IBM Cloud Object Storage. End users can still access the data with the same permissions and metadata from their source NAS systems, even though the files have been transferred to cloud-based object storage on the back end.
“When we move a file out of NAS into object storage, we put in a link in that NAS so when a user goes to open the file, it looks like the same file. It has all the metadata properties and everything,” Subramanian said. “A Komprise Observer will actually respond, and it will map the object back to file and return it. But that whole handshake is transparent to users and applications.”
Komprise vs. cloud gateways
Subramanian said, unlike many cloud gateway appliances, Komprise does not aim to shift all data to the public cloud and cache the hottest data in local appliances. She said that approach can become expensive if the customer needs to retrieve data from the cloud.
“We’re simply providing a more cost efficient way to manage the data. Essentially we’re saying, ‘Look, your NAS is great for your hot data. But for the 80% of your data that is rarely getting cached, and you need to keep either for business or compliance reasons, let us move that to a cost-efficient store,’” Subramanian said.
The Komprise software also enables customers to migrate data from one NAS system to another NAS system, if they want to replace or decommission file-based storage devices.
Subramanian said Komprise and IBM had joint customers in industries such as financial services, insurance and health care storing large volumes of data. Komprise is part of the Ready for IBM Storage and Ready for IBM Cloud validated solution directory. The new worldwide reseller agreement applies to IBM product sales and services teams as well as IBM channel partners, Subramanian said.
Komprise prices its software based on the amount of data under management. The startup offers a subscription model, at about a half penny per GB per month, or a perpetual license, at $120 to $130 per TB, according to Subramanian.
Earlier this month, Komprise shipped a new 2.8 software update. New features included support for European euros, British pounds and Japanese yen in the product’s ROI calculator, which originally displayed cost savings only in U.S. dollars. The 2.8 product upgrade also enhanced the NAS migration capabilities for SMB environments, adding the ability to preserve access control.
Hyper-converged pioneer Nutanix has a history of rapid revenue growth and wide losses. For last quarter, both of those were larger than expected.
Nutanix revenue grew 41% year-over-year to $289 million — $9 million above the top end of its previous forecast. Nutanix also lost $35 million, larger than expected and much more than the $20 million loss a year ago.
For this quarter, Nutanix forecasted revenue from $295 million and $300 million but heavier losses than analysts expected.
In an interview following the earnings call last Thursday, Nutanix CEO Dheeraj Pandey said the company’s move to a software-centric business model was partly to blame for the earnings miss. Nutanix is two quarters into the model, which no longer credits revenue from partners’ hardware sold with Nutanix software. Nutanix had $12 more in deferred revenue than expected last quarter due to the accounting change. Pandey said changes in the billing to revenue ratio makes it harder to accurately forecast revenue during the transition.
“That’s part of the software transition,” he said. “We don’t fully have an exact view of billings to revenue as we go through this transition.”
Nutanix also went on a hiring spree, adding more employees than in any previous quarter. CFO Duston Williams called it a “full court press” on hiring. That included the addition of 60 sales teams and 85 new employees through acquisitions of Netsil and Minjar. That helped run Nutanix’s expenses to $232 million in the quarter, $12 million above its guidance.
Nutanix remains in a growth stage, hoping to take advantage of the still rapidly expanding hyper-converged infrastructure (HCI) market. According to IDC, the hyper-converged market expanded 69.4% year-over-year for the fourth quarter of 2017. That should leave plenty of room for future Nutanix revenue increases.
“We are still discovering the total addressable market (TAM) of this architecture,” Pandey said. “Five years ago, nobody gave hyper-convergence a $100 million TAM but now it’s already close to $10 billion. A lot of the TAM expansion depends on category creators like Nutanix. We go and expand workloads, geographies, different kinds of mission critical applications, hardware plans on which we run. We keep expanding that, and the TAM should continue to grow.”
Much of that expansion is due to traditional storage systems becoming hyper-converged. Pandey said about 65% to 70% of Nutanix deals come from converting traditional storage products to HCI.
“Other HCI competitors don’t have that much HCI focus,” he said, referring to server vendors that have moved into HCI. “They have other three-tier products that they still sell, while we have conviction in our architecture.”
Despite its losses, Nutanix has more than $900 million in cash. Pandey said the vendor will continue to increase spending to add features and new products from internal development and acquisitions. The acquisition strategy will continue to focus on small deals that add key pieces of technology rather than large established companies.
“We’re not in the business of buying growth,” Pandey said. “We won’t just go and buy a customer base. We’re looking for companies that have great technology and awesome people that want to build this company into something bigger.”
Nutanix AHV hypervisor adoption is growing, although it doesn’t directly add to revenue because it is part of the Acropolis software stack. Nutanix claims 33% of its HCI nodes sold last quarter used AHV compared to 30% in the previous quarter. Pandey said AHV saves people from having to buy VMware licenses, but new features also sway customers.
“Well, I think we don’t lead with cheap,” he said on the earnings call. “Nutanix is not that cheap of a product. If anything, we get a lot of flak for being a premium product, but we lead with ease of use and that our stuff works.”
“We’re making microsegmentation simple to use, so you don’t have to pay for NSX,” he said, referring to VMware’s software-defined networking product. “Our microsegmentation is one-click, and one node at a time.”
BOSTON – At the first ZertoCON in 2016, analyst John Morency said that “IT resilience” is becoming the new “disaster recovery.” The concept at the time stressed continuous availability and proactively avoiding all recovery situations, versus just being able to recover from huge disasters.
Two years later, at ZertoCON 2018, IT resilience was the dominant theme. But that concept itself is evolving.
“The definition has changed,” Morency, a research vice president at Gartner, said in his keynote address Wednesday. “The scope has changed.”
Gartner defines resilience as “the ability of an organization to protect, absorb, recover and adapt in a complex and rapidly changing environment to enable it to deliver its objectives and to rebound and prosper.” That’s different from the classic recovery model with a focus on recovery time and recovery point objectives, Morency said.
“Backup can only take us so far,” Morency said.
Zerto’s new Elastic Journal, for example, which is scheduled for release with Zerto 7 in early 2019, provides continuous recovery points across data, files or virtual machines, going from seconds to years back. Zerto pitched it as a new way to do backup. The feature is a part of Zerto’s newly branded IT Resilience Platform.
“We don’t ever go to our backup solution for recovery,” said senior system engineer Jayme Williams, of material manufacturer TenCate, a Zerto customer since 2012.
Instead, TenCate uses Zerto for its journal file-level recovery.
Data protection criteria are changing, Morency said. In the past, functionality was the most important for products. Now it’s cost, ease of use and capability to support multiple data protection uses.
Needs and planning tips for IT resiliency
According to Morency, governance requirements driving organizations’ need for IT resilience include:
- close-to-continuous IT and business operations
- workload mobility
- sustainable data integrity, consistency, availability and accessibility
- cyberthreat mitigation
- IT service configuration, deployment and change agility
- detection of and response to potentially disruptive events in order to sustain business and IT operations
Ransomware is changing the game for cyberthreats. Speakers at ZertoCON 2018 noted that a ransomware attack is a “when” not “if” scenario. And traditional backup — with recovery point objectives measured in hours — may not cut it, as organizations will want to recover from just before the attack hit.
Gartner estimates that by 2020, 30% of organizations targeted by major cyberattacks will spend more than two months cleansing backup, resulting in delayed recoveries.
IT resilience management is also driving product convergence: in backup software, runbook automation, software-defined managers and cloud management platforms.
“It’s not about backup. It’s not about runbook automation,” Morency said. “It’s all of the above.”
For many organizations, the IT resilience scope is hybrid. According to Gartner research, nearly 80% of organizations say their data center capacity profile in five years will include some combination of on-premises and cloud.
Morency provided an action plan for organizations looking to begin an IT resilience journey:
- Monday morning: Benchmark your organization’s resilience and identify people, process and technology gaps that are specific to the support of mission-critical business processes and applications.
- 30 to 90 days: Develop and execute a plan for improving relevant resilience gaps for mission-critical processes and applications.
- 90 to 180 days: Prioritize gap closure for critical and important processes and applications; mitigate the resilience risks posed by key vendors and service providers.
NetApp CEO George Kurian says things have never been better for the storage vendor since he joined the company in 2011.
Kurian maintains NetApp’s all-flash platform is a hit, it’s the first major vendor with end-to-end NVMe array, it has solid connections with all major cloud vendors and now has a viable hyper-converged product. And he said NetApp is still making two SAN array displacements a day, while its main rival Dell EMC is struggling with its midrange storage and cloud strategies.
“What a difference a year makes,” Kurian said during NetApp’s Wednesday evening earnings call, reporting a better than expected 11% year-over-year revenue increase to $1.64 billion last quarter. “We improved the consistency of our results, expanded our market opportunities, and successfully accelerated our momentum. We are undoubtedly in the best position since beginning the transformation of NetApp.”
NetApp’s forecast failed to capture Kurian’s optimistic words, though. Investors were disappointed by NetApp’s guidance range of $1.365 billion to $1.465 billion for this quarter, a significant drop from last quarter. The stock price fell 3.9% to $66.79 Wednesday after the earnings report and then dropped more to $64.00 at today’s opening.
When asked about the tepid guidance, Kurian said: “We are very bullish on the strength of our product portfolio. Our philosophy is to build a plan that we can meet or beat and provide you more updates as we see more visibility through the course of the year.”
When Kurian became NetApp CEO three years ago, the vendor was a laggard in the emerging all-flash market. He said all-flash revenue grew 43% year-over-year last quarter and is at a $2.4 billion run rate for the year, putting it at or near the top of the overall storage market for that segment. He said most new arrays that ship now are all-flash configurations.
NetApp this month launched its All-Flash FAS (AFF) 800, part of a wave of NVMe arrays from major vendors. Dell EMC, Hewlett Packard Enterprise, Pure Storage and IBM also have new NVMe arrays out or coming, as do several startups specifically targeting that market.
Speaking of competitors, Kurian said Dell EMC still has a lot of work to do following its 2016 merger. He said the market leader needs to do more than rationalize its overlapping midrange arrays to stem share losses.
“I think what Dell has to do is not only rationalize their portfolio, but then to develop a coherent cloud strategy. That takes years of work. They’re years behind on everything from flash to cloud,” he said.
NetApp’s cloud strategy revolves around its Data Fabric and Cloud Volumes, which make its file services available in public clouds. NetApp Cloud Volumes is generally available for Amazon Web Services and in private previews for Microsoft Azure and Google Cloud Platform.
“The opportunity created by this part of our business is incredibly exciting,” Kurian said.
Kurian said NetApp’s FlexGroup feature that allows customers to cluster FlexVols has helped the vendor win deals from Dell EMC’s Isilon scale-out NAS product. “We have taken back several footprints from Isilon, and frankly, they’re trying to chase us now,” he said.
NetApp HCI is still in its early days, less than a year after launching. Kurian said the vendor is not chasing traditional hyper-converged use cases but concentrating on enterprises running mixed workloads. NetApp uses its SolidFire all-flash platform as the storage for HCI.
“We are not targeting the entire hyper converged market, but a very specific large segment of it where we think we’ve got a winning architecture,” he said.
Should we believe Kurian’s words or NetApp’s guidance? Either he is not as optimistic as he sounds, or he is setting things up for NetApp to beat expectations again. Check back in three months to find out.