SAN FRANCISCO – Pure Storage will push the theme of a “data-centric architecture” at its annual Accelerate user conference that begins Wednesday.
Data-centric architecture is Pure’s description for its new flash strategy. The strategy revolves around the concept that the storage array is becoming a commodity item, an afterthought for IT. Enterprise data centers instead want fast flash to deliver data as a service to any type of application.
Among the expected product highlights is a major upgrade to the flagship Pure Storage FlashArray block and file system, featuring a handful of highly dense models that extend nonvolatile memory express (NVMe) rack-scale flash across the product line.
This will be the first Pure Accelerate under new CEO Charles Giancarlo, who replaced Scott Dietzen last April. Dietzen remains as Pure’s charman.
Shared accelerated storage: Jargon or meaningful distinction?
Pure Storage has had an eventful year so far. Pure became a public company in 2015, and this year capped off a pair of key milestones: $1 billion in sales and achieving non-GAAP profit. Pure Storage launched in 2009 as one of the first vendors to sell only all-flash arrays.
Pure now wants to shift the focus away from hardware specs to software-defined storage features that exploit advances in flash technology. With the launch of Pure AIRI this year, the vendor moved into artificial intelligence.
Pure wants to lump its all-flash arrays under a recently developed hardware category known as shared accelerated storage. The term was coined by IT analyst firm Gartner to describe hardware gears equipped with NVMe over Fabrics capabilities.
“Our vision for data-centric architecture is that IT organizations need to think less about managing storage and more about being storage service providers to the rest of the organization” said Matt Kixmoeller, Pure’s vice president of strategy. “That’s a bit of a different mindset than just buying and running storage arrays.”
At Pure Accelerate in 2017, the vendor previewed a FlashStack converged infrastructure product based on Cisco servers and networking. The first iteration of that product will be made generally available this week.
As all-flash array pioneer Pure Storage celebrates its $1 billion in annual revenues, two of its former executives are planning the next big thing in flash.
Pavilion Data, whose CEO Gurpreet Singh and VP of global sales Dan Heydenfeldt came from Pure, today picked up $12 million in funding to market its NVMe over Fabrics (NVMe-oF) storage system. Singh said Pavilion is going after customers running applications built on a “new modern stack dominated by open source, massively parallel, scale-out, clustered databases and file systems.” In other words, it’s targeting storage for databases such as MongoDB, Spark, MySQL and Cassandra instead of Oracle and Microsoft SQL Server.
“Somebody gets to build the next billion dollar company riding on this modern data stack,” Singh said. “We believe we have the best architecture for these modern applications. We call it disaggregated shared storage, or rack-scale flash.”
Singh said current all-flash arrays — including Pure’s — are fine for traditional applications but not modern apps. “The old-school dual controller architecture, server-centric design exposes a lot of challenges when running these applications,” he said. “For example, performance density is just not there.”
He said a storage system to run these new apps must have performance, latency and bandwidth characteristics of direct attached storage yet easy to scale and use. “Today there are compromises,” Singh said. “You can go shared storage, but you lose performance. Or you stick in four to six NVMe cards per server, 40 servers per rack, tens of racks and you lose the serviceability and data management. “It requires a complete re-think of how you develop and architect a storage system. You can’t retro-fit that. The basic math doesn’t add up.”
Pavilion’s answer is the Pavilion Memory Array, which started shipping in early 2018.
Pavilion Data CTO VR Satish said a 4u Pavilion Memory Array can drive 120 GBps second with around 100 microseconds of latency and 20 million IOPS. The system uses x86 hardware and up to 72 standard 2.5-inch NVMe flash drives for a maximum capacity of 1 PB.
The back of the box resembles a network switch with a minimum of two line cards, each card containing four 100 Gigabit Ethernet ports and two controllers. Customers can expand to 10 line cards, 40 GigE ports and 20 controllers.
The array does not scale beyond one system, but Satish said 1 PB of storage with 72 drives “is more than enough.”
“Hyperscalers don’t want to be caught dead running Fibre Channel in their data centers,” Singh said.
Satish said Pavilion developed a clustered file system that supports RAID 6 data protection, non-disruptive upgrades, multi-pathing, thin provisioning, snapshots and clones. The array does not require any software to run on host servers.
Founded in 2014, Pavilion Data has just under 60 employees split between the U.S. and India. But the company has not had all smooth sailing so far. A founder and the original CEO Kiran Malwankar left Pavilion early this year. Pavilion also had some layoffs around that time, although it is hiring now.
Singh said another founder and current VP of engineering Sundar Kanthadai is in charge of development, and there has been no change in direction since Malwankar left.
He said Malwankar “left to pursue other opportunities” although Malwankar’s LinkedIn entry describes him as a “free bird.”
“It’s the natural course and evolution of a company,” Singh said. “People leave and new people come in.”
The new funding brings Pavilion Data’s total to $33 million. New investors Korea Investment Partners and DAG Ventures participated along with previous investors Kleiner Perkins Caufield & Byers, Artiman White Space Investments, and SK Telecom.
While Veeam Software uses its VeeamON user conference this week to further its push into the enterprise, Quest Software is making its own attempt to go the same route with its NetVault Backup.
With NetVault Backup 12.0, Quest has made the application more scalable – particularly for virtual machines. NetVault Backup 12.0 can run VMware plug-ins on any available proxy so users can back up VMs with a unified view and scale thousands of VMs. A new heuristic algorithm can load balance backup jobs across clients acting as backup proxies.
NetVault Backup 12.0 also supports application-aware storage array snapshots for the first time, although the only array supported now is Dell EMC SC Series (Compellent). Other enhancements include a single sign-on so users can log onto NetVault Backup by using Active Directory credentials, a push install feature to streamline product updates and installations, a new granular catalog search function and a new widget-based dashboard.
“We’ve decided to up the game with NetVault a little bit,” said Adrian Moir, Quest’s lead technology evangelist. “We wanted to go further into the enterprise space. We wanted to add more scale around protection of VMware, and give a single view of a larger environment by placing proxies under a single point of management. This is also our first entry into array-based snapshot. We’ve built a framework and will expand it, adding other arrays as quickly as we can.”
Also like Veeam, Quest is avoiding integrated systems in this age of converged data protection. Moir said Quest wants NetVault Backup to work with as many backup target options as possible rather than package it on an appliance.
“We’re quite happy to run on anyone’s hardware,” he said. “If people want to use a specific hardware platform, that’s fine with us. We’d rather offer the flexibility rather than give them something that doesn’t match the rest of their infrastructure. An appliance might be right-fitted, but sometimes it doesn’t match everything else. We’d rather be flexible and let them match the rest of their environment.”
CHICAGO – Veeam unveiled a copy data management feature at its VeeamON 2018 user conference Tuesday as part of its newly branded Hyper-Availability Platform.
Veeam DataLabs, part of the vendor’s flagship Availability Suite, allow organizations to create copies of production environments for uses beyond standard data protection. Those scenarios include test and development, DevOps and DevSecOps.
DataLabs is an element that has evolved over the past year, said Peter McKay, Veeam president and co-CEO.
“It’s a continuation of extending our platform for the enterprise,” McKay said at VeeamON. “Data is becoming more and more important. It’s the lifeblood of companies today.”
An organization can boot virtual machines from a backup or replica into a secure, isolated virtual environment. Veeam DataLabs expand on the functionality of Veeam’s Virtual Labs, enabling production-like instances of virtual environments. The feature replaces Veeam Virtual Labs.
Specifically, they help to ensure teams are developing and testing with the most recent data copies. Developers can spin up instances of the production environment as they design new features. The DataLabs provide sandbox environments to test new patches and updates. The copies also enable testing of security vulnerabilities without disrupting production systems.
In addition, an organization can examine and classify data, which helps comply with regulations such as the General Data Protection Regulation.
More than ‘always on’
Veeam DataLabs is one piece of how the vendor is expanding its data management functionality.
Veeam laid the groundwork at VeeamON 2018 for its strategy and vision of what it’s calling the “Hyper-Availability Platform for Intelligent Data Management,” a shift from its recent “Availability for the Always-On Enterprise” platform.
Hyper-availability is more than availability and being always on; it’s enabling artificial intelligence and being able to move data fluidly across a multi-cloud environment, said Danny Allan, vice president of product strategy.
Veeam executives noted the difficulties that organizations are having with their data management. Not only is the volume of data increasing exponentially, it’s also often on multiple platforms, in multiple locations.
Combining AI and automation will drive the most innovate disruptions of the next decade, Allan said. Veeam is building intelligence into its platform, but didn’t provide specifics Tuesday at VeeamON.
“Leveraging intelligence for better data outcomes is where we need to go as an industry,” Allan said.
Veeam recently reached 300,000 customers and claims to be growing by 4,000 per month.
Nutanix .NEXT 2017 focused on the hyper-converged infrastructure pioneer’s move to the public cloud with Xi Cloud Services. Nutanix .NEXT 2018 last week was a showcase for Nutanix Beam, the vendor’s software-defined networking service. Those launches went in directions most expected the vendor to eventually take hyper-convergence.
Nutanix .NEXT 2019 will likely have an edge computing launch, although it’s not clear that will be the main focus. But the HCI vendor last week unveiled Project Sherlock at Nutanix .NEXT 2018. A Nutanix executive described Sherlock as an enterprise cloud software stack to harness hundreds of zettabytes of data that will be captured by Internet of Things (IoT) devices.
Satyam Vaghani, Nutanix’s VP of IoT and AI, demonstrated Sherlock during one of the .NEXT keynotes. He showed how the technology could allow a retail store to use it to identify customers and automate check out in stores using technologies such as facial recognition. The software could tie into the store’s sensors, cameras, kiosks and rack servers at the edge and connect to the cloud. The demo included a SaaS console to support “planet scale” IoT infrastructure.
“Enterprise IoT is very complicated because of the variety of devices involved,” Vaghani said. “It’s planet scale, not data center scale. Platforms for IoT operations are very different than platforms we are used to running in the data center. Wouldn’t it be great if there was a way for Nutanix to provide a managed service that can be used to manage all of this planet scale infrastructure, all from a central place?”
Vaghani said the technology could allow users to walk out of a story without stopping at a checkout counter to complete a purchase, board an airplane without a boarding pass or wear an ID badge at work. “These are the headliner use cases of IoT, the biggest use case in the edge cloud,” he said.
Nutanix chief product and development officer Sunil Potti said he expects “more concrete announcements” for Sherlock at .NEXT 2019. “Now it’s in what we call the Series A financing mode,” he said.
Considering Xi still isn’t ready for wide-scale consumption a year after its launch, there is a good chance Sherlock won’t make it into next year’s use conference. Like Xi, Sherlock requires a significant amount of engineering work.
“Just like we couldn’t take our current OS and put it in Xi, we had to re-do some core components” for Sherlock, Potti said. “When we extend our stack to the edge, we can’t take our current storage stack. The amount of real time processing needed, the latency requirements, the footprint requirements … everything is different.”
Government IT wrestles with cloud costs, management
Nutanix Beam, a SaaS offering for managing cloud costs, resonated with users at Nutanix .NEXT 2018 who need to get a handle on public cloud costs. Cloud sprawl was a topic during a breakout session with admins from government agencies, a vertical that often has mandates to use public clouds.
Beam provides a dashboard showing all of an organization’s cloud services and monthly pricing. Admins can manage their cloud resources from the dashboard, and Beam recommends cost saving changes.
David Gokey, chief engineer for Mission Systems at NORAD-U.S. Northern Command (USNORCOM), said he saw Beam for the first time at the show and it fit the bill for what he’s been looking for.
“I need a cloud economist,” he said of the task of keeping track of the costs of all his agency’s cloud services. “I need somebody 100 percent dedicated to look at all the products out there and know how we can control that. It has to be automated.”
Derek Williams, director of data center operations for the state of Louisiana, said his group operates as a service provider for state agencies. That requires a great deal of cost control.
“Parking four petabytes of data in the cloud is not cheap,” he said. “The way we handle it is to establish a service catalog per bid. We own the back end on the IT side, then we look at the business case and determine where that infrastructure should live. The business shouldn’t really care where that server is. They have a business case, not an IT case. We figure out what’s cost effective on the back end. People used to say ‘I’ve got this much money has to be allocated for this project.’ Now it’s turned around for grants. You’re no longer getting money to buy hardware, this is a service. You have to change the way you think now. You say I have this much money for someone to provide a hosting service. We put a rate on that based on the internal IT cost.”
Using multiple clouds will complicate things more. Gorey and Williams said they find the public clouds do different things well, making it necessary to use more than one.
Gorey said Microsoft Azure is way ahead of market leader AWS for high performance computing, and Google has the best AI capabilities. “By 2030, AI and machine learning will be the largest piece of our workload,” he said. “We’re going to need a multi-cloud strategy. We want to give everybody options. People should have choices between where they want to push those workloads.”
Williams said the Louisiana IT team found AWS ran its workloads well and “we saw no reason to go and learn a different stack for every cloud.” But that is changing with advances such as Kubernetes that makes it easier to move workloads.
“If we can get it to where we can move workloads without having to care about where it goes, and security rules follow and we can get it all locked down, that will be the turning point for us where we don’t care what cloud it’s in,”Williams said. “But we haven’t hit that state yet.”
Dell EMC upgrades XC Nutanix-based HCI
Dell chose Nutanix .NEXT 2018 rather than Dell Technologies World the previous week to roll out its new Dell EMC XC appliance based on Nutanix software. Like the Dell EMC VxRail HCI platform that uses VMware vSAN software, XC runs on Dell PowerEdge servers.
The Dell EMC XC940-24 is the first XC quad-core appliance, and can hold 6 TB of memory per appliance. The XC940-24 is available with all flash or as a hybrid with flash and hard disk drives and supports 10-Gigabit Ethernet or 25 GbE networking. Dell positions it as the XC version for running high-performance applications such as in-memory and memory-intensive databases.
The XC940-24 also has a new interface that integrates with the Nutanix Prism software stack. “The interface looks like Prism but Dell designed it,” said John Shirley, director of product management for the Dell EMC XC family.
Dell sees XC as the HCI choice for customers running multiple hypervisors, while VxRail is for VMware-only customers. But the multiple hypervisors of choice for XC customers rarely includes Nutanix AHV, Shirley said. He said Microsoft Hyper-V is the second most common choice behind VMware’s ESX, although Nutanix claims more than 35% of its customers are using AHV.
Dell EMC is not on the list of backup vendors that support AHV. Its Avamar software and Data Domain backup targets are not integrated with AHV. “If we see big traction, we’ll take a look at it,” Shirley said. “It’s not on our roadmap now.”
XC sales contributes to Dell’s No. 1 spot in HCI market share. Dell ranks first in IDC hardware and Dell-owned VMware is tops in HCI software, with Nutanix standing second in both lists. Nutanix CEO Dheeraj Pandey said his company is second due to “funky accounting,” but Dell is chasing world domination in HCI.
“We haven’t been shy about saying we want to be number one in hyper-converged,” Shirley said.
Will HCI SDN go with the Flow?
Potti said Nutanix Flow software-defined networking will converge one more IT role, bringing HCI administration more in line with the cloud.
“You can consolidate the networking admin with the storage admin and the server admin,” he said. “In AWS, there is no networking admin. There is no storage admin. There is no server admin. There’s only the cloud admin or cloud operations or cloud DevOps. So that hyper-convergence of roles has to happen. That’s the core reason we’re doing the network piece.”
Potti said AWS was more of the model for Flow than VMware NSX, which gives SDN to VMware’s vSAN HCI software. “You go to Amazon, provision virtual machines, then suddenly there’s a new policy for security,” he said. “You can put 10 VMs in a security group and the system makes sure they’re segmented.”
With two weeks to go before the compliance deadline for the General Data Protection Regulation, cloud-to-cloud backup vendor OwnBackup is helping its customers prepare for the comprehensive set of rules.
GDPR updates data protection, privacy and access laws across the European Union, and goes into effect on May 25, 2018. It affects not only companies in the EU, but any company that processes data on European Union residents.
The OwnBackup GDPR feature set allows customers to find a data subject’s information within backups.
Built on OwnBackup’s backup and recovery service, the features help customers respond to data subject rights requests, as they apply to personal data within backups and archives, according to the vendor.
The OwnBackup GDPR functionality may be a trend-setter in backup for software as a service (SaaS) applications, such as Salesforce, Slack and ServiceNow, all platforms that the vendor protects. SaaS data is created in the cloud and often needs enhanced protection beyond the basics offered by the applications.
“We’re happy to put a stake in the ground,” said Lee Aber, OwnBackup’s chief information security officer. “We’re trying to move the SaaS backup market forward.”
Key elements of GDPR include:
- The right to be forgotten: Data subjects can request personally identifiable data to be erased from a company’s storage.
- The right to rectification: Data subjects can expect inaccurate personal information to be corrected.
- The right of portability: Data subjects can access personal data that a company has about them and transfer it.
- The right of access: Data subjects can review data that an organization has stored about them.
The OwnBackup GDPR features include:
- Erasure requests, submitted through the OwnBackup application, which support data subjects’ right to be forgotten;
- Rectification requests, submitted through the OwnBackup application, which support data subjects’ right to have their personal data updated;
- Audit logs and notifications, sent to the data controllers’ administrators confirming that an erasure or rectification request has been processed;
- Exporting or transferring a subject’s personal data to support the right of portability;
- The capability to search for data subject information across backups and archives, including within attachments; and
- The ability to set custom data backup expiration dates.
Aber said he has been impressed with how the SaaS community has stepped up to get the word out about GDPR. Salesforce and others have provided guidance and education.
OwnBackup CEO Sam Gutmann said customers have focused on GDPR in the last four to five months.
“Once the regulations are live, I think you’ll see a lot more focus in the U.S.,” Gutmann said.
Gutmann said the OwnBackup GDPR feature set will be live next week. Customers will see a specific tab for GDPR in the administrative console, with its own subset of tools. There is no upcharge for the features, as it’s part of OwnBackup’s core offerings.
OwnBackup plans to add more features, including the ability to apply a group of requests in bulk.
Aber said he thinks interpretation of GDPR will evolve, as will OwnBackup’s approach. Once enforcement begins, Aber suspects the authorities will go after the most egregious rule-breakers.
In general, GDPR provides common sense guidelines around data protection, transparency and privacy that should help organizations.
“It’s not just a compliance obligation,” Aber said. “It actually makes sense.”
Dell Technologies World 2018 has come and gone with hardly a mention of the corporate reorganization Michael Dell and his team is considering as the next step after Dell’s $60 billion-plus 2017 acquisition of storage giant EMC.
Dell has been weighing a return to public ownership as part of strategy to pay down a mountain of debt related to the merger with EMC . A potential reverse merger with its VMware subsidiary is one option Dell is considering, along with a possible spinoff of the Dell EMC storage division as a public company.
But Dell chairman and CEO Michael Dell stuck to technology during his Monday keynote address kicking off the conference. He also sidestepped questions about the issue during a media briefing following his keynote, suggesting any interested parties should read the SEC filing the company made listing its options.
“Have you read our Form 13D (securities filing)?” Dell responded to a question from one reporter. “If you haven’t read it, I suggest you read it in full. It contains everything we’ve said on the matter. We filed it because we publicly said we were thinking about some things. If I say anything else about it, we’ll have to make another filing.”
The coals of Dell’s post-merger considerations are being stoked by roughly $46 billion of debt related to the EMC transaction. Getting a handle on the debt service was an immediate concern when Dell and EMC storage fused into a single company. Analysts and industry observers have predicted from the outset that Dell would have to lop off business segments that no longer fit its long-range goals, including services and software.
In the Jan. 31 filing with the U.S. Securities and Exchange Commission, actually made by publicly traded VMware, Dell Technologies referred to several options under review. One option involves Dell EMC storage pursuing an initial public offering (IPO) of stock. That would mark a sharp departure from the origins of the EMC deal.
Taking Dell EMC storage public likely would net much larger proceeds than the $555 million from Dell subsidiary Pivotal Software’s recent IPO. At the time it acquired EMC, Dell touted the fact the deal would shield the storage business from the scrutiny of Wall Street.
Since the merger, Dell EMC storage revenue has sagged, with Dell Technologies claiming the lion’s share of revenue from sales of traditional servers and networking gear. Would investors get behind an EMC IPO, considering the industrywide decline in externally networked storage sales? That an open question and one undoubtedly getting batted around during Dell’s executive deliberations.
A reverse merger with VMware would shift the Dell EMC debt to the virtualization vendor’s books. That would entail Dell, which owns 81% of VMware, to be the acquired by its much smaller subsidiary. VMware could absorb the debt by selling additional shares to amortize the cost, but shareholders might balk if they perceive such a decision will dilute the value of VMware holdings.
“It’s not a guarantee that activist investors won’t jump in. But there is so much cash being generated by VMware (it could make sense) as a way of shuffling that paper debt around,” said Greg Schulz, senior advisory analyst at consulting firm Server and Storage IO, based in Stillwater, Minn.
Then again, it’s entirely possible that no changes lay ahead for VMware or Dell EMC storage. Listed among the various options on Dell’s securities filing is “maintaining the status quo.” Most industry observers, in fact, say they would be surprised if Dell made any substantive change to its corporate structure.
For now, though, Dell’s flight path toward a decision appears to be stuck in an indefinite holding pattern, searching for a runway and soft landing.
Pivot3, one of the last small independent hyper-converged infrastructure (HCI) vendors left, is on a growth path that its CEO said could lead it to going public within two years.
The vendor report its first quarter 2018 bookings increased more than 60% over last year, with two-thirds of its sales coming from Fortune 1000 companies.
Pivot3 CEO Ron Nash said the early HCI vendor is on pace to hit profitability and complete an initial public offering of stock within 18 to 24 months. He said the 250-employee company that has not raised venture funding in two years will probably not need any more funding until then. Since its 2003 launch, Pivot3 raised $253 million in funding with the last round of $55 million coming in early 2016.
“We substantially reduced our loss over the past three or four quarters,” Nash said. “We’re on our path to break even, and we’re planning to drive to profitability before we IPO. I want us to be ready to be a public company before we do our IPO.”
Nash attributed recent Pivot3 revenue growth to it selling to larger customers, customers running several applications instead of one app, buyers expanding their deployments.
He said Pivot3’s early video surveillance customers are now buying Acuity HCI appliances, and Acuity customers are adding Pivot3 video product. Nash said the larger and multi-app deployments are partly a result of Acuity’s early use of NVMe inside HCI appliances as well as its quality of service.
The claimed Pivot3 revnue growth is in line with —although slightly below — the rest of the HCI market. No industry-wide figures are available yet for the first quarter, but IDC put HCI worldwide revenue growth at 68% in the third quarter of 2017 and 69.4% in the fourth quarter. Pivot3 claimed its revenue increased 50% in the third quarter and 65% in the fourth.
Like many emerging technologies, HCI began with startups before large vendors jumped in. The early players were Nutanix, SimpliVity, Pivot3, Scale Computing, Maxta and a few others. Now Nutanix is a public company, SimpliVity is part of Hewlett Packard Enterprise and the HCI market is dominated by large vendors.
The recent Gartner Magic Quadrant for HCI listed Nutanix, Dell EMC, VMware and HPE as leaders. Pivot3 made the challenger quadrant along with Cisco and Huawei. That puts Pivot3 up against large established vendors in most of its deals.
One of Pivot3’s biggest challenges is what Nash calls “big legacy companies playing big-boy bad behavior.” By that, he means the large vendors who were slow to embrace hyper-convergence will try to raise concerns about smaller companies when they talk to potential customers. If that fails, they will deeply discount their products.
“First, the big boys deny the new technology is better,” Nash said. “Then when the new technology companies start to make breakthroughs, the big boys give a discount. But over the long term, no one has ever protected a market long term by doing that. It only slows the erosion.”
The Unitrends VMware backup product that launched today is going directly after one of the data protection vendor’s major competitors.
VM Backup Essentials (vBE) converges many features the vendor previously offered — including virtual backup software and ransomware detection — into a product specifically targeted at VMware administrators in SMBs, a customer base that Unitrends says Veeam Software is overlooking. Unitrends calls vBE a “Veeam killer.”
Unitrends CTO Mark Jordan said Veeam is leaving behind VMware administrators as it shifts its focus to enterprise customers.
“We are using this to fill a niche,” Jordan said.
Unitrends CEO Paul Brady said he’s heard complaints about Veeam’s increasing prices and lack of attention. The number of customers expressing frustration gained steam in the second half of last year, he said.
George Crump, founder and president of analysis firm Storage Switzerland, said Unitrends is banking on the fact that Veeam has left its core market exposed.
“I don’t know if that’s really happening,” Crump said. “If the market is exposed, I think [Unitrends] could be pretty successful there.”
For 2017, Veeam hit 62% year-over-year growth in enterprise deals and 500% annual growth for deals exceeding $1 million. Veeam’s push for more enterprise business has also been clear in its product updates. The latest version of its flagship data protection product, Veeam Availability Suite, improved management of virtual, physical and cloud data. In addition, through an OEM deal with Cristie Software, the Veeam Availability Platform added physical protection.
Danny Allan, vice president of product strategy at Veeam, said the vendor has nearly 300,000 customers and continues to add 4,000 new ones each month. Veeam claims its first-quarter bookings increased 21% year over year, and expects to reach $1 billion for 2018.
“For smaller companies, we have Veeam Backup Essentials 9.5, an affordably priced solution that meets the needs of SMBs,” Allan wrote in an email. “For large enterprises, we have the enterprise-plus edition of Veeam Availability Suite, which is designed for large, distributed, multi-cloud architectures that scale to multiple petabytes. What’s more, we help power over 3,200 partners’ backup-as-a-service and DR-as-a-service offerings to tens of thousands of Veeam customers.
“That said, if anyone has concerns about our commitment to VMware administrators, we encourage them to ask someone from Veeam about it when they attend their next [VMware User Group]. Because Veeam will be there.”
How will Unitrends vBE measure up?
The Unitrends VMware backup product is a virtual appliance for a hypervisor-level environment only. It manages protection for VMware workloads in public clouds such as Google Cloud Platform and Amazon Web Services, and also integrates with the Unitrends Cloud purpose-built for DR as a service. The product manages VMware data backup, off-site replication and recovery, according to Unitrends.
The ransomware detection was previously included in Unitrends Recovery Series physical appliances and Unitrends Backup virtual appliances. The technology uses predictive analytics to determine the probability that ransomware is operating on a server, workstation or desktop computer. Unitrends alerts customers when it detects ransomware, so they can immediately restore from the last legitimate recovery point.
“It’s a feature set that wins us business every single day,” Jordan said.
The Unitrends VMware backup also includes enterprise-grade global and inline deduplication.
Though there’s nothing that stands out in terms of new technology, Crump said, Unitrends has simplified the interface and removed options that no longer apply because the product is focused on VMware.
Crump said he feels Veeam’s sweet spot is still VMware protection. The product is solid technology-wise, he said, and though it’s become more expensive, he doesn’t see Veeam losing a significant market share.
For Crump, Unitrends’ success comes down to two questions: How well does it execute? And how accurate is its description of the discomfort in the Veeam community?
From a technology standpoint, vBE is focused, and there are going to be customers only on VMware, who now have at least two options.
“In that scenario, I think Unitrends will win a portion of the business,” Crump said, noting the platform’s price and ability to expand. He estimated the target customer is at the upper end of the SMB market.
The product is meant to be used with other vendors’ storage or compute, Jordan said.
The Unitrends VMware backup can protect up to six sockets. It costs $105 per socket.
By the middle of the year, Unitrends plans to add more features to vBE, including complete disaster recovery orchestration, copy data management and analytics.
Crump cautioned against adding too much.
“When you’re trying to go simple, the more features you add to it, the more confusing it becomes,” he said.
The product will also expand to unlimited socket protection, Jordan said.
Kaleao hit an obstacle when trying to convince its ARM-based hyper-converged systems fit well with traditional enterprises. While its features and services might fit enterprises, the original Kaleao box was a terrible fit.
So Kaleao is adding a 4U KMAX-EP system to go with its original KMAX, which is now called KMAX-HD (high density). The redesign is aimed at bringing the startup into more enterprise deals.
U.K.-based Kaleao aimed its KMAX systems at hyperscalers and cloud providers when it launched in early 2017. Those customers didn’t need KMAX to fit into traditional data center racks and “don’t care how many wires run into and out of the box,” said Kaleao chief scientific officer John Goodacre.
“We’re finding that we fit in the enterprise, but it is taking more effort to plug into an enterprise than to plug into a hyperscalers’ data center,” Goodacre said.
Goodacre said Kaleao has “a few customers, not in the hundreds.” The potential audience is much larger for enterprises, and there is also a hunger for hyper-convergence in the enterprise now. But Goodacre said those cutomers want to put their hyper-converged appliances in their traditional server racks.
The KMAX-EP and KMAX-HD each hold 192 eight-core processor sockets and 48 solid-state drives (up to 370 TB of flash), and up to 960 Gigabits per second of Ethernet networking attached. Both use low-power ARM processing. But Kaleao is counting on enterprises finding the 4U chassis and the new design a better fit in their space than the 3U KMAX-HD.
“This is for people who want to replace 20 Dell servers with one of these, but it [KMAX-HDI] doesn’t quite fit as well into their cabinets,” Goodacre said. “The data center and machine rooms for cloud providers and hyperscalers aren’t exactly what you see in enterprises. You probably share the cabinet that you put your KMAX into. Hyperscalers are buying cabinets just for their system.”
With the new hardware design, Kaleao is also now selling both form factors in Server and Appliance versions. The original KMAX was a Server Edition, built on an open-source Linux platform for ARM-based application development and deployment. In the Appliance version, the hardware is managed by OpenStack software and Kaleao’s microsever-based small footprint hypervisor called a Microvisor.