CHICAGO – Veeam unveiled a copy data management feature at its VeeamON 2018 user conference Tuesday as part of its newly branded Hyper-Availability Platform.
Veeam DataLabs, part of the vendor’s flagship Availability Suite, allow organizations to create copies of production environments for uses beyond standard data protection. Those scenarios include test and development, DevOps and DevSecOps.
DataLabs is an element that has evolved over the past year, said Peter McKay, Veeam president and co-CEO.
“It’s a continuation of extending our platform for the enterprise,” McKay said at VeeamON. “Data is becoming more and more important. It’s the lifeblood of companies today.”
An organization can boot virtual machines from a backup or replica into a secure, isolated virtual environment. Veeam DataLabs expand on the functionality of Veeam’s Virtual Labs, enabling production-like instances of virtual environments. The feature replaces Veeam Virtual Labs.
Specifically, they help to ensure teams are developing and testing with the most recent data copies. Developers can spin up instances of the production environment as they design new features. The DataLabs provide sandbox environments to test new patches and updates. The copies also enable testing of security vulnerabilities without disrupting production systems.
In addition, an organization can examine and classify data, which helps comply with regulations such as the General Data Protection Regulation.
More than ‘always on’
Veeam DataLabs is one piece of how the vendor is expanding its data management functionality.
Veeam laid the groundwork at VeeamON 2018 for its strategy and vision of what it’s calling the “Hyper-Availability Platform for Intelligent Data Management,” a shift from its recent “Availability for the Always-On Enterprise” platform.
Hyper-availability is more than availability and being always on; it’s enabling artificial intelligence and being able to move data fluidly across a multi-cloud environment, said Danny Allan, vice president of product strategy.
Veeam executives noted the difficulties that organizations are having with their data management. Not only is the volume of data increasing exponentially, it’s also often on multiple platforms, in multiple locations.
Combining AI and automation will drive the most innovate disruptions of the next decade, Allan said. Veeam is building intelligence into its platform, but didn’t provide specifics Tuesday at VeeamON.
“Leveraging intelligence for better data outcomes is where we need to go as an industry,” Allan said.
Veeam recently reached 300,000 customers and claims to be growing by 4,000 per month.
Nutanix .NEXT 2017 focused on the hyper-converged infrastructure pioneer’s move to the public cloud with Xi Cloud Services. Nutanix .NEXT 2018 last week was a showcase for Nutanix Beam, the vendor’s software-defined networking service. Those launches went in directions most expected the vendor to eventually take hyper-convergence.
Nutanix .NEXT 2019 will likely have an edge computing launch, although it’s not clear that will be the main focus. But the HCI vendor last week unveiled Project Sherlock at Nutanix .NEXT 2018. A Nutanix executive described Sherlock as an enterprise cloud software stack to harness hundreds of zettabytes of data that will be captured by Internet of Things (IoT) devices.
Satyam Vaghani, Nutanix’s VP of IoT and AI, demonstrated Sherlock during one of the .NEXT keynotes. He showed how the technology could allow a retail store to use it to identify customers and automate check out in stores using technologies such as facial recognition. The software could tie into the store’s sensors, cameras, kiosks and rack servers at the edge and connect to the cloud. The demo included a SaaS console to support “planet scale” IoT infrastructure.
“Enterprise IoT is very complicated because of the variety of devices involved,” Vaghani said. “It’s planet scale, not data center scale. Platforms for IoT operations are very different than platforms we are used to running in the data center. Wouldn’t it be great if there was a way for Nutanix to provide a managed service that can be used to manage all of this planet scale infrastructure, all from a central place?”
Vaghani said the technology could allow users to walk out of a story without stopping at a checkout counter to complete a purchase, board an airplane without a boarding pass or wear an ID badge at work. “These are the headliner use cases of IoT, the biggest use case in the edge cloud,” he said.
Nutanix chief product and development officer Sunil Potti said he expects “more concrete announcements” for Sherlock at .NEXT 2019. “Now it’s in what we call the Series A financing mode,” he said.
Considering Xi still isn’t ready for wide-scale consumption a year after its launch, there is a good chance Sherlock won’t make it into next year’s use conference. Like Xi, Sherlock requires a significant amount of engineering work.
“Just like we couldn’t take our current OS and put it in Xi, we had to re-do some core components” for Sherlock, Potti said. “When we extend our stack to the edge, we can’t take our current storage stack. The amount of real time processing needed, the latency requirements, the footprint requirements … everything is different.”
Government IT wrestles with cloud costs, management
Nutanix Beam, a SaaS offering for managing cloud costs, resonated with users at Nutanix .NEXT 2018 who need to get a handle on public cloud costs. Cloud sprawl was a topic during a breakout session with admins from government agencies, a vertical that often has mandates to use public clouds.
Beam provides a dashboard showing all of an organization’s cloud services and monthly pricing. Admins can manage their cloud resources from the dashboard, and Beam recommends cost saving changes.
David Gokey, chief engineer for Mission Systems at NORAD-U.S. Northern Command (USNORCOM), said he saw Beam for the first time at the show and it fit the bill for what he’s been looking for.
“I need a cloud economist,” he said of the task of keeping track of the costs of all his agency’s cloud services. “I need somebody 100 percent dedicated to look at all the products out there and know how we can control that. It has to be automated.”
Derek Williams, director of data center operations for the state of Louisiana, said his group operates as a service provider for state agencies. That requires a great deal of cost control.
“Parking four petabytes of data in the cloud is not cheap,” he said. “The way we handle it is to establish a service catalog per bid. We own the back end on the IT side, then we look at the business case and determine where that infrastructure should live. The business shouldn’t really care where that server is. They have a business case, not an IT case. We figure out what’s cost effective on the back end. People used to say ‘I’ve got this much money has to be allocated for this project.’ Now it’s turned around for grants. You’re no longer getting money to buy hardware, this is a service. You have to change the way you think now. You say I have this much money for someone to provide a hosting service. We put a rate on that based on the internal IT cost.”
Using multiple clouds will complicate things more. Gorey and Williams said they find the public clouds do different things well, making it necessary to use more than one.
Gorey said Microsoft Azure is way ahead of market leader AWS for high performance computing, and Google has the best AI capabilities. “By 2030, AI and machine learning will be the largest piece of our workload,” he said. “We’re going to need a multi-cloud strategy. We want to give everybody options. People should have choices between where they want to push those workloads.”
Williams said the Louisiana IT team found AWS ran its workloads well and “we saw no reason to go and learn a different stack for every cloud.” But that is changing with advances such as Kubernetes that makes it easier to move workloads.
“If we can get it to where we can move workloads without having to care about where it goes, and security rules follow and we can get it all locked down, that will be the turning point for us where we don’t care what cloud it’s in,”Williams said. “But we haven’t hit that state yet.”
Dell EMC upgrades XC Nutanix-based HCI
Dell chose Nutanix .NEXT 2018 rather than Dell Technologies World the previous week to roll out its new Dell EMC XC appliance based on Nutanix software. Like the Dell EMC VxRail HCI platform that uses VMware vSAN software, XC runs on Dell PowerEdge servers.
The Dell EMC XC940-24 is the first XC quad-core appliance, and can hold 6 TB of memory per appliance. The XC940-24 is available with all flash or as a hybrid with flash and hard disk drives and supports 10-Gigabit Ethernet or 25 GbE networking. Dell positions it as the XC version for running high-performance applications such as in-memory and memory-intensive databases.
The XC940-24 also has a new interface that integrates with the Nutanix Prism software stack. “The interface looks like Prism but Dell designed it,” said John Shirley, director of product management for the Dell EMC XC family.
Dell sees XC as the HCI choice for customers running multiple hypervisors, while VxRail is for VMware-only customers. But the multiple hypervisors of choice for XC customers rarely includes Nutanix AHV, Shirley said. He said Microsoft Hyper-V is the second most common choice behind VMware’s ESX, although Nutanix claims more than 35% of its customers are using AHV.
Dell EMC is not on the list of backup vendors that support AHV. Its Avamar software and Data Domain backup targets are not integrated with AHV. “If we see big traction, we’ll take a look at it,” Shirley said. “It’s not on our roadmap now.”
XC sales contributes to Dell’s No. 1 spot in HCI market share. Dell ranks first in IDC hardware and Dell-owned VMware is tops in HCI software, with Nutanix standing second in both lists. Nutanix CEO Dheeraj Pandey said his company is second due to “funky accounting,” but Dell is chasing world domination in HCI.
“We haven’t been shy about saying we want to be number one in hyper-converged,” Shirley said.
Will HCI SDN go with the Flow?
Potti said Nutanix Flow software-defined networking will converge one more IT role, bringing HCI administration more in line with the cloud.
“You can consolidate the networking admin with the storage admin and the server admin,” he said. “In AWS, there is no networking admin. There is no storage admin. There is no server admin. There’s only the cloud admin or cloud operations or cloud DevOps. So that hyper-convergence of roles has to happen. That’s the core reason we’re doing the network piece.”
Potti said AWS was more of the model for Flow than VMware NSX, which gives SDN to VMware’s vSAN HCI software. “You go to Amazon, provision virtual machines, then suddenly there’s a new policy for security,” he said. “You can put 10 VMs in a security group and the system makes sure they’re segmented.”
With two weeks to go before the compliance deadline for the General Data Protection Regulation, cloud-to-cloud backup vendor OwnBackup is helping its customers prepare for the comprehensive set of rules.
GDPR updates data protection, privacy and access laws across the European Union, and goes into effect on May 25, 2018. It affects not only companies in the EU, but any company that processes data on European Union residents.
The OwnBackup GDPR feature set allows customers to find a data subject’s information within backups.
Built on OwnBackup’s backup and recovery service, the features help customers respond to data subject rights requests, as they apply to personal data within backups and archives, according to the vendor.
The OwnBackup GDPR functionality may be a trend-setter in backup for software as a service (SaaS) applications, such as Salesforce, Slack and ServiceNow, all platforms that the vendor protects. SaaS data is created in the cloud and often needs enhanced protection beyond the basics offered by the applications.
“We’re happy to put a stake in the ground,” said Lee Aber, OwnBackup’s chief information security officer. “We’re trying to move the SaaS backup market forward.”
Key elements of GDPR include:
- The right to be forgotten: Data subjects can request personally identifiable data to be erased from a company’s storage.
- The right to rectification: Data subjects can expect inaccurate personal information to be corrected.
- The right of portability: Data subjects can access personal data that a company has about them and transfer it.
- The right of access: Data subjects can review data that an organization has stored about them.
The OwnBackup GDPR features include:
- Erasure requests, submitted through the OwnBackup application, which support data subjects’ right to be forgotten;
- Rectification requests, submitted through the OwnBackup application, which support data subjects’ right to have their personal data updated;
- Audit logs and notifications, sent to the data controllers’ administrators confirming that an erasure or rectification request has been processed;
- Exporting or transferring a subject’s personal data to support the right of portability;
- The capability to search for data subject information across backups and archives, including within attachments; and
- The ability to set custom data backup expiration dates.
Aber said he has been impressed with how the SaaS community has stepped up to get the word out about GDPR. Salesforce and others have provided guidance and education.
OwnBackup CEO Sam Gutmann said customers have focused on GDPR in the last four to five months.
“Once the regulations are live, I think you’ll see a lot more focus in the U.S.,” Gutmann said.
Gutmann said the OwnBackup GDPR feature set will be live next week. Customers will see a specific tab for GDPR in the administrative console, with its own subset of tools. There is no upcharge for the features, as it’s part of OwnBackup’s core offerings.
OwnBackup plans to add more features, including the ability to apply a group of requests in bulk.
Aber said he thinks interpretation of GDPR will evolve, as will OwnBackup’s approach. Once enforcement begins, Aber suspects the authorities will go after the most egregious rule-breakers.
In general, GDPR provides common sense guidelines around data protection, transparency and privacy that should help organizations.
“It’s not just a compliance obligation,” Aber said. “It actually makes sense.”
Dell Technologies World 2018 has come and gone with hardly a mention of the corporate reorganization Michael Dell and his team is considering as the next step after Dell’s $60 billion-plus 2017 acquisition of storage giant EMC.
Dell has been weighing a return to public ownership as part of strategy to pay down a mountain of debt related to the merger with EMC . A potential reverse merger with its VMware subsidiary is one option Dell is considering, along with a possible spinoff of the Dell EMC storage division as a public company.
But Dell chairman and CEO Michael Dell stuck to technology during his Monday keynote address kicking off the conference. He also sidestepped questions about the issue during a media briefing following his keynote, suggesting any interested parties should read the SEC filing the company made listing its options.
“Have you read our Form 13D (securities filing)?” Dell responded to a question from one reporter. “If you haven’t read it, I suggest you read it in full. It contains everything we’ve said on the matter. We filed it because we publicly said we were thinking about some things. If I say anything else about it, we’ll have to make another filing.”
The coals of Dell’s post-merger considerations are being stoked by roughly $46 billion of debt related to the EMC transaction. Getting a handle on the debt service was an immediate concern when Dell and EMC storage fused into a single company. Analysts and industry observers have predicted from the outset that Dell would have to lop off business segments that no longer fit its long-range goals, including services and software.
In the Jan. 31 filing with the U.S. Securities and Exchange Commission, actually made by publicly traded VMware, Dell Technologies referred to several options under review. One option involves Dell EMC storage pursuing an initial public offering (IPO) of stock. That would mark a sharp departure from the origins of the EMC deal.
Taking Dell EMC storage public likely would net much larger proceeds than the $555 million from Dell subsidiary Pivotal Software’s recent IPO. At the time it acquired EMC, Dell touted the fact the deal would shield the storage business from the scrutiny of Wall Street.
Since the merger, Dell EMC storage revenue has sagged, with Dell Technologies claiming the lion’s share of revenue from sales of traditional servers and networking gear. Would investors get behind an EMC IPO, considering the industrywide decline in externally networked storage sales? That an open question and one undoubtedly getting batted around during Dell’s executive deliberations.
A reverse merger with VMware would shift the Dell EMC debt to the virtualization vendor’s books. That would entail Dell, which owns 81% of VMware, to be the acquired by its much smaller subsidiary. VMware could absorb the debt by selling additional shares to amortize the cost, but shareholders might balk if they perceive such a decision will dilute the value of VMware holdings.
“It’s not a guarantee that activist investors won’t jump in. But there is so much cash being generated by VMware (it could make sense) as a way of shuffling that paper debt around,” said Greg Schulz, senior advisory analyst at consulting firm Server and Storage IO, based in Stillwater, Minn.
Then again, it’s entirely possible that no changes lay ahead for VMware or Dell EMC storage. Listed among the various options on Dell’s securities filing is “maintaining the status quo.” Most industry observers, in fact, say they would be surprised if Dell made any substantive change to its corporate structure.
For now, though, Dell’s flight path toward a decision appears to be stuck in an indefinite holding pattern, searching for a runway and soft landing.
Pivot3, one of the last small independent hyper-converged infrastructure (HCI) vendors left, is on a growth path that its CEO said could lead it to going public within two years.
The vendor report its first quarter 2018 bookings increased more than 60% over last year, with two-thirds of its sales coming from Fortune 1000 companies.
Pivot3 CEO Ron Nash said the early HCI vendor is on pace to hit profitability and complete an initial public offering of stock within 18 to 24 months. He said the 250-employee company that has not raised venture funding in two years will probably not need any more funding until then. Since its 2003 launch, Pivot3 raised $253 million in funding with the last round of $55 million coming in early 2016.
“We substantially reduced our loss over the past three or four quarters,” Nash said. “We’re on our path to break even, and we’re planning to drive to profitability before we IPO. I want us to be ready to be a public company before we do our IPO.”
Nash attributed recent Pivot3 revenue growth to it selling to larger customers, customers running several applications instead of one app, buyers expanding their deployments.
He said Pivot3’s early video surveillance customers are now buying Acuity HCI appliances, and Acuity customers are adding Pivot3 video product. Nash said the larger and multi-app deployments are partly a result of Acuity’s early use of NVMe inside HCI appliances as well as its quality of service.
The claimed Pivot3 revnue growth is in line with —although slightly below — the rest of the HCI market. No industry-wide figures are available yet for the first quarter, but IDC put HCI worldwide revenue growth at 68% in the third quarter of 2017 and 69.4% in the fourth quarter. Pivot3 claimed its revenue increased 50% in the third quarter and 65% in the fourth.
Like many emerging technologies, HCI began with startups before large vendors jumped in. The early players were Nutanix, SimpliVity, Pivot3, Scale Computing, Maxta and a few others. Now Nutanix is a public company, SimpliVity is part of Hewlett Packard Enterprise and the HCI market is dominated by large vendors.
The recent Gartner Magic Quadrant for HCI listed Nutanix, Dell EMC, VMware and HPE as leaders. Pivot3 made the challenger quadrant along with Cisco and Huawei. That puts Pivot3 up against large established vendors in most of its deals.
One of Pivot3’s biggest challenges is what Nash calls “big legacy companies playing big-boy bad behavior.” By that, he means the large vendors who were slow to embrace hyper-convergence will try to raise concerns about smaller companies when they talk to potential customers. If that fails, they will deeply discount their products.
“First, the big boys deny the new technology is better,” Nash said. “Then when the new technology companies start to make breakthroughs, the big boys give a discount. But over the long term, no one has ever protected a market long term by doing that. It only slows the erosion.”
The Unitrends VMware backup product that launched today is going directly after one of the data protection vendor’s major competitors.
VM Backup Essentials (vBE) converges many features the vendor previously offered — including virtual backup software and ransomware detection — into a product specifically targeted at VMware administrators in SMBs, a customer base that Unitrends says Veeam Software is overlooking. Unitrends calls vBE a “Veeam killer.”
Unitrends CTO Mark Jordan said Veeam is leaving behind VMware administrators as it shifts its focus to enterprise customers.
“We are using this to fill a niche,” Jordan said.
Unitrends CEO Paul Brady said he’s heard complaints about Veeam’s increasing prices and lack of attention. The number of customers expressing frustration gained steam in the second half of last year, he said.
George Crump, founder and president of analysis firm Storage Switzerland, said Unitrends is banking on the fact that Veeam has left its core market exposed.
“I don’t know if that’s really happening,” Crump said. “If the market is exposed, I think [Unitrends] could be pretty successful there.”
For 2017, Veeam hit 62% year-over-year growth in enterprise deals and 500% annual growth for deals exceeding $1 million. Veeam’s push for more enterprise business has also been clear in its product updates. The latest version of its flagship data protection product, Veeam Availability Suite, improved management of virtual, physical and cloud data. In addition, through an OEM deal with Cristie Software, the Veeam Availability Platform added physical protection.
Danny Allan, vice president of product strategy at Veeam, said the vendor has nearly 300,000 customers and continues to add 4,000 new ones each month. Veeam claims its first-quarter bookings increased 21% year over year, and expects to reach $1 billion for 2018.
“For smaller companies, we have Veeam Backup Essentials 9.5, an affordably priced solution that meets the needs of SMBs,” Allan wrote in an email. “For large enterprises, we have the enterprise-plus edition of Veeam Availability Suite, which is designed for large, distributed, multi-cloud architectures that scale to multiple petabytes. What’s more, we help power over 3,200 partners’ backup-as-a-service and DR-as-a-service offerings to tens of thousands of Veeam customers.
“That said, if anyone has concerns about our commitment to VMware administrators, we encourage them to ask someone from Veeam about it when they attend their next [VMware User Group]. Because Veeam will be there.”
How will Unitrends vBE measure up?
The Unitrends VMware backup product is a virtual appliance for a hypervisor-level environment only. It manages protection for VMware workloads in public clouds such as Google Cloud Platform and Amazon Web Services, and also integrates with the Unitrends Cloud purpose-built for DR as a service. The product manages VMware data backup, off-site replication and recovery, according to Unitrends.
The ransomware detection was previously included in Unitrends Recovery Series physical appliances and Unitrends Backup virtual appliances. The technology uses predictive analytics to determine the probability that ransomware is operating on a server, workstation or desktop computer. Unitrends alerts customers when it detects ransomware, so they can immediately restore from the last legitimate recovery point.
“It’s a feature set that wins us business every single day,” Jordan said.
The Unitrends VMware backup also includes enterprise-grade global and inline deduplication.
Though there’s nothing that stands out in terms of new technology, Crump said, Unitrends has simplified the interface and removed options that no longer apply because the product is focused on VMware.
Crump said he feels Veeam’s sweet spot is still VMware protection. The product is solid technology-wise, he said, and though it’s become more expensive, he doesn’t see Veeam losing a significant market share.
For Crump, Unitrends’ success comes down to two questions: How well does it execute? And how accurate is its description of the discomfort in the Veeam community?
From a technology standpoint, vBE is focused, and there are going to be customers only on VMware, who now have at least two options.
“In that scenario, I think Unitrends will win a portion of the business,” Crump said, noting the platform’s price and ability to expand. He estimated the target customer is at the upper end of the SMB market.
The product is meant to be used with other vendors’ storage or compute, Jordan said.
The Unitrends VMware backup can protect up to six sockets. It costs $105 per socket.
By the middle of the year, Unitrends plans to add more features to vBE, including complete disaster recovery orchestration, copy data management and analytics.
Crump cautioned against adding too much.
“When you’re trying to go simple, the more features you add to it, the more confusing it becomes,” he said.
The product will also expand to unlimited socket protection, Jordan said.
Kaleao hit an obstacle when trying to convince its ARM-based hyper-converged systems fit well with traditional enterprises. While its features and services might fit enterprises, the original Kaleao box was a terrible fit.
So Kaleao is adding a 4U KMAX-EP system to go with its original KMAX, which is now called KMAX-HD (high density). The redesign is aimed at bringing the startup into more enterprise deals.
U.K.-based Kaleao aimed its KMAX systems at hyperscalers and cloud providers when it launched in early 2017. Those customers didn’t need KMAX to fit into traditional data center racks and “don’t care how many wires run into and out of the box,” said Kaleao chief scientific officer John Goodacre.
“We’re finding that we fit in the enterprise, but it is taking more effort to plug into an enterprise than to plug into a hyperscalers’ data center,” Goodacre said.
Goodacre said Kaleao has “a few customers, not in the hundreds.” The potential audience is much larger for enterprises, and there is also a hunger for hyper-convergence in the enterprise now. But Goodacre said those cutomers want to put their hyper-converged appliances in their traditional server racks.
The KMAX-EP and KMAX-HD each hold 192 eight-core processor sockets and 48 solid-state drives (up to 370 TB of flash), and up to 960 Gigabits per second of Ethernet networking attached. Both use low-power ARM processing. But Kaleao is counting on enterprises finding the 4U chassis and the new design a better fit in their space than the 3U KMAX-HD.
“This is for people who want to replace 20 Dell servers with one of these, but it [KMAX-HDI] doesn’t quite fit as well into their cabinets,” Goodacre said. “The data center and machine rooms for cloud providers and hyperscalers aren’t exactly what you see in enterprises. You probably share the cabinet that you put your KMAX into. Hyperscalers are buying cabinets just for their system.”
With the new hardware design, Kaleao is also now selling both form factors in Server and Appliance versions. The original KMAX was a Server Edition, built on an open-source Linux platform for ARM-based application development and deployment. In the Appliance version, the hardware is managed by OpenStack software and Kaleao’s microsever-based small footprint hypervisor called a Microvisor.
Four months after the release of Veeam’s Universal Storage API, Veeam and Pure Storage have teamed up to streamline data protection and storage, and speed recovery.
The integration — initiated through the API — between the Veeam Availability Platform and the Pure Storage FlashArray allows joint customers to lessen the impact on production environments by using Pure snapshots, and back up more often to reduce recovery point objectives. Also through the Veeam and Pure Storage integration, customers can create snapshot-only jobs to add more frequent recovery points.
With Veeam Explorer for Storage Snapshots, customers recover individual items or entire virtual machines directly from Pure Storage snapshots for faster recovery time objectives.
Veeam and Pure Storage customers can also use Pure snapshots to create an on-demand isolated test environment to use production workload copies for testing and development, analytics and security.
“It’s really a ‘better together’ story,” especially for virtualized environments, said Ken Ringdahl, vice president of global alliance architecture at Veeam.
In addition, joint customers can use Pure Storage FlashBlade, a scale-out storage system for unstructured data, as a Veeam Ready Repository. That enables customers to use all Veeam software features.
A Pure partnership with Veeam
The API was released as part of the Veeam Availability Suite 9.5 Update 3 in December.
Pure Storage, like Veeam, is investing in the enterprise, said Carey Stanton, vice president of global business and corporate development at Veeam.
In the past year, Veeam has released a number of product updates targeted toward enterprise customers. For example, Veeam Availability Suite 9.5 Update 3 includes a central console to manage backup and recovery across virtual, physical and cloud workloads. Veeam also added physical support through an OEM deal with Cristie Software.
At the end of 2017, Veeam closed its acquisition of N2WS, which extended its enterprise and cloud data protection reach.
Veeam and Pure Storage have had a partnership for years, with features such as snapshot integration. Michael Sotnick, vice president of global channels and alliances at Pure, said that more integration will follow.
“It’s only the first inning in this game,” Stanton said.
The integration, available for download now, is a storage plug-in to the Veeam Availability Platform.
After four straight quarters of growth, IBM storage hardware revenue crashed in the first quarter of 2018. IBM reported a 15% decline in storage hardware revenue, a drop that dragged the entire company’s overall revenue to fall short of expectations.
On IBM’s earnings call Tuesday night, CFO Jim Kavanaugh blamed the storage turnaround on increased competition, pricing pressures and “some sales execution challenges.” He did say he expects a strong IBM storage product portfolio and new launches in late 2018 to fuel a rebound.
And Kavanaugh emphasized the declines were only on the IBM storage hardware side. He said IBM’s software-defined and cloud object storage revenue increased, but those are attributed to other IBM segments.
Kavanaugh described the storage market as “aggressive,” but said he has confidence in the IBM storage team.
“We were disappointed in our storage performance and it contributed to a modest shortfall to our own expectations of IBM’s revenue growth in the quarter,” Kavanaugh said.
“But I’ll tell you we have a great team with a proven track record. This is the same team that has proven that they’ve revitalized the portfolio in the past, took market share for four consecutive quarters. We have all the confidence in the world that we can get our storage business back to where it needs to be as we move forward in the second half of 2018.”
IBM attributed its 2017 rebound to surges in flash and cloud storage. According to IT research firm IDC, IBM gained share in external (networked) storage in every quarter of 2017. For the full year, IBM increased its revenue to $2.184 billion from $2.073 billion in 2016 while the overall external storage systems market was flat. IBM increased its share from 9% in 2016 to 9.6% in 2017. But unless the overall storage market tanked, IBM dropped share in the first quarter of 2018.
With a fresh $60 million in funding, Scality CEO Jerome Lecat said he expects the object storage vendor to become profitable within two years.
Scality, which launched its Ring object storage in 2010 and added a distributed file system five years later, has now raised $152 million over five funding rounds. Its previous largest round before today brought in $57 million in August of 2015.
The vendor relies largely on partnerships with Hewlett Packard Enterprise and Cisco to drive sales, so Lecat said most of the new funding will go into product development. Lecat said the new Zenko multi-cloud management software will be a main focus but he expects more product launches in 2018.
“We intend to show we can run a profit,” Lecat said. “We intend to be a break-even profitable company by 2020. From there, who knows where we go? We can raise more money and make acquisitions, but that’s not a decision I’ll make now. Sixty million dollars is a lot of money, and our goal is to strengthen our product strategy.”
Scality, which has more than 200 employees, finished 2017 with nearly 90 people on its engineering team. Lecat said that number will grow with the funding.
The 2017 Gartner Magic Quadrant for Distributed File Systems and Object Storage lists Scalilty as a leader, along with Dell EMC and IBM.
Zenko, based on open-source software, stores data and applications under one interface regardless of whether they reside on-premises or in clouds. Scality uses the Amazon S3 API that works with RING or other back-end cloud storage. Scality released Zenko code to the open source community in 2017 and is preparing to make a commercially supported version available in 2018. Zenko can complement Scality Ring or run as a standalone product.
Lecat said he expects Zenko to go GA within a few months. By then, he expects beta customer Bloomberg Media to put Zenko in production.
Multi-cloud management is important for Scality, whose customers include service providers and organizations with large data stores. Lecat said the average Scality Ring implementation is for 4 PB of data, often spread across clouds.
“The future of enterprise IT and storage is multi-cloud,” he said. “What we mean by multi-cloud is large enterprises will continue to operate some storage by themselves in a private cloud and also leverage several public cloud services. A few years ago, people said everything would go in Amazon, but we’re not seeing that trend. We see people also want [Microsoft] Azure, Google and possibly other public clouds.”
As a private company, Scality does not disclose revenue but Lecat said the vendor added 51 enterprise customers in 2017. He put its total customer count at more than 200.
The funding round included new investor Harbert European Growth Capital, with previous Scality investors Menlo Ventures, Iris Capital, Idinvest Partners and Galileo Partners participating.