NetApp is the first storage vendor to sell Toshiba’s new highly secure solid-state drives (SSDs).
NetApp hybrid FAS and E-Series arrays will include Toshiba PXO4S 12 Gb per second SAS SSDs, built with Federal Information Processing Standard (FIPS) 140-2 Level 2 encryption.
The self-encrypting SSDs have sequential read and write speeds of up to 1,900 MiBps and 1,100 MiBps.
“Most SAS SSDs have non-encryption versions and encrypted versions. This is different because these SSDs went through a FIPS certification process, which is the highest government-level security,” said Cameron Brett, a director of SSD product marketing at Toshiba’s storage products business unit. “It’s the highest level of security for SAS SSDs because it’s certified by the government. It’s not an easy process. It’s very time-consuming.”
NetApp has shipped the SSDs in arrays since last month.
FIPS is a set of standards that describe document processing, encryption algorithms and other information technology processes for use within non-military federal government agencies and contractors who work with those agencies.
“The FIPS certification is very specific to the components and the firmware,” Brett said. “The drive itself cannot be breached.”
“This is an important step for NetApp to make this part of their campaign,” Coughlin said. “They are making it a key part of their offerings and that is the first time I’ve seen that. And this is not software encryption. This is encryption that is built into the storage devices. When it’s built into the drives, the key is more secure.”
Coughlin said software-based encryption carries a large overhead tax for encrypting and decrypting the data.
“Software-based encryption puts overhead on the system whereas drives that are encrypted don’t have that overhead,” he said. “With software encryption, you have to rewrite the data with the encryption or change the keys. With hardware encryption, the data on the drives is always encrypted and the keys never leave the drives. All you need is good, strong passwords.”
Commvault’s push to store customer data in the cloud is paying off.
CEO Bob Hammer said the amount of data stored in the cloud with Commvault software has more than doubled since the start of 2016. That helped the data protection and management software vendor to better revenue growth than expected last quarter.
Commvault Tuesday reported revenue of $159.3 million last quarter, a 13% increase over the previous year. Its software revenue of $70.5 million increased 22% year-over-year. Overall and software revenue both beat analysts’ expectations.
Commvault lost $800,00 in the quarter, down from $4.6 million a year ago. Hammer said he expects continued growth of the cloud and other standalone products will accelerate revenue increases over the next few quarters, and that should lead to consistent profits.
Commvault addresses the cloud with standalone products such as its Edge Drive file sharing product, cloud replication and DR packages and through partnerships with Microsoft and Amazon. The Edge Drive is part of the standalone applications that can plug into the Commvault Data Platform, along with apps for virtual machine backup, end point protection, archiving and others data management use cases.
Moving customer data to the cloud was a focus on the Commvault Go user conference this month.
“This represents a significant driver of our software growth,” Hammer said of the cloud during Commvault’s earnings call. “I can tell you the cloud is a significant, material part of our revenue and of our revenue growth.
Commvault said revenue from enterprise deals (more than $100,000) increased 44% from last year and made up 57% of its revenue in the quarter. The number of enterprise deals increased 45% from last year, averaging $268,000 per deal.
“At the 100,000-foot level, the major driver of our business is large enterprises as they move to the cloud,” Hammer said. “We’re helping them manage data from on-premise to the cloud, manage data in the cloud and then help them manage data in these hybrid environments.”
Hewlett Packard Enterprise launched its high-end enterprise storage array upgade today at the same time as Hitachi Data Systems. The HPE XP7 and HDS Virtual Storage Platform (VSP) use the same underlying hardware and software, supplied by HDS’ Japanese parent company Hitachi Ltd.
HPE has licensed the Hitachi technology for 15 years, for customers who want the highest availability mainframe storage. The XP7’s mainframe support and ability to virtualize any hardware array on the back end distinguish the platform from HPE’s flagship 3PAR StoreServ platform.
“The XP stands for advanced replication, mission critical RAS [resiliency, availability, serviceability], and 100 percent access to data and applications, regardless of any hardware or site failure,” said Vish Mulchand, senior director of product management for HPE storage.
You can read more about the XP7 speeds and feeds in this story on the HDS VSP upgrade. As with HDS, HPE calls out the platform’s all-flash options and data reduction technologies as key enhancements. While HPE adds its own availability software to the XP7 platform, the inline and post-process data reduction comes from the HDS flash module drives (7 TB and 14 TB), ASICs inside the FMDs and optimized software. “For this class of storage, customers want configuration flexibility,” Mulchand said. “We have customers using all-flash XP7s and hybrid XP7s”
Unlike HDS, HPE provides pricing info for its new platform. Street pricing starts at $20,800 for the new controllers and $22,200 for FMDs. HPE claims an all-flash configuration can cost $1.20 per GB with 4-1 data reduction. Software-based compression and deduplication costs $11,600 per XP7 frame.
FalconStor Software took aim at hybrid cloud deployments with a new pricing model and product upgrade for its FreeStor storage virtualization and block-based data services.
The Melville, NY-based software vendor now charges customers only for the primary copy of data – not the total storage capacity under management – with its subscription-based pricing model. The FreeStor software provides common tools, single-pane management and block-based services such as data migration, protection, recovery, and analytics for use with heterogeneous storage.
FalconStor CEO Gary Quinn estimated that 70% of FreeStor’s customers are managed service providers (MSPs). He said providers offer services such as backup or disaster recovery (DR) and want the ability to store an additional copy of their customers’ data in public clouds such as Amazon Web Services (AWS) or Microsoft Azure.
Quinn said FalconStor’s enterprise customers have also been asking for similar options to move to AWS or Azure for virtual backup, DR and test and development use cases.
“It doesn’t really cost me anything to make a copy of the data or replicate the data to another location and manage it through the FreeStor management server. So our view is that customers should pay once,” Quinn said.
He said the list price for the FreeStor software, inclusive of data services, is three cents per GB per month to use on the primary data copy. The customer supplies the hardware.
Eric Burgener, a research director at International Data Corp., said he has seen pay-as-you-go models from other vendors but nothing like FalconStor’s aggressive pricing.
FalconStor changed the pricing model in anticipation of a new version of its FreeStor software, which extends support to public clouds. FalconStor added support for Amazon, Microsoft, Alibaba, Huawei and Oracle to go with its prior support of OpenStack-based deployments.
Tim Sheets, vice president of marketing at FalconStor, said, in an Amazon environment, the FreeStor Virtual Appliance (FSS VA) would run on the AWS Elastic Compute Cloud (EC2). The FSS VA could either use Amazon’s Elastic Block Store (EBS) or present block services through AWS Storage Gateway (ASG) to load into Amazon’s object-based Simple Storage Service (S3) container, he said.
“You don’t have to go learn a new set of tools from Amazon if you haven’t done it before. We’ve already got the configuration set up to really simplify it for those customers,” he said. “And you also get the analytics, all the insights, through a single pane of glass with the FreeStor management server that you wouldn’t get if you had to use the Amazon or an Azure gateway,” Sheets said.
Customers could also use FreeStor to manage data across multiple supported public clouds or to move data from one public cloud to another, so long as the FSS VA runs in each cloud.
“I’m sure that Amazon’s not going to provide tools to leave Amazon and go to Azure,” Quinn said. “That’s what we’re doing here, the same way as if you wanted to move from EMC to HP on disk or EMC to Pure on flash. It’s just being done in the cloud.”
The new FreeStor software also beefs up external security with support for the Lightweight Directory Access Protocol (LDAP) and Microsoft Active Directory for authentication, authorization, and auditing.
Other newly supported features include enhanced analytics to enable core-to-edge visibility down to the applications and service-level agreement (SLA) management, improved support for NVMe to boost performance and lower latency, and Linux 7 compliance.
The FreeStor updates arrive as FalconStor battles financial woes. FalconStor reported $8.1 million in revenue for the second quarter, down from $9.6 million in Q2 of 2015, with only $9.4 in cash on hand. But Quinn said at the time that FalconStor was making solid progress selling FreeStor subscriptions to MSPs, enterprises and OEMs.
EMC’s XtremIO all-flash SAN is getting a file-system injection thanks to Dell Fluid File System (FluidFS).
Dell EMC previewed the NAS capabilities for XtremIO at Dell EMC World, saying they would be generally available by late 2017. FluidFS is a scale-out NAS technology that Dell acquired from Exanet in 2010 and used to add file capabilities to its Compellent and EqualLogic SAN arrays. But even before Dell acquired EMC for more than $60 billion, the development teams from XtremIO and FluidFS – both based in Israel – were collaborating on their integration.
Chris Ratcliffe, Dell EMC senior vice president of core technologies, jokingly referred to the joint development as a “black ops” operation. The integration will add NFS, SMB, Hadoop Distributed File System (HDFS) and NDMP to XtremIO’s current Fibre Chanel and iSCSI block storage support.
As when Dell added FluidFS to Compellent and EqualLogic, XtremIO will require a separate piece of hardware to deliver file services. XtremIO CTO Itzik Reich called the appliance an extension to XtremIO rather than a full gateway, and said the XtremIO approach will not impact performance. He also said file storage will be managed through the same interface as block storage with “the same look and feel.”
Reich said the original design goal for ExtremIO included adding data services in later iterations. “What’s in the market today is just the beginning,” he said of the product that EMC claims has more than 3,000 customers and $3 billion in revenue in three years on the market. He also said there will be a lot more added to the next-generation XtremIO, including more drives, higher capacity SSDs and software-defined storage capabilities.
“We were looking for ways to complement our scale-out architecture,” he said. “We wanted it to be more than just Fibre Channel. When we heard talk of a partnership, I gave Michael (Dell) a call and said this is a good project for us to add file services.”
Dell EMC this week announced plans to deliver an all-flash version of its Isilon scale-out NAS platform in 2017. Isilon is aimed at traditional scale-out NAS use cases such as media/entertainment, life sciences and Hadoop analytics. Ratcliffe said XtremIO’s NAS would be more for traditional SAN customers. “This is scale-out NAS for transactional environments that require sub-millisecond response times,” he said.
Reich estimated it would have taken at least five years to build file services from scratch into XtremIO. His team looked at filesy stems from EMC’s Unity unified and Isilon scale-out NAS but determined FluidFS fit better with XtremIO’s architecture.
“Unity doesn’t scale out,” he said. “Isilon scales out like nobody’s business, but it doesn’t provide the latency we need.”
EMC’s Unity, Isilon and VMAX All-Flash arrays already support 15 TB SSDs, but they won’t be available on XtremIO until the next generation. Reich said his team wants to make sure using the higher capacity drives will not impact performance. “People don’t realize, the larger the drive capacity gets, the worse the performance gets,” he said. “We are not willing to sacrifice our predictable performance.”
IBM marked the one-year anniversary of its Cleversafe acquisition with the launch of a “pay-as-you-go” cloud object storage service enabling customers to use the same technology on site and off premises.
IBM foreshadowed its plans to facilitate hybrid cloud deployments on Oct. 5, 2015, when it acquired Cleversafe. But until this month, IBM made available the Cleversafe object storage software for use only on-premises or in a dedicated environment in the IBM Cloud.
Russ Kennedy, vice president of product strategy and customer success at IBM, said IBM has done considerable work to extend its public cloud’s previously limited multi-tenancy capabilities to support millions of concurrent tenants and to integrate the core Cleversafe technology.
Kennedy said customers have the flexibility to store application data in the cloud and move it back on premises, or vice versa, if they choose. He said IBM is looking to provide more automation capabilities in the future, “where decisions are made based on utilization or access or certain parameters that may drive the workloads in one direction or another.”
IBM Cloud Object Storage services are now available in the U.S. and Europe in three configurations:
–Standard – Cleversafe-based high-performance offering for active workloads; supports object storage application programming interfaces (APIs) such as Amazon S3 and OpenStack Swift.
–Vault – lower-cost offering that targets archive, backup and other workloads where data is infrequently accessed.
–Dedicated – single-tenant IBM Object Storage running on dedicated servers in IBM Cloud data centers; available as an IBM managed service or a self-managed option.
Kennedy said SecureSlice technology from Cleversafe eliminates the need for customers to manage encryption keys. SecureSlice automatically encrypts each data segment before it is erasure coded and distributed. IBM Cloud Accesser technology can reassemble the data at the customer’s primary data center, and SecureSlice decrypts it.
IBM Cloud Object Storage has regional and cross-regional options. The cross-regional service sends sliced data to at least three geographic regions. The regional service stores data in multiple data centers in a specific region.
Kennedy said IBM operates close to 50 data centers worldwide, including 12 to 15 in North America. IBM Cloud Object Storage is due to become available in the Asia-Pacific region by year’s end, with other locations to follow in 2017, according to Kennedy.
IBM Cloud Object Storage pricing, found at this link, is based on per GB per month basis. There are also fees for transactions. IBM’s on-premise object storage software can be licensed based on capacity or through a subscription model.
Scott Sinclair, a senior analyst at Enterprise Strategy Group (ESG) Inc., said a 2016 ESG poll of current enterprise Amazon Web Services (AWS) customers identified Microsoft Azure and IBM as the most viable competitors to AWS.
Sinclair said using the same object storage software on premises and off premises could provide advantages. He said storage vendors often differ in how they implement protocols, so users might have piece of mind with the same technology in both places. He said they also know what to expect for service and support, working with a partner that understands both their on-premises and off-premises needs.
“The more vendors that you have to manage in your IT organization creates work,” Sinclair said. “And that work requires people.”
Kennedy said the exponential growth of information is driving users to recognize the cost, scalability and management benefits that object storage can provide over traditional storage, especially when they need petabytes or exabytes of capacity.
“There are still headwinds for object storage,” he said. “Not all the applications in the world have the ability to write to object storage like they do to traditional file-based or block-based storage. But that’s changing. And it’s changing quite rapidly with the popularity of moving to the cloud.”
Magneto-resistive RAM chipmaker Everspin Technologies is trying to make sure its in-memory magnetic storage technology will spin into the future. The Chandler, Ariz.-based vendor this month raised $40 million in an initial public offering needed to “continue as a going concern” beyond 2016, according to its security filing.
Everspin shares started trading on the Nasdaq Global Market Oct. 7 under ticker symbol MRAM. Investors purchased 5 million shares at $8 per share. Underwriters retain an option to scoop up an over-allotment of 750,000 shares at the offering price, which would increase proceeds by roughly $6 million to $46 million.
In a concurrent transaction, Everspin said it expects to get an additional $5 million via a private placement of 625,000 shares with China-based NOR memory maker GigaDevice Semiconductor (HK) Limited.
Shares in Everspin peaked at $9.99 during the first day of trading before pulling back Monday to $6.69 on volume of 161,170 shares, a drop of 33%.
The early-stage vendor has lost money every year since its 2008 spinout from Freescale Semiconductor Inc., now a subsidiary of Netherlands-based NXP Semiconductors. A net loss occurs when a company’s debt, operating expenses and taxes exceed its profit.
Everspin has accumulated an $89.7 million deficit and has $2.6 million in cash and cash equivalents on hand. Through June, Everspin’s net loss was $10 million on revenue of $12.8 million. Net losses in fiscal year (FY) 2015 were $19 million on total revenue of $26.5 million, which followed a $10 million net loss on revenue of $24.9 million in FY 2014.
Magneto-resistive RAM technology stores data as a magnetic state. Traditional semiconductor memory uses an electrical charge. Everspin MRAM is available as a discrete device or an embedded system on a chip. The products are designed to read and write data at speeds on par with DRAM and SRAM.
Everspin’s magnetic storage combines the persistence of nonvolatile memory with the speed and endurance of random access memory. The MRAM devices can be used as byte-addressable memory channel storage. Everspin stacks chips in a vertical plane to boost cell density and enable MRAM to function as persistent memory.
Everspin integrates magnetic storage by integrating industry-standard CMOS wafers during manufacturing at its Chandler fab plant. The storage uses a magnetic tunnel junction device to make calls to system memory. Manufacturing of higher density MRAM chips are outsourced to fab partner Global Foundries, which holds a $5 million ownership stake in Everspin.
Newcomers like Everspin and RAM card makers could represent the next wave of storage technology, following the dizzying pace of adoption of NAND flash. Everspin’s lower density MRAM products range from 128 kilobits to 16 megabits (Mb). Common uses are found in automotive, industrial and transportation application. Higher density chips range from 64 Mb to 256 Mb and provide magnetic storage to the enterprise storage market, including server, SSD and appliance vendors. Everspin counts Broadcom, Dell, IBM and Lenovo as storage customers.
Cloud archiving startup HubStor is fortifying its data protection, adding two methods to apply Azure encryption of at-rest data in Microsoft public cloud storage.
Customers of the Ottawa-based software company may opt to use Azure Storage Service Encryption (SSE) or Hubstor’s virtual cloud gateway to apply Azure encryption of cloud-hosted data. Which method to choose depends on an organization’s expected level of cloud-based searching and the accompanying security level.
The Microsoft SSE setting automatically encrypts data as it is written to persistent storage in Azure.
Encryption with Hubstor’s virtual cloud gateway preserves a subset of encrypted data locally, then synchronizes file shares to Azure. HubStor lets customers retain local control by applying 256-bit AES encryption ciphers, although its approach limits indexing and search capabilities.
Hubstor’s indexing engine performs transparent decryption of data to render content. For that reason, the vendor suggests customers use it locally. It recommends Azure encryption for enterprises that regularly perform search queries on cloud data sets.
“Before these enhancements, any content encrypted before (moving to) the cloud was work for the customer and the data was hard to manage. We now make it easier to encrypt before the cloud, and the encrypted data integrates with HubStor’s data-aware storage platform to make it easy to isolate it in searches and policies,” Hubstor CEO Geoff Bourgeois said.
Hubstor launched its eponymous software suite in July. Thus far, Microsoft Azure is the only public cloud it supports. The Hubstor enterprise archive suite is an overlay atop the Azure cloud, particularly to tier cold or seldom used data related to legal regulations. Hubstor is installed behind a corporate firewall, but presents a cloud archive tenant for file storage in Azure.
Encryption is not the company’s only news. Hubstor plans a feature for cloud storage chargeback, aimed at law firms and project-oriented enterprises that store unstructured data for long periods. Chargeback equips Hubstor customers to track, visualize, and report on storage costs associated with particular projects, clients, or departments.
Bourgeois said Hubstor analytics are getting a boost in November from planned taxonomy and user tagging. Hubstor also is developing a file system driver that virtualizes inactive array-based data and tiers it to Azure.
Avere Systems recently announced its latest hardware, the Cloud-Core NAS (C2N) hybrid system that is integrated with object storage and can scale up to five petabytes.
The system is comprised of the FXT 5000 nodes for NAS and the CX200 nodes for object storage that is based on OpenStack Swift software. A full system includes a minimum configuration of three 1U CX200 storage nodes for a total of 120 TB of usable capacity when using triple replication for data protection.
The other minimum configuration is six CX200 storage nodes for 480 TB of usable capacity when using erasure coding for data protection. The erasure coding offer N+4 availability, so four servers or four drives can be lost and the system will keep running. It also offers a geo-dispersal capability for disaster recovery using three sites. The CX200 nodes are loaded with 10TB disk drives and capacity can be expanded in 80 TB increments.
“It’s a scaleable system that can go from three nodes all the way to 72 1U servers that gets over 5 PB of capacity,” said Jeff Tabor, senior director of product management and marketing at Avere Systems. “It provides NAS simplicity but also provides the efficiency of the cloud and it’s all integrated.
“The key part of the operating system is the data protection. One is erasure coding and the other is triple replication. Triple replication can be inefficient so the erasure coding gives both resiliency and efficiency,” Tabor said.
The FXT compute performance tier for NAS, which supports NFS and SMB, is an all-flash configuration that scales to 480 TB using solid state drives. The system supports snapshots, data migration, mirroring, compression and encryption.
Tabor said Avere Systems is targeting customers who are dealing with large file data. The system integrates private and public object storage with an organization’s existing NAS infrastructure so it allows customers to create a hybrid cloud and manage an entire heterogeneous infrastructure as a single, logical pool of storage. The C2N is integrated with Avere Systems’ global namespace.
“Historically, you would store that on NAS but NAS has some challenges,” he said. “The trend is to move away from NAS and move to the cloud. But it’s difficult moving that data to the cloud. What C2N provides is a simple way to get into the cloud. This is a complete edge-to-core configuration supported by Avere. C2N has a built in operating system, so it’s our cloud.”
Enterprise storage containers aren’t about to supplant virtual machines, but the trend line for Docker data center adoption is going up. Hurdles of persistent storage and enterprise data protection are being removed, allowing organizations to move from “monolithic applications” to containerized microservices, according to a recent industry webinar sponsored by the Storage Networking Industry Association (SNIA).
The Oct. 6 event was the first of two events planned as part of SNIAs’s Cloud Storage Initiative. SNIA-CSI chairman Alex McDonald, part of NetApp’s Office of the CTO, moderated the session with panelists Keith Hudgins of Docker and Chad Thibodeau of Veritas Technologies.
Typical Docker data center use cases have mostly centered on application development and testing, but the panel said container storage is undergoing big changes.
“Micro-service architecture is designed to enable applications to be deployed extremely fast and make them much more portable to run on a variety of platforms. Containers really are optimized for speed of deployment, portability and efficiency,” Thibodeau, a principal product manager at backup vendor Veritas, told an audience of about 140 attendees.
He said companies often get started by launching containers inside virtual machines, “but ideally, containers are designed to (give you) the most advantage by running on bare metal.”
Containers are similar to virtual machines, yet also distinctly different. Whereas virtualization abstracts underlying hardware, Docker software virtualizes the operating system, eliminating the need to supply each virtual instance with a hypervisor and guest operating system. Multiple workloads share compute, operating system and storage resources, yet run in segregation on the same physical machinery.
According to Docker, data center downloads of its Linux-based software have topped five billion since its launch in 2013. It claims more than 650,000 registered users. Microsoft threw its support behind Docker containers as part of Windows Server 2016.
Sensing its growing importance, most major storage vendors now have tools to use their arrays as a persistent storage back end for Docker. Data center demand is ticking upward, albeit gradually. Financial services firms spawn persistent storage containers to authenticate end users.
Hudgins listed payroll-processing giant ADP and government IT contractor Booz Allen Hamilton among major firms using Docker in some fashion. Hudgins, the director of tech alliances at Docker, said ADP approached Docker to build nimble infrastructure for application microservices, using private and public cloud storage.
“ADP wanted a fast, easy way to change their payroll processing as needed. They deployed Docker Data Center internally to run all their data processing in a micro-services-based way… using Docker Data Center on both an internal OpenStack private cloud and public components running in Amazon for people to check their pay stubs. (ADP’s) entire system is now running on Docker Data Center,” Hudgins said.
Docker is a common service platform that Booz Allen uses to host customized applications for its government clients at the federal General Services Administration. Hudgins said Booz Allen wanted to migrate from “monolithic applications toward a smaller component-ized structure,” running a commercial version of Docker hosted in Amazon Web Services.
“They greatly reduced their time to market for (customer) applications… and also reduced the surface attack area and improved security,” Hudgins said.
SNIA said a Dec. 7 webinar will highlight best practices on Docker data management.