Quantum CEO Jon Gacek teased what he called the “cloud offering” several times during the backup vendor’s earnings call this week but didn’t go deep into details beyond “our vmPro technology, along with our deduplication technology, is the basis of a cloud-based data protection offering that we will be introducing in the coming months.” In an interview after the call, he let on that the DXi would provide the backup, and there will likely be a service provider partner.
“We’ll probably launch with a partner first and go from there,” Gacek said.
Last October, Quantum revealed it plans to let SMB customers replicate data to the cloud from a new Windows-based NAS product. But that’s apparently not the same as what Gacek talked about this week. The SMB replication uses Datastor Shield software, which is different than the DXi software.
LSI CEO Abhi Talwalkar said during his company’s earnings call that the WarpDrive PCI Express (PCIe) card will be used in the EMC product.
“We’re expanding and increasing our focus in storage and server application acceleration, bringing the performance advantage of the Flash to enterprise servers, storage and networking applications,” Talwalkar said. “We are pleased to be participating in the EMC Lightning program.”
The LSI WarpDrive holds 300 GB of single-level cell (SLC) flash, and Talwalkar said LSI is close to releasing a second-generation WarpDrive that includes SandForce storage processors and supports multi-level cell (MLC) and eMLC flash. LSI and Micron have frequently been mentioned as likely partners for the PCIe flash component ever since EMC previewed Project Lightning in May. Industry sources say it is likely that EMC will use two PCIe flash sources, with Micron’s P320h PCIe card as the other.
EMC demonstrated Project Lightning at VMworld and highlighted the technology last October at Oracle Open World. It uses EMC’s FAST tiering software and PCIe flash to improve application response time and throughput by servicing reads in flash while passing writes through to the storage array.
EMC will include PCIe flash and a SAN host bus adapter in an appliance called a VFCache Driver. The PCIe flash will be used to access read data stored in cache while writes will be passed through the HBA to storage.
EMC CEO Joe Tucci has changed his mind about retiring at the end of the year.
During EMC’s earnings conference call today, Tucci said he has agreed with a request from EMC’s board to stay on as chairman and CEO into 2013. He said when he does step down, his replacement will come from EMC’s senior management team but the vendor is not yet ready to name the successor.
“After much soul-searching, I have agreed to extend my role as chairman and CEO into 2013,” Tucci said. “I’ve started to increase the responsibilities of my senior team. When the time is right, my successor will be named.”
Tucci last September told the Wall Street Journal that he would relinquish his CEO title by the end of 2012 and remain on as chairman for two years.
The top candidates to replace Tucci are Pat Gelsinger, president of EMC’s information infrastructure product group; CFO Dave Goulden; and Howard Elias, president of EMC’s cloud services group.
I was asked to do a disaster recovery review for a small non-profit corporation recently. While larger organizations regularly bring in somebody to review their preparedness for disasters, small businesses rarely bring in an outsider. This company had fewer than 20 employees, all at a single location.
As always, the first step was to interview the key people in the company. The purpose of these interview is to find out about the current situation and to understand what the staff believes about the DR plan. With a company of this size, it did not take long to understand the situation. The staffers generally believed they could handle a disaster and had no immediate concerns. However, the current situation did not give me the same confidence.
There were two servers used for the major applications, an accounting system and a CRM system. These servers were also used for general file sharing. Each of the two applications had customized reports added. The individual laptops and desktops each had their own office software installed as well as many unshared files and copies of the shared files.
A tape backup was run every night for the servers, and one of the staff took the tape home and rotated through a week’s worth of backups. The tapes were never checked. The service provider who would restore tapes was a part-time administrator who ran a business providing services for other organizations.
The administrator, known inside this company as “the guy,” would come over on demand when there was an issue. The DR plan was to take the tapes to the part-time administrator’s office and restore the data on servers there. This had never been done, not even as a test.
The company’s DR plan did not address the possibility of a regional disaster where the personnel were not available or operations were impacted by lack of power or a network failure. The feeling was that the operations could tolerate being unavailable for a week, and any longer impact was highly unlikely and had greater consequences that would overshadow being out of operation. The possibility of losing key personnel was not included in this review but was part of an overall staffing plan.
The shortcomings were obvious, but the real issue was the lack of understanding of their limitations and the practices required. There was an unwarranted belief that there would be no issue restoring data from tape and that any server could immediately assume the role of the application servers and did not need to be exercised regularly. This obviously meant that the company needed education around the topic of DR and best practices, and that the local service provider chosen may not have the skill or desire to do what was really best for the customer.
I wrote a report and made recommendations of what should be done. The flexibility to address the problems was more limited with the small business than with companies that I would typically deal with, so I needed to consider the expenses and training.
Small businesses need a disaster recovery plan and a set of practices to implement. They also need education about how to develop a plan, what to look for and some criteria around choosing a services provider (“the guy”). It will be interesting to follow-up and see what changes are made.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
Nexenta scored a $21 million funding round this week, and the open-source ZFS-based software vendor will use the money to expand globally and market its new virtual desktop infrastructure (VDI) product.
Nexenta’s NexantaStor software runs on commodity servers, turning them into multiprotocol storage systems. Nexenta CEO Evan Powell said Nexanta software was sold in $300 million of its partners’ hardware deals last year. The startup has more than 250 resellers. The largest is Dell, which uses Nexanta software in the Compellent zNAS product.
Powell said 50% of Nexenta’s sales are already international, and the vendor only has one person working outside of the U.S – in Beijing. He plans to add staff in China and open offices in Japan and the Netherlands and probably other countries.
On the product front, the vendor is preparing to launch NexentaVDI, a virtual appliance that integrates with VMware View. NexentaVDI lets customers quickly provision storage for virtual desktops, and helps optimize performance by allowing thresholds for IOPS per desktop.
Nexenta previewed the VDI software during VMware World Europe in Copenhagen last October. NexentaVDI is in beta, and Powell said he expects to launch around April.
Powell said another change coming is that he expects Nexenta software running on more solid-state device (SSD) storage systems this year. NexentaStor has been optimized to run on SSDs, but the hardware will continue to come from partners.
“As a software company, we can remove the pernicious vendor lock-in on storage,” Powell said. “Storage is one of the last bastions of lock-in business models. Customers want to know how much they’re going to pay for storage in the future, and there’s a pent-up demand to get back at storage vendors who have exploited their customers for 10 or 20 years. We publish our prices and we don’t lock you in [to hardware]. But users like to buy arrays, they want to buy a box, plug it in, see the lights blink, and they have storage. So we reach out to vendors who sell arrays.”
Nexenta could lose its biggest array partner, however. Dell has made it clear that it is integrating clustered NAS technology it acquired from Exanet into Compellent SAN arrays to make them multiprotocol systems. After that, will Dell need Nexenta?
Powell is hoping that Dell will continue offering zNAS as an option for Compellent. He said one prospective customer is looking at a multi-petabyte deployment including zNAS. “I believe there’s room for both proprietary scale-out NAS with Exanet and zNAS with NexentaStor,” Powell said.
We’ll have to wait to see if Dell agrees.
WhipTail, the all-flash storage array vendor tucked away in Whippany, N.J., closed a Series B funding round and revealed a high-profile customer this week.
Although WhipTail failed to disclose the amount of its funding, but industry sources say it was about $9.5 million. That’s not in the same ballpark of the $35 million and $40 million funding rounds its rival Violin Memory secured last year, but WhipTail CEO Dan Crain said his company is close to profitable with close to 100 employees and is picking up about 20 customers per quarter.
“We are well-capitalized,” Crain said.
WhipTail bills its XLR8r as a cost-effective enterprise all-flash array, using multi-level cell (MLC) memory drives. The vendor goes after customers with a virtual desktop infrastructure (VDI), but Crain said it serves many types of industries.
AMD’s System Optimization Engineering Department said it replaced 480 15,000 RPM Fibre Channel drives with WhipTail’s solid-storage arrays for a 50-times improvement in latency and 40% performance increase.
AMD did not say how much flash capacity it bought from WhipTail, but Crain said is average deal is in the 25 TB to 30 TB range.
WhipTail isn’t the only all-flash array vendor out there. Nimbus Data, SolidFire, Texas Memory Systems, and Violin have all-SSD systems, Pure Storage is in beta and the large storage vendors will likely follow. Unlike a lot of the all-flash vendors, though, Crain said “We don’t compete on price. We solve a myriad of problems around performance.
“The field is still narrow for credible SSD manufacturers. The storage industry inherited NAND, and there is a lot of science and engineering that has to go into making NAND work in the enterprise,” he said. “We understand this stuff. We treat NAND and flash memory like flash, we don’t treat it like a hard disk.”
Symantec Corp. plans to use its newly acquired LiveOffice, a cloud-based archiving service, to provide end users with better search and data analysis capabilities for legal documents stored in the cloud.
Symantec announced on Monday it acquired LiveOffice for $115 million, a transaction that was completed on Jan. 13 and now gives Symantec an in-house, cloud-based archiving tool for legal documents, e-mail, file-sharing services and communications on social media sites such as Facebook, LinkedIn and Twitter. Symantec and LiveOffice have had an OEM relationship since 2010 and the archiving service was rebranded as EnterpriseVault.cloud in April 2011.
LiveOffice already has some level of integration with Symantec’s Enterprise Vault and Clearwell eDiscovery platform to provide email storage management, legal discovery and regulatory compliance. Now Symantec can more tightly integrate LiveOffice with Clearwell so end users can perform more detailed data analysis and generate narrower results when searching for legal documents. The archiving tool serves as the knowledge repository while the eDiscovery platform provides the analysis capability.
“When you are looking for these legal documents, it’s like trying to find a needle in a haystack,” said Brian Dye, vice president for Symantec’s Information Intelligence Group. “Many times in these cases what you are looking for boils down to four or five documents. If you can get tighter and tighter results, you are transferring less data.”
Symantec also plans to build a common user interface and account provisioning tool for LiveOffice and its anti-spam Symantec MessageLabs Email Security.cloud
“We don’t have a time frame [for delivering the enhancements] right now,” Dye said. “We will have one quickly.”
LiveOffice has nearly 20,000 customers, Forrester analyst Brian Hill wrote in a blog about the deal. The company “historically marketed to small- and mid-sized financial services firms. Over the past couple of years, however, the vendor has steadily bolstered its archiving and broader information governance functionality, lined up productive partnerships with major technology vendors, and met with success in selling to larger organizations across a wider set of vertical markets,” Hill wrote.
There are similarities to the advances in storage systems and the advances we’ve seen in automobiles. When you’ve spent most of your life working on both, the similarities become noticeable.
Storage systems today have a focus on improving simplicity. That’s simplicity from the standpoint of being easy to install and operate. Installation simplicity is measured by the number of steps or time it takes to provision volumes or file systems.
Beyond that, storage systems simplify management with self-tuning capabilities. The tiering built into many of the more sophisticated storage systems is an example of simplified management. Tiering can be automated movement of data between different classes of storage device – the most popular being solid-state devices (SSDs) and high capacity disks. Tiering can also be done by using caching with solid-state technology or DRAM. Most of these tiering features operate automatically.
These developments mean administrators no longer need specific training to handle many storage systems. The phrase used to describe the person that manages this storage is an “IT Generalist.” This development changes the requirements for IT staffing.
The analogy between storage systems and automobiles may be superficial, but makes for an interesting discussion. Tuning used to be a complex process. Tuning up an automobile meant setting the points and adjusting the timing by using settings on the flywheel that have been replaced with electronic ignitions. No more tuning is required – or possible – in most cars. Adjusting the carburetor was another seemingly never-ending task. You had choke control settings, air mixture valve settings, and don’t forget balancing in a multi-barrel carburetor. Fuel injection systems have changed all that. There are no adjustments now.
There are also many other monitoring and reporting systems for cars. Rather than listen or do examinations (sometimes called an Easter-egg hunt) to find a problem, it can be located through on-board diagnostics. This all makes is much difficult to make any adjustments and to “fix it yourself.” Few people have the detailed knowledge of the systems in their cars. Fewer still would know what to do about a problem.
So the car now has an IT generalist who can take the car to a specialist who owns the right equipment when there is an issue. With a storage system, the vendor support group — with the right tools — will diagnose and make repairs. As for tuning the storage system, there are systems that allow that to be done. But it takes a specialist with the correct training and tools to do it.
Overall, this is better for IT. The savings in training and personnel costs are evident. But there’s still that Ford with a 289-cubic-inch engine with a Holley carburetor that needs some minor adjustments.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
EMC failed to push its Project Lightning server-based PCIe solid-state flash product out the door by the end of 2011 as the vendor pledged, but industry sources say it will officially launch soon under the name of VFCache.
EMC recently registered a trademark for the VFCache name under the description: “Computer hardware, namely, data caching devices including flash memory devices and computer software for data storage and data management.” The vendor previewed Project Lightning at EMC World in May and released it to beta later in the year. The only surprises left are what partners – if any – EMC will use for the product. VFCache is expected to work with EMC’s FAST tiering software, but whose PCIe flash will EMC use?
Along with launching a new backup deduplication appliance, Dell made other storage additions and enhancements today in London at its first Europe Dell Storage Forum. The biggest rollout, besides the DR4000 backup box, was an upgrade to Compellent Storage Center 6.0 software with new 64-bit support that doubles memory size.
The upgrade -– along with extended VMware support –- is part of Dell’s strategy to make the Compellent Fibre Channel SAN platform a better fit for the enterprise. The 64-bit support is a precursor to the addition of Ocarina Networks’ primary data reduction technology to Compellent systems, because deduping and compressing data will require more processing power. Another advantage for Compellent 6.0’s 64-bit support is it enables tiering data in smaller block sizes to automatically tier data more efficiently.
Compellent also now supports the full copy offload and hardware-assisted locking features that are part of VMware vSphere Storage APIs for Array Integration (VAAI). The storage vendor also added a Dell Compellent Storage Replication Adapter (SRA) for VMware’s Site Recovery Manager 5, and vSphere 5 Client Plug-in and Enterprise Manager to help manage virtualized storage pools with the latest version of vSphere.
Randy Kerns, senior strategist for the Evaluator Group IT consultant firm, said the 64-bit support will enable Compellent to better take advantage of next-generation Intel chip advances. He said that’s a nice benefit because of Compellent’s architecture and licensing model. “People underestimate the importance of this, but Compellent is about storage as an application and the applications are loaded on powerful Intel servers,” he said. “With Compellent, you buy a license and you don’t have to re-buy a license when you upgrade. This also lets them track the new technology brought out by Intel and leverage Intel’s research and development.”
Before the Dell acquisition, Compellent sold into the midrange. Dell already has the EqualLogic platform for the midrange and is looking for something more competitive with EMC VMAX, Hewlett-Packard 3PAR, IBM DS8000, Hitachi Data Systems Universal Storage Platform and NetApp FAS6200 systems. But to become a true enterprise option, Compellent may have to scale beyond its current two-controller limit.
“When Dell did not get 3PAR, Compellent was the only option left worth looking at, but it doesn’t go high enough,” said Arun Taneja, consulting analyst for the Taneja Group. “Dell is feverishly working on taking Compellent upstream. One of the elements needed is 64-bit support. But to compete with the likes of 3PAR and VMAX, Compellent has to go to more than two controllers. What if Dell cannot take it to four or eight controllers, what are they going to do? The next 12 months will be telling. For five years, the Compellent people have been telling me they can go beyond two controllers. We’ll find out if they were telling the truth.”
Dell also added support for 10-Gigabit Ethernet Force10 switches on its EqualLogic iSCSI SAN platform and support of Brocade 16-Gbps Fibre Channel switches for Compellent.