It’s the New Year and it’s time to organize things around the home. Throw out things that you no longer need. Take things that you probably won’t need but you’re not sure about, and put them in the attic. It’s a fairly simple exercise, but not done regularly. It happens now because the advent of a new year reminds us to do it.
This is similar to what storage administrators need to do this time of year. They need to organize their data, delete data that is no longer needed, and archive data that is unlikely to be used in regular processing. Maybe the archive target device should be called the “attic.” What they are really doing is making decisions about information. What is the value of the information, who owns it, what restrictions regarding compliance are there for that information, and can it be deleted?
There are several approaches to making decisions about the information. Some people don’t make decisions because there is no clear guidance and they might make the wrong choice. An interesting strategy from one storage administrator was to archive data to tape without migrating the data to new tape technology when the old generation of tape drives became obsolete. Eventually, there would be no drives left that could read those tapes and the administrator did not have to worry about the decision to delete the data.
Archiving has become a misused or at least misunderstood term. It really is about taking data that is not expected to be needed and moving it to another location. This is done for economic advantages, and potentially to meet regulatory requirements. Over time, the term has expanded to “active archive” and “deep archive.” Active archive is for data that is retained in the original context so an application or user can retrieve it without an intervening process. Deep archive is for data that is not expected to be needed, but may be. Both locations can support immutability, versioning, and other compliance requirements. There are advocates of using cloud-based storage for deep archive.
Managing information effectively includes archiving data by making intelligent decisions about what gets archived and where. The economic value of moving data off primary storage systems is great. An ongoing policy to move data periodically compounds that value. It should not take a trigger such as a new year to make a decision about organizing and moving unneeded stuff to the attic. It is a valuable IT process.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
Nearly half of employees use online file sharing services even though their companies have a policy against it, according to a recent report conducted by cloud storage vendor Nasuni.
For its Special Report on Shadow IT in the Workplace, Nasuni surveyed 1,300 corporate IT users. Online file sharing and bring your own devices (BYOD) are gaining in popularity as employees use PCs, iPads, smartphones and tablets to access work-related files. And they use these devices to work from airports, homes, cafes and other locations outside of the corporate office.
Companies are starting to call this “shadow IT” because these services and devices often are not controlled by traditional IT departments. For many companies, it’s a growing security problem.
“We were surprised that the percentage of users is one out of five. That is not a trivial number. There is an incredible amount of usage and it was higher than we expected,” said Connor Fee, Nasuni’s director of marketing. “We talked to a lot of IT guys and they complained about Dropbox specifically.”
Other services mentioned include iCloud and Google Drive. The survey included 300 IT managers and 1,000 employees. The survey found that one in five employees uses Dropbox for work files, and half of employees who use file sharing services do not know if their companies have a policy against it.
The report also found that corporate executives are the worst offenders, with vice presidents and directors most likely to use Dropbox despite the security risks. About 58% of employees with personal smart phones or tablets access work files from those devices. The survey found that 54% of respondents work at organizations that do not allow access to file sharing.
“The fact that corporate leaders are the worst offenders tells me IT is failing to deliver on something that is needed,” Fee said. “Having and educating about a policy is not enough. It needs to be address beyond policies.”
IBM has announced plans to acquire partner StoredIQ, an Austin, Texas-based software company that specializes in classifying and managing big data, particularly for regulatory compliance and legal reasons.
The financial terms of the acquisition were not disclosed, but IBM expects to close on the deal around March of 2013.
StoredIQ has been an IBM partner for two years. IBM’s Tivoli data management and archive storage systems are already certified to work with StoredIQ applications. The technology dynamically identifies and classifies data that either needs to be retained or decommissioned, while also governing who has access to what data.
Big Blue plans to make StoredIQ’s technology a part of IBM’s Information Lifecycle Governance (ILG) suite that has the ability to set management policies, mine and assess what data is valuable and what data should be deleted. The company’s software does not overlap with IBM’s current products in this area, said Ken Bisconti, IBM’s vice president for Enterprise Content Management (ECM) software.
“They have the ability to dynamically identify data that is in place,” Bisconti said. “Most vendors require companies to move files to a separate repository before it’s classified and managed. StoredIQ dynamically collects that information wherever it resides. We have the ability to collect data in a repository but we did not have the ability to dynamically collect it.”
Bisconti said IBM will retain StoredIQ’s staff, which they consider essential intellectual property. The company has 50 employees to date. Also, more integration will be done with the ILG suite. StoredIQ’s software already works with the ILG portfolio. IBM’s policy governance technology manages the overall plan of identifying high-value data versus low-value data. Those instructions are sent to the StoredIQ engine, which executes the policy.
“For example, StoredIQ can do e-discovery for data that is in place if you have to respond to a request for discovery material,” Bisconti said. “Typically, companies have a difficult time to get to data that was not moved to a repository.”
StoredIQ was founded in 2001 under the name of DeepFile Corp., a data classification company that created metadata to manage different types of unstructured data. The company later changed its name to StoredIQ and focused on compliance and e-discovery to help companies figure out what data to keep and what to delete.
IBM’s information lifecycle governance business is part of its software group.
Exablox came out of stealth today by disclosing it has $22 million in funding, but CEO Doug Brockett is only dropping hints about its product until a full-blown rollout next spring.
Brockett and director of marketing Sean Derrington described the product as NAS-like, with flash as well as hard drive storage and managed at least partially through the cloud. They said it will share characteristics with cloud gateway products, but won’t be a direct competitor to cloud NAS vendors Nasuni, Ctera, Panzura, and TwinStrata.
“It’s more than a gateway or caching appliance for file serving,” Derrington said. “We want to enable customers to easily manage capacity and performance on-premise. If they choose to locate information outside of their primary data center, they have the flexibility to do it.”
Brockett said the key characteristic for Exablox storage will be the ability to scale without complexity.
“How do you manage a scale-out infrastructure spread across lots of locations,” he said. “We think the answer is having a management system that runs on the cloud itself instead of the device.”
For now, Brockett is focused on scaling out the Mountainview, Calif., company that so far consists of 28 employees and five contractors. “It’s me and Sean and a bunch of engineers,” Brockett said. “We need to build a go-to-market team.”
Brockett said the product is designed for companies with from 50 to 500 employees, and many are already running the system on a trial basis.
The funding comes from venture capital firms DCM, Norwest Venture Partners and U.S. Venture Partners. The $22 million consists of two rounds of funding, with the first dating to 2010. Brockett comes from SonicWall and Derrington worked in storage product and cloud product marketing for Symantec.
Competition in the all-flash market will grow intense in 2013, and startups Whiptail and SolidFire this week moved to strengthen their companies.
Whiptail closed a $31 million funding round today. Ignition Partners led the round, with BRE Ventures and Spring Mountain Capital participating, along with strategic investors SanDisk, an unnamed “Silicon Valley industry titan,” and debt financing from Silicon Valley Bank. Whiptail also hired a new CFO, Catherine Chandler.
SolidFire bolstered its senior executive team, adding RJ Weigel as president, John Hillyard as CFO and Tom Pitcher as VP/International.
Whiptail is the first startup to receive funding from SanDisk Ventures’ new $75 million fund for strategic investments. Alex Lam, director of SanDisk Ventures, said SanDisk picked Whiptail among the flash array vendors because its arrays can scale into tens of terabytes today with plans to drastically extend that.
“There’s a lot of noise in the industry from companies talking about the size of the round they raised or they got investments from Sequoia or somebody like that,” Lam said. “But you really want to look at the core technology. I look at the ability to take a terabyte and scale to petabytes without the customer having to purchases a new platform.”
Whiptail recently said its upcoming Infinity storage will scale to 360 TB in early 2013, and CEO Dan Crain said it will eventually go to petabyte scale
SanDisk and Whiptail had no previous relationship, but you can expect Whiptail to get its flash memory from SanDisk now. Lam said SanDisk will likely look to invest in server-side flash and flash software companies next.
“This is our first stake in the ground to show we’re serious,” he said. ”We view flash as disruptive in enterprise storage. We want to build up an ecosystem of enterprise flash technologies.”
Whiptail received a much smaller funding round – less than $10 million – in January 2012.
SolidFire stands out from other all-storage array vendors because it sells almost solely to cloud providers. Weigel fits with that strategy, because he ran sales and field operations at 3PAR in the early days when a good part of its customers were service providers such as Savvis and Terremark. Weigel said cloud providers hold great potential for SolidFire.
“So much of what’s in the cloud now is not the most critical apps,” he said. “People are putting test/dev and backup in the cloud, but customers have been waiting for quality of service and guaranteed SLAs for the most critical apps, such as Oracle. We’re going to deliver on that promise. Cloud service partners will be able to put together a business practice around our storage.”
But Weigel and SolidFire chief marketing officer Jay Prassl said they can see the day when SolidFire moves into the enterprise as well.
“Cloud providers are step one before taking the next step into large enterprises,” Prassl said. “We get a lot of calls today from enterprises, and we don’t hang up the phone. We’ll be announcing some of them next year. But companies that have a specific focus like SolidFire can do well in the cloud space.”
Weigel added: “Obviously, there will be a time and place where other markets make sense, but we are focused on the service provider cloud space today. It’s a great growth market for us.”
Object storage is a method of storing information that differs from the popular file storage and venerable block storage that are the most familiar in IT. It is another a type of storage where information and metadata are both stored, although the metadata may be stored with the actual information or separately.
We often see new object storage products these days with slightly different implementations. While many of these new object storage offerings are designed to solve specific problems for customers, all have the opportunity to be used across many different applications and environments.
The object storage of today is different than what some may have been familiar with in the past. Previously, a content address was used to identify data put into a storage system such as the EMC Centera. The new object storage, for the most part, is storing files with associated metadata frequently using HTTP and REST. The metadata can be different depending on the implementation or the application or system, and contains information such as data protection requirements, authorizations and controls for access, retention periods, regulatory controls, etc.
New object storage systems address storage challenges, including:
• Massive scaling to support petabytes and even exabytes of capacity with billions of objects.
• Hyper performance data transfer demands that go beyond the traditional storage systems used in IT today.
• Compliancy storage for meeting regulatory controls for data including security controls.
• Longevity of information storage where data can be stored and automatically transitioned to new technologies transparent to access and operational processes.
• Geographic dispersion of data for multiple site access and protection from disaster.
• Sharing of information on a global scale.
For the vendors offering new object storage systems, success with narrowly targeted usages can eventually spread to opportunities in enterprises. They address problems that already apply in the enterprise, but perhaps not at the scale that requires object storage yet.
Some of the vendors offering object storage today include:
Data Direct Networks Web Object Scaler (WOS)
HDS Hitachi Content Platform
Scality Ring Storage
Many of these vendors offer a file interface to their object storage as well as the native object API using HTTP and REST.
The types of object storage are developing so fast that the terminology is inconsistent between vendors. I attended the Next Generation Object Storage Summit recently that was convened by Greg Duplessie and The ExecEvent. This event was a great opportunity for vendors and analysts to discuss the technology, and how to describe it and understand the current market place. It was clear in the summit that the initial focus for new object storage should first be on the problems being solved today and then on the opportunities to move into more widespread usage.
This will be a developing area in the storage industry and Evaluator Group will develop a matrix to compare the different solutions.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
With major storage vendors in various stages of preparation to launch all-flash arrays in 2013, the startups already selling flash storage are working to stay a step ahead. For some, this means adding storage management and data protection, while others work on making systems redundant and still others try to reduce costs.
Whiptail’s plan for staying ahead of the game is to make its all-flash arrays the most scalable in the market. The startup is preparing to launch its Infinity architecture in the first quarter of next year. Infinity is an expansion of the vendor’s current Invicta platform, except for it scales to 30 nodes and 360 TB of flash compared to Invicta’s six nodes and 72 TB.
And that’s just the beginning, says Whiptail CEO Dan Crain. “Our largest tested configuration is 30 nodes,” he said. “We can probably go 10 times that, but we haven’t tested it.”
It’s unlikely that anybody will need – or want to pay for – 3.6 PB of flash in one system for a while, so Whiptail has time to test larger configurations. But Crain said his strategy is to have an architecture in place for his early customers to grow into as flash takes hold.
“Our basic message always has been organized around building a platform that folks can invest in and keep building onto,” he said. “People can take anything they’ve ever bought from us and organize it into Invicta.”
Whiptail claims it has achieved 2.1 million IOPS and 21.8 GB per second throughput in testing with a 15-node 180 TB set-up, and projects more than 4 million IOPS and 40 GBps with 30 nodes.
Infinity requires several pieces of technology, including version 5.0 of Whiptail’s Racerunner operating system, and enhancements to the array’s silicon storage routers.
Crain said he doesn’t expect flash to take over the storage world overnight. He predicts it will be a gradual process as early customers use it for high-performance applications and eventually move other critical data onto flash.
That’s why he wants to get an early customer base that will grow into Whiptail storage as it supports higher scale.
“We’ve always said we’re going to build into the market,” he said. “We never go out and tell everybody we’re going to take over the world because that’s not rational. Adoption of our technology is in its infancy.”
Crain said Whiptail already does things such as real-time error correction, clustering, auto-failover and asynchronous replication. Deduplication, a potentially key feature for SSD because of its limited capacity, remains a roadmap item.
“Over time we’ll have dedupe,” he said. “We’re very sensitive on performance latency, so we tend not to compete on cost per gig. Dedupe has benefits in general, but it’s still not yet widely deployed on primary storage.”
Hewlett-Packard has announced a single architecture across storage systems that can span different sizes of enterprises. This is the HP 3PAR StoreServ that now includes the 7000 model to complement the high-end enterprise models currently available.
That gives HP one architecture that covers from the small enterprise though the largest enterprise data center systems.
On the surface, the announcement of the HP 3PAR StoreServ 7000 appears to be a new system for the mid-tier and small enterprise. In reality, it represents a fundamental decision about leveraging investment in an architecture that can scale across multiple market segments and meet market demands such as performance, capacity, resiliency, and advanced operational features at different price points. By leveraging its investment in 3PAR, HP can maximize R&D and support for storage. The 7000 now allows HP a broad breadth of coverage with the single architecture.
Except for NetApp’s FAS platform, no major storage vendor has one architecture that spans from low-end SAN through the high-end of the enterprise. HP does have other storage platforms, such as the StoreVirtual (formerly LeftHand) and the XP P9500 that is re-branded from Hitachi for mainframe storage, but these fit on the extreme high and low ends. Extending 3PAR’s architecture allows HP to phase out its aging EVA midrange platform.
The advantage of leveraging a single storage architecture seems obvious but has been contradictory to the method most vendors use to deliver products to different segments of the market. That’s because they usually gain products through acquisition. That method is expedient but creates independent offerings that require separate (and costly) development and support teams. HP gained 3PAR through acquisition, but the architecture was flexible and scalable enough to address the range of customers from the small enterprise to the data center.
Leveraging one architecture has benefits for both the customer and for the vendor. For the vendor, focusing on one team for R&D and support drives down costs and makes for a simpler sales engagement.
The most important benefit for the customer is a longer product lifespan. With the vendor not having to invest in a diverse set of products, there’s an obvious commitment to the product line. That dramatically reduces customers’ worries that an end-of-life decision will be made based on the economics of investment in that product.
Other benefits include the availability of what may have been considered high-end enterprise features on lower-end systems. For the customer, the continuity of the storage architecture reduces the interruptions that occur when changing processes or moving from one model to another.
A single architecture is part of an evolving landscape for storage. The leverage of hardware technologies and embedded software has been in progress for some time. HP terms that Converged Storage and it is represented in other products included in the major storage announcements beyond the HP 3PAR StoreServ 7000. They include the HP StoreOnce for data protection, HP StoreAll for file and object storage, and HP StoreVirtual for flexible, economic iSCSI storage.
You can expect to see more major vendors going to a single storage architecture with highly leveraged hardware and embedded storage. It makes economic sense for the vendor and the customers. The key is that the architecture must be able to scale to meet the demands in the different usage models for performance, capacity, and price.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
The end of the year is a busy time for storage pros as salesman push to meet year-end quotas and IT plans for year-end operations.
Year-end operations can involve beginning the processes required to close the books of a company and many other business tasks. For the storage group, it typically means additional projects that can only be done when user activity is reduced. Most companies limit operations between Christmas and New Year’s, making it an opportune time for storage projects such as:
• Moving data from one storage system to another. This is done for several reasons: balancing workloads for data access to improve storage systems’ performance; balancing capacity to meet expected demands; and to utilize space more efficiently.
• Moving data off storage systems that are due to be retired as they come off maintenance or warranty.
• Deploying new storage systems to meet increased demand for capacity or performance. Deploying new storage is quickly followed by moving data again to distribute it according to application requirements.
• Increasing the size of data stores such as databases based on demands.
• Performing an end-of-year data protection cycle that will retain information based on business governance demands.
These projects are all critical to operations. Based on the conservative nature of storage professionals, they view these as tasks best done where the potential for impact is least.
So Happy Holidays for the storage (and other IT guys). The time will be spent either in the data center performing the tasks or at home monitoring and controlling the tasks remotely.The image of the storage guys watching multiple screens for operational status while the holiday parties rage on without them is real and has been the experience of most of us that have been in the industry for a long time.
So, why are there no Holiday-proof storage systems or data management software in wide usage? These would be ones that can balance data across the different systems (from different vendors). There are some systems and data management software that can balance or migrate data across like systems or a narrow subset of heterogeneous systems. There are even a few products that can work across any storage platform. But, for the most part, IT storage people still schedule these activities for reduced demand times to minimize potential impacts because they have experienced some impacts in the past and that memory was painful.
Some of the new products (software and hardware) seem to be quite good but the capabilities are limited to a few at this point. The confidence in using them is built over time and eventually the automation will seem like a commonplace activity and not something that requires special attention. But for now it’s the type of activity that makes storage pros fear a “Danger Will Robinson, danger” moment.
This will change eventually, and the confidence will grow for the systems and software that can do these activities across multiple operational environments and not only for very specific usages. Vendors continue to make advances and the successes will allow for greater usage in the non-holiday time. Then maybe the storage guys can attend those parties without having to be on call or check on status.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
Seattle-based startup Qumulo closed a whopping $24.5 million Series A funding round last week, without even dropping the F-word (flash) or C-word (cloud) that many startups rely on to woo venture capitalists these days.
The Qumulo press release did delve into data growth and played up the team’s Isilon connection. CEO Peter Godman, CTO Aaron Passey and VP of engineering Neal Fachan helped develop Isilon’s OneFS clustered file system that propelled that company to an IPO in 2006 and a $2.25 billion buyout by EMC in 2010.
Qumulo’s executives left Isilon in between IPO and acquisition. Now Godman says he would like to recreate the Isilon culture, even if he can’t replicate the software because EMC now owns the intellectual property. The Isilon connection helped sway Highland Capital Partners, Madrona Venture Group, and Valhalla Partners to invest in Qumulo’s first round.
“Our Isilon experience was a relevant factor in our fund raising [with the VCs], but Isilon was also an extraordinary event in our lives,” Godman said. “It was a vibrant and unique culture, and I give credit to Isilon founders Sujal [Patel] and Paul [Mikesell] for creating that experience.”
Mikesell is VP of engineering at Clustrix and Patel is training for marathons after leaving EMC last month, but Qumolo is sure to have other former Isilon employees on the team. Godman said he plans to expand Qumolo’s 18-person team by 50 or so with almost all the hires based in Seattle.
Godman won’t talk about specifics of the product they are developing, but Qumulo’s press release said the startup will solve manageability, scalability and efficiency problems in storage. Those same characteristics apply to Isilon’s OneFS but Qumulo can’t copy that technology.
“We’re respectful of the IP ownership issue,” Godman said. “Everyone who is an engineer has had to deal with that need to stand clear of things you know are incumbent. But the flip side is it’s easiest to avoid infringing on things you know about.”
He said Qumulo will reveal the timeframe for its product next year. He is willing to address more general storage topics, such as how much the underlying technology has changed in the more than a decade since Isilon began developing its clustered file system.
“Object storage is starting to come into its own now, with a lot of vendors and Amazon S3 using it,” he said. “That part has changed a lot. Also, NAND flash is here now. Its rapidly dropping cost and performance characteristics are disrupting storage technologies, so the kind of storage you build for that looks different than the storage you build for hard disk drives. That couples nicely with the emergence of virtualization. Virtual machines place stress on storage that NAND flash is uniquely suited to address in a cost-efficient way.”
Are there any hints there in what Qumulo is doing? We’ll find out in 2013.