Actifio bills itself as the “radically simple data copy storage company,” but apparently it’s pricing and packaging models weren’t radically simple enough. So Actifio 5.1 released today has new packaging and pricing, including a turnkey solution combining storage and software, a gateway version without storage, and a utility pricing model.
Actifio claims it made 62 sales in the fourth quarter of 2012 for an average price of more than $200,000 with its appliances that change the data protection process. But the pricing could confuse customers used to more traditional backup.
“Customers told us, ‘It can be difficult to comprehend exactly what I’m buying,’” said Andrew Gilman, Actifio’s senior director of global marketing. “So we spent time crafting our packaging. We have pre-packaged configurations that make it easier to understand what’s showing up on the data center floor, and how it can be deployed quickly.”
The Actifio 100T turnkey system comes in eight standard configurations, ranging from 7 TB of protected data for remote offices to 100 TB for enterprises. Customers can stack 100 TB configurations up to 2 PB.
Actifio Gateway lets managed service providers use their existing hardware to build clouds and services based on Actifio’s copy data technology. Its software provides multitenancy and is designed to work on a mix of virtual, physical servers and storage devices.
The utility pricing model lets customers pay as their data volumes grow. It includes software, storage, maintenance and installation, starting at around $4,000 per TB per year for 30 TB and dropping to around $600 per TB in larger volumes.
Gilman said Actifio began piloting the utility model in late 2012, and it is now available on all models with more than 30 TB. Customers can sign up for three-year or five-year subscriptions.
Gilman said Actifiio will let its channel and service provider partners decide whether to use the utility model for their customers.
Actifio also added features in 5.1. They include bare metal restore to enable rebuilds of Windows and Linux hardware, the ability to non-disruptivet failover testing, and the ability to failback and SyncBack after a disaster by only moving changed application data from the DR site to the primary site.
EqualLogic founder Paula Long has a new storage startup, and she says its goal is to do for data scientists what EquaLogic did for storage administrators.
Long was one of three founders of iSCSI storage pioneer EqualLogic, which proved Ethernet can play a major role in network storage. She teamed with former EqualLogic marketing VP John Joseph a year ago to start DataGravity, which today pulled in $30 million in Series B funding to give it a total of $42 million with no products due for at least a year.
DataGravity CEO Long and president Joseph won’t give too much detail on what the company is up to, except to say it’s about gaining business intelligence from data in an easier way than is possible today.
“At EqualLogic our mission was to put an IT administrator in every storage array so an IT generalist could become a storage expert quickly and not spend a lot of time managing storage,” Long said. “We also went for the all-inclusive approach instead of dim sum although everybody thought we were nuts. We took a complicated solution and made it simple.
“Now we’re taking storage experts and marrying them with analytics and database experts, and deriving what happens when you put all those people together. We’re taking EqualLogic’s core principles of simplicity and everything packaged together, and moving that into delivering data intelligence in a simple way.”
The dim sum approach she referred to is the way the major storage vendors sold all of their software features as separate licenses. EqualLogic included all of that in the price of the array, which fit well with its strategy of offering Ethernet storage that was easier to manage than Fibre Channel and didn’t require expensive FC networking devices. EqualLogic was poised to go public when Dell acquired it for $1.4 billion in 2008.
“The last time we looked at iSCSI, SAS, SATA, and all the key enabling tools,” Joseph said. “In this company, we’re looking at a different level of the problem. The focus is now on data and the value that data brings to someone’s industry.”
Long said she expects DataGravity to start shipping product in 2014. She did say it would sell devices and not just software. When asked if the systems would be similar to big data analytics products such as Oracle Exadata, EMC Greenplum, and IBM Netezza, she said “I don’t think there’s anybody similar to us. I’ve seen slices of what we’re doing, but not the same solution.”
Venture capitalist firm Andreessen Horowitz led today’s round, with previous investors Charles River Ventures and General Catalyst Partners also participating. Andreesen Horowitz partner Peter Levine joins DataGravity’s board.
Long said DataGravity went for a B round early to get Andreesen Horowitz and Levine on board. “We’re frugal New Englanders,” she said. “We weren’t planning on raising money until the fourth quarter of this year. But Peter approached us. We thought having him join us early so he can help shape the company would be a great help.”
Levine is the former CEO of XenSource and executive vice president of Veritas Software long before Symantec acquired it. He’s also a current board member of startup Actifio. He said a startup’s leadership plays a key role in his decisions to invest.
“The reason we invested in this company is because of Paula and John,” he said. “No one in storage is more strategic or better able to articulate the storage market than Paula. The concept of turning data into information is quite profound.”
When asked what challenges DataGravity faces, Levine said “Getting a product out is always a challenge. And any time you build a new category, there’s the education and evangelism involved with going after a new market.”
Taking a step back and looking at the new or updated storage systems coming out reinforces trends I see for storage hardware in the industry. By storage hardware, I’m referring to physical system components such as the controller cards, processors and data flow electronics.
The majority of the underlying hardware used in storage comes from standard components. Many of the available storage systems use Intel-based server designs. The value to the vendor from the sale of a storage system comes primarily from the embedded software running on that hardware and associated server software used in conjunction with that storage system.
The embedded software features and capabilities that differentiate one storage system from another are the most focused on product characteristics. For most storage vendors, the software-based features are categorized and put into suites of functionality that have additional charges over and above the basic system costs.
Storage systems that cannot continue to add unique value gravitate to a price based on the hardware costs with some added margins. Characterized as a “race to the bottom,” the undifferentiated systems are generally referred to as commodity storage.
The view that storage hardware is generally moving to commodity may be a natural evolution but is personally disappointing. The manufacture of the common platform for storage systems has moved to lowest cost regions of the world. The equitability of low wages is a different topic, but the commoditization does have a direct relationship with larger scale production and magnifies cost reduction value.
There is a parallel that comes to mind, which is the U.S. steel industry. The technology and production processes for the steel industry were developed and refined in the U.S. over time. But production has moved to lower-cost regions in the world based on price pressures. With the movement of a manufacturing industry, the investment to build production facilities is used to modernize the production.
The surface level similarity between the storage hardware system base and the steel industry has several factors: production moved to follow lowest cost manufacturing, underlying technology was not changing, and investments in the industry were to make manufacturing more efficient.
There still are custom hardware based storage systems in the industry, but they are a minority and the investments in new ideas and hardware technologies seem limited. The competitive price pressures and the continued investments required for innovation make it more challenging to sustain a company with novel technology.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
Symantec executives Wednesday indicated backup and storage would continue to play a big role in its future, especially after a strong fourth quarter for NetBackup.
Symantec folded its earnings report into a two-hour presentation focused mainly on outlining the company’s strategic direction under CEO Stephen Bennett, who replaced Enrique Salem last July.
Symantec executives said NetBackup revenue increased 12% year-over-year, highlighting the company’s overall four percent revenue increase. Bennett and CFO James Beer said NetBackup appliances continue to sell well. They did not mention Backup Exec, which the execs have said needed fixing in the past few earnings call. Backup Exec had a tough year, following a poorly received upgrade of the SMB backup application at the start of 2012.
Symantec’s storage management product revenue grew one percent, which is the second quarter of growth after several of decline.
Despite the good results, last quarter’s earnings were overshadowed during the presentation by Symantec’s future plans. NetBackup would seem to figure prominently in those plans, especially on appliances.
The execs also said they would simplify licensing and improve technical support for all products. Bennett said he intends to make it easier to get support through other methods than by phone. “We weren’t delivering technical support experiences for our customers that we were proud of,” he said.
The Symantec execs included an object storage platform on its list of 10 product areas it will focus on. It’s not clear if that would be a new platform or object storage built into FileStore. In 2010, Symantec folded development of its nascent S4 object storage file system into FileStore.
We’ll have to wait for details on Symantec’s object storage plans. As Bennett said Wednesday, “We have nothing to announce today” as far as product specifics.
Symantec didn’t announce layoff specifics either, although Bennett said there would be a $275 million severance charge. Bloomberg reported that up to 1,000 jobs will be cut, mostly middle managers. Some will be replaced by new hires in research and development. And Reuters reported that Bennett does not plan to sell any of Symantec’s assets for now.
“We’re going to have bigger, better jobs for fewer people,” Bennett said.
Bennett will actually have one fewer job. He is giving up the chairman role he also held. Symantec split the CEO and chairman roles, naming board member Dan Schulman as the new chairman.
Bennett said he spent much of the last quarter of 2012 on a world tour meeting with customers, employees and investors while putting together the company’s forward strategy. He said the next year or so will be one of transition.
“We have to change the tires while the car is running,” Bennett said. We know which customer he got that idea from.
When startups raise funding, contributors often include strategic investors who hope to benefit from the technology the new company develops. Storage startups usually turn to technology industry giants for strategic investments.
Cloud storage startup Symform is going in a different direction. Symform today said it received a $3 million strategic investment from Second Century Ventures (SCV), a venture capital fund of the National Association of Realtors (NAR).
The Seattle-based Symform’s cloud consists of local drive space contributed by its subscribers. So maybe you can say it’s in the realtor business because it sells real estate on customers’ storage.
But the real estate trade association didn’t fund Symform because it considers it a kindred spirit. NAR will make Symform subscriptions available to its 1.1 million members as a membership benefit. The realtors add cloud backup, and Symform adds to its subscription base.
Symform bills itself as a crowd-sourced cloud network. Organizations join the network by contributing extra local drive space in exchange for fast and secure backup. Besides providing backup, the network also synchronizes data from any device.
“We see ourselves as the Skype of data storage,” Symform CEO Matthew Schiltz said.
The startup’s technology encrypts data 256-bit AES. It breaks data into 64 MB blocks, divides each block into 64 1 MB fragments, and randomly distributes them across separate devices in the Symform cloud storage network. The architecture regenerates and redistributes missing fragments even when devices fail, according to Symform technical documents.
The technology is integrated with NAS storage devices from Synology, Netgear and QNAP. The company claims it has active users in 138 countries, up from 46 at the end of 2011. It also claims more than 7 billion fragments now are stored in the Symform cloud.
The $3 million is officially part of an $11 million Series B funding round Symform announced last April. It initially raised $1.5 million in seed money, so the company has raised total of $20 million in funding. Its previous investors include Longworth Venture Partners, OVP Venture Partners and WestRiver Capital.
Brocade completed its CEO search this week, hiring Lloyd Carney to replace Mike Klayko.
Klayko said last August that he would step down as soon as the board found a replacement, ending an eight-year tenure as Brocade’s CEO. During that time, Brocade acquired its main storage switch rival McData, outdueled Cisco for the top spot in storage networking revenue and spent $2.6 billion to get into the Ethernet market by acquiring Foundry Networks. But Klayko failed to attract a buyer for Brocade despite a great deal of speculation that the company was for sale several times over the past few years.
If you’re looking for hints on where Carney might take Brocade, two things about his resume stand out. First, he has little storage background and plenty of network experience. Second, he sold off two companies he ran – Xsigo Systems to Oracle last year and Micromuse to IBM in 2005.
Xsigo is the closest Carney has come to a storage company. Xsigo actually did I/O virtualization and was more of a networking play, but did work with storage gear. After Oracle bought Xsigo, it tried to recast its technology as software-defined networking. IBM acquired network management software vendor Micromuse for around $865 million, and Carney stayed with IBM for one year to run the Micromuse division.
He has also been COO at Juniper Networks, president of Nortel’s wireless internet division and a vice president at Bay Networks as well as CEO of his own angel investment firm. Carney obviously knows his way around Silicon Valley, which could help if Brocade puts itself up for sale again. If not, you can expect the vendor to continue its push to become an Ethernet network leader while holding on to the No. 1 Fibre Channel network spot for as long as that market remains lucrative.
There have been plenty of acquisition rumors around Brocade over the years, despite Klayko’s insistence in 2009 that the company was not for sale. Hewlett-Packard and Dell were believed to be considering buying Brocade before they acquired other networking companies, and there has also been talk of private investors buying Brocade.
A two-tier data archiving approach can help free primary storage capacity, reduce expenses from regular data protection, and meet compliance or business requirements for specific data.
A two-tier strategy divides archive data based on the probability of accessing that data. Archive data can be accessed online as one tier and a deep archive as another where access may be more involved.
An online archive has these characteristics:
- Data can be transparently and directly accessed by users or applications without other intervening processes.
- The time to access data is only nominally affected compared with primary storage, with no impact on users and applications.
- Typically NAS is used because the largest amount of archived data is in the form of files. There may also be support for objects depending on systems in use.
- The online archive has support for compliance requirements such as immutability with versioning of files, audit trails of access to data, and regular integrity checking of the data.
- The storage is much less expensive than primary storage.
- Only changed files are replicated for protection.
- Systems have built-in longevity with automatic, transparent migration to another platform. The migration is non-impacting to operations or staff.
A deep archive would have different characteristics than the online archive, including:
- Data moved to the deep archive is not expected to be needed again for any normal processing.
- Access from a deep archive may require greater time than applications than tolerate.
- Data may be stored in the form of objects with metadata about ownership and retention controls in order to permit massive scaling. The storage could be on local systems or in a cloud service.
- Longevity is handled automatically by the systems or service with transparent migrations.
- Compliance features are fully supported including digital data destruction.
- Protection is automatic with geographically separated replicated copies.
There is justification for a two-tier archive. You can gain large savings from moving data not expected to be used again to the lowest costing storage without compromising protection, integrity, or longevity. Economic models show the advantage and the compounding value over time as data is retained and more is added. Development of new systems and software that support object storage for very large scale of items in the archive and transparent migration for longevity are enabling wider usage. For all of these reasons, a two-tier archive is a good fit for a storage strategy.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
It’s the New Year and it’s time to organize things around the home. Throw out things that you no longer need. Take things that you probably won’t need but you’re not sure about, and put them in the attic. It’s a fairly simple exercise, but not done regularly. It happens now because the advent of a new year reminds us to do it.
This is similar to what storage administrators need to do this time of year. They need to organize their data, delete data that is no longer needed, and archive data that is unlikely to be used in regular processing. Maybe the archive target device should be called the “attic.” What they are really doing is making decisions about information. What is the value of the information, who owns it, what restrictions regarding compliance are there for that information, and can it be deleted?
There are several approaches to making decisions about the information. Some people don’t make decisions because there is no clear guidance and they might make the wrong choice. An interesting strategy from one storage administrator was to archive data to tape without migrating the data to new tape technology when the old generation of tape drives became obsolete. Eventually, there would be no drives left that could read those tapes and the administrator did not have to worry about the decision to delete the data.
Archiving has become a misused or at least misunderstood term. It really is about taking data that is not expected to be needed and moving it to another location. This is done for economic advantages, and potentially to meet regulatory requirements. Over time, the term has expanded to “active archive” and “deep archive.” Active archive is for data that is retained in the original context so an application or user can retrieve it without an intervening process. Deep archive is for data that is not expected to be needed, but may be. Both locations can support immutability, versioning, and other compliance requirements. There are advocates of using cloud-based storage for deep archive.
Managing information effectively includes archiving data by making intelligent decisions about what gets archived and where. The economic value of moving data off primary storage systems is great. An ongoing policy to move data periodically compounds that value. It should not take a trigger such as a new year to make a decision about organizing and moving unneeded stuff to the attic. It is a valuable IT process.
(Randy Kerns is Senior Strategist at Evaluator Group, an IT analyst firm).
Nearly half of employees use online file sharing services even though their companies have a policy against it, according to a recent report conducted by cloud storage vendor Nasuni.
For its Special Report on Shadow IT in the Workplace, Nasuni surveyed 1,300 corporate IT users. Online file sharing and bring your own devices (BYOD) are gaining in popularity as employees use PCs, iPads, smartphones and tablets to access work-related files. And they use these devices to work from airports, homes, cafes and other locations outside of the corporate office.
Companies are starting to call this “shadow IT” because these services and devices often are not controlled by traditional IT departments. For many companies, it’s a growing security problem.
“We were surprised that the percentage of users is one out of five. That is not a trivial number. There is an incredible amount of usage and it was higher than we expected,” said Connor Fee, Nasuni’s director of marketing. “We talked to a lot of IT guys and they complained about Dropbox specifically.”
Other services mentioned include iCloud and Google Drive. The survey included 300 IT managers and 1,000 employees. The survey found that one in five employees uses Dropbox for work files, and half of employees who use file sharing services do not know if their companies have a policy against it.
The report also found that corporate executives are the worst offenders, with vice presidents and directors most likely to use Dropbox despite the security risks. About 58% of employees with personal smart phones or tablets access work files from those devices. The survey found that 54% of respondents work at organizations that do not allow access to file sharing.
“The fact that corporate leaders are the worst offenders tells me IT is failing to deliver on something that is needed,” Fee said. “Having and educating about a policy is not enough. It needs to be address beyond policies.”
IBM has announced plans to acquire partner StoredIQ, an Austin, Texas-based software company that specializes in classifying and managing big data, particularly for regulatory compliance and legal reasons.
The financial terms of the acquisition were not disclosed, but IBM expects to close on the deal around March of 2013.
StoredIQ has been an IBM partner for two years. IBM’s Tivoli data management and archive storage systems are already certified to work with StoredIQ applications. The technology dynamically identifies and classifies data that either needs to be retained or decommissioned, while also governing who has access to what data.
Big Blue plans to make StoredIQ’s technology a part of IBM’s Information Lifecycle Governance (ILG) suite that has the ability to set management policies, mine and assess what data is valuable and what data should be deleted. The company’s software does not overlap with IBM’s current products in this area, said Ken Bisconti, IBM’s vice president for Enterprise Content Management (ECM) software.
“They have the ability to dynamically identify data that is in place,” Bisconti said. “Most vendors require companies to move files to a separate repository before it’s classified and managed. StoredIQ dynamically collects that information wherever it resides. We have the ability to collect data in a repository but we did not have the ability to dynamically collect it.”
Bisconti said IBM will retain StoredIQ’s staff, which they consider essential intellectual property. The company has 50 employees to date. Also, more integration will be done with the ILG suite. StoredIQ’s software already works with the ILG portfolio. IBM’s policy governance technology manages the overall plan of identifying high-value data versus low-value data. Those instructions are sent to the StoredIQ engine, which executes the policy.
“For example, StoredIQ can do e-discovery for data that is in place if you have to respond to a request for discovery material,” Bisconti said. “Typically, companies have a difficult time to get to data that was not moved to a repository.”
StoredIQ was founded in 2001 under the name of DeepFile Corp., a data classification company that created metadata to manage different types of unstructured data. The company later changed its name to StoredIQ and focused on compliance and e-discovery to help companies figure out what data to keep and what to delete.
IBM’s information lifecycle governance business is part of its software group.