There’s probably no more exciting, imagination-capturing branch of military research than DARPA (Defense Advanced Research Projects Agency). They’ve brought us robo-hummingbird spies, self-driving hummers and, last but not least, the Internet.
So they can be excused for wanting just a little something in return: To be able to use their iPhones and Androids securely. From a recent Request for Information:
The primary purpose of this RFI is to discover new technologies and methods to support full disk and system encryption of the CMDs (specifically Apple and Android platforms) to include a pre-boot environment to load the operating system. The solution must use an AES-256 bit encryption algorithm compliant with FIPS 140-2 as published by the National Institute of Standards and Technology (NIST). In order to meet this objective, DARPA extends an invitation to industry and universities to submit a whitepaper with ideas/concepts that describe an innovative existing technology approach that can be deployed in less than 90 days.
Currently only Blackberries and high-end secured phones are allowed in many DoD environments, meaning Angry Birds is out. It sounds like DARPA is looking for a full-drive encryption bootloader to pick up where the consumer-friendly Droids and iPhones have left off. To be fair, Apple has beefed up its security offerings in recent iterations (with a few nay sayers), but the business need isn’t new. In my time reporting on mobile devices, I’ve heard any number of security schemes to get around security concerns: Everything run as SaaS, with no sensitive local data stored; A specialized encrypted card that held or encrypted and decrypted the data; and a number of virtualized environments that sat (supposedly) securely inside the everything goes-consumer devices.
It will be interesting to see what DARPA picks: Freedom of Information request, anyone?
Open-source software isn’t anything to write home about anymore, but hardware design is less transparent. Facebook is changing the game today with its Open Compute Project, designed to share the specs and design of the custom servers built for Facebook’s data center in Prineville, OR. As Larry Dignan pointed out, this is a symbiotic move on the social networking company’s part:
In many respects, Facebook is open sourcing its data center and server designs. Jonathan Heiliger, vice president of technical operations, said the Open Compute Project is a way of giving back. It’s also a way to get vendors with more scale to incorporate Facebook’s designs to meet its needs with cheaper systems.
Despite what it means to vendors who may be scrambling to replicate Facebook’s designs and solutions, the Prineville data center’s PUE rating speaks volumes: 1.07 PUE versus the 1.5 PUE average, and the 1.4 – 1.6 PUE of Facebook’s leased data center.
- Vanity-free packaging with little-to-no screws, which resulted in 22% fewer materials, and 6 lbs less in weight.
- No duct work and no air cooling system in the data center.
- Dual AMD Opteron® 6100 Series socket motherboard with 24 DIMM slots.
- Intel Xeon® 5500 or Intel Xeon® 5600 socket motherboard with 18 DIMM slots.
- Localized uninterruptable power supplies serving six server racks.
How do you see this changing industry standards or vendor offerings, if at all? Share your feedback in the comments section or send me an email at Melanie@ITKnowledgeExchange.com.
You know what Dell thinks is cool? A billion dollars, which is just about what it spent in its recently closed acquisition of storage virtualization trailblazer Compellent Technologies, which had been receiving industry nods for its top-of-the-line hardware/software one-two punch.
Dell has been making strong inroads in the storage market, but some high-profile defeats have stymied its progress, most notably an embarrassing loss (after early bold announcements) of 3PAR in a very public bidding war with HP. A Gartner Magic Quadrant report on the storage market noted “the failed bid to acquire 3PAR may add additional shadow on the existing relationship with EMC.”
But Dell isn’t licking its wounds. That $2.4 billion it didn’t spend was burning a hole in its pocket, launching a $2 billion cloud spending spree. All that dropped money might help position Dell nicely as businesses look to bridge the public/private storage cloud divide, particularly since it’s playing nice with VMware and Azure in addition to its various in-house options.
The big question is if all the pieces come together neatly enough to compete with the established players in the area.
DataCore Software, known for its storage virtualization software, has released a survey comprising over 450 IT organizations across North America and Europe, “The State of Virtualization.” The findings can be a little disturbing, especially to a company who creates a product that many medium and large enterprise IT orgs are leaving out of their virtualization plans: Storage. The study found that 43 percent had mistaken the impact storage would have on server and desktop virtualization or had shied away from a virtualization project because storage-related costs were too high.
Delayed virtualization projects weren’t the only downside to this apparent misunderstanding of storage virtualization costs; even among those who had already deployed server virtualization, 66 percent view the increase in storage costs as their biggest problem. Higher costs doesn’t mean higher quality, either. Almost 40 percent reported unhappiness with their storage infrastructure due to slowed or limited availability for applications. To exacerbate the decreased performance, 22 percent of IT admins feel locked-in to their storage hardware provider, with about 40 percent of respondents using two different storage systems from the same vendor.
In order to achieve the agile, cost-effective, and enduring IT infrastructures you seek, those old ties to physical storage devices must be broken, just like you’ve done with servers. To do so requires tackling the next “Big Problem” plaguing data centers today – dissolving the expensive and restrictive dependency on disk hardware.
Are storage cost forecasts keeping your company from virtualizing, or do you wish now that it had? Share your stories in the comments or send me an email directly at Melanie@ITKnowledgeExchange.com.
Despite growing knowledge and interest in server virtualization, storage virtualization seems to be lagging behind. The truth is that if you’re interested in or are currently deploying server virtualization, storage dilemmas – and budgets – can hold you back.
Former leader of research and development at VMware, Kieran Harty’s new company, Tintri, is aiming to take virtualization to the next level. When Harty understood that many companies hesitate to include mission-critical operations in their virtualization deployments because of storage performance issues, the seed for VMstore was born. Harty’s team members at Tintri have backgrounds in virtualization at VMware and Citrix and storage background at Data Domain and NetApp.
The goal of VMstore is to create a storage solution specifically for virtualized environments from the ground up, reducing complexity. While traditional storage accounts for 20 percent of a typical enterprise’s IT budget, virtualization budgets see closer to a 60 percent dedication to storage. Harty told the Computer Technology Review that “the key bottleneck slowing virtualization adoption is the legacy storage systems that were architected before virtualization was even a consideration. Our products are designed to help enterprises virtualize 80 percent or more of their IT infrastructure.”
From the delightful DrunkenData, a delightful post (well, translation really) by Jon Toigo on the inevitable Rites of Spring, including the coming of fresh flowers, melting ice and, what else, storage marketing hype.
It’s worth a read simply for the echoes of T.S. Eliot applied to the economics of single-sourced hardware, but in the thousand-plus word piece dives into another topic: Why is storage, one of the most commoditized of the IT Dark Arts, still so expensive? For years, we’ve been regaled by tales of storage de-duplication, cross-platform standardization, and cloud/shared/virtualized storage solutions that all promise to decimate costs, but budgets keep going up. What gives?
There’s the simple, obvious explanation: That more stuff is being stored, dummy, so even as price-per-gigabyte goes down, total costs go up. But Toigo writes that this doesn’t paint the full picture:
In 2011, roughly 40 years after the distributed computing revolution, the costs of distributed computing have only increased – especially in storage. Despite the onset of commoditization in drives and enclosures (all disks come from four OEMs these days, and all chassis from a half dozen or so enclosure makers), array prices have accelerated at a rate of about 120 percent per year.
Part of the explanation is value-add software that vendors insist on joining to proprietary array controllers. Another part is the failure of the industry to define truly standardized interconnects so that two switch providers can build products to a common standard with absolute certainty that they will not work together in the same fabric.
All these sexy management frills added to an un-sexy commodity service are keeping prices high and preventing real savings in the storage department. In the end, that helps make the storage market hot and steamy for acquisitions, but helping enterprises actually cut costs might make for a beautiful love story that lasts past one passionate night.
Michael Morisy is the editorial director for ITKnowledgeExchange. He can be followed on Twitter or you can reach him at Michael@ITKnowledgeExchange.com. Image from Flickr user zigazou76 and licensed with Creative Commons.
Well, folks, it seems that Windows 7 month is on its way out. To celebrate all we’ve learned this month, we’ve compiled some of the highlights from around the community (we’re suckers for nostalgic look-backs). March isn’t over yet though, so dive on in and share your Windows 7 opinions and experiences with us!
Don’t get blindsided by all of the hype regarding the download numbers of Firefox 4 and Internet Explorer 9. Feedback and insight from actual users is one of the best ways to gauge what enterprise tools are best for you and your company.
Since there’s been a lot of discussion about Windows 7 this month, it only seemed fitting to get the community’s impressions of Internet Explorer 9 from Microsoft, compatible with Windows 7 and Vista. We polled the IT Knowledge Exchange members and got some interesting and varied responses.
In addition to even further muddying its already confused cloudspeak, Microsoft’s latest and greatest management tool, System Center Configuration Manager 2012 (SCCM 2012), has a few handy tricks up its sleeves to help keep Redmond on top, even when the company is eating competitors’ mobile dust.
SCCM 2012 brings in new support for managing and configuring mobile devices including, as Mary Jo Foley reports, Android, Apple iOS and even Symbian devices. It’s a surprisingly agnostic move with interesting implications, as Foley notes:
My first thought when I heard that SCCM 2012 will enable management of iPads is that Microsoft may be undercutting its own argument that Windows-based slates and tablets are inherently better because IT managers can’t manage iPads as easily and well as they can Windows-based PCs and devices. If Microsoft makes enterprise management of iPads a key selling point, won’t the company be removing one of the advantages it claims for Windows slates?
Possibly, but maybe it’s also just responding to market demands, particularly since Windows Phone 7 OS will probably be a laggard in the field for a few years, at least (although it does also support Windows Phone 7 management), despite whatever catchy advertising campaigns they run.
Hat tip to Ron Miller.