The first one was an article by Ben Kepes (@benkepes) that called out a recent JPMC report that focused on the transformation of networking equipment. It highlighted that in comparisons between ‘whiteboard’ (or ‘bare-metal’) networking equipment and Cisco equipment, that support costs were adding a significant level of cost to the overall price.
“….However, one of the insights we come away with is that Cisco’s services fees and licenses significantly inflate their pricing. Over time we would expect these fees to come under the most pressure as Cisco is forced to compete with lower cost solutions….” – JPMC
The second one came from EMC with their announcement that the ViPR storage software would be offered for free, with community support. EMC is obviously not the first company (or community project) to release free software, but this is definitely a different path for a traditional Enterprise vendor to take with a new product. Make it easy to download, trial and use the technology. Update the wiki-based documentation as problems are found. Consult with others using the software, in real-time, on the communities. It’s not your grandfather’s Enterprise vendor engagement model anymore.
Still other activities are starting to take the place of traditional tech-support. Sites like StackOverflow are there to help software developers. Projects like OpenStack are driving community efforts around documentation and training, to augment existing tools like IRC to drive community-based support. And programs like #EMCElect, #CiscoChampions, and VMware vExperts are trying to identify their best and brightest, with some people taking this a step further and setting up various fee-based activities to capitalize on their knowledge and community building.
Instead of continuing to build (or outsource) large tech-support organizations, will we begin to see Support-as-a-Service groups start to emerge? For fee, the best-of-the-best, knowledge repositories and discussions to supplement or replace support over time? Can this be done in real-time as tools/software/products report their status and intelligent analytics systems seek rememdiation in real-time?
What do you see as the future of support in the Enterprise?
Many years ago, NIST set out to define Cloud Computing. In doing so, they laid out a stacked model of IaaS, PaaS and SaaS. The leading public provider is lumped into the IaaS category, but offers many PaaS-like services. Some people are saying that maybe PaaS is dead, or at least needs to be redefined.
But recently, we’ve seen more and more people saying that maybe there will be a layer in between the IaaS and PaaS defintions. As the ability to offer “everything as a service” has evolved, so too has the thinking around the concept of “IaaS+”. IaaS+ is the idea that it’s more than just provisioning VMs and the underlying infrastructure, but also a rich set of services that can be used in conjunction with the apps in those VMs. And it implies a more modular approach than the platform/frameworks of PaaS.
- Gartner talks about IaaS+: http://cloudpundit.com/2014/01/28/the-end-of-the-beginning-of-cloud-computing/
- Moving from IaaS to IaaS+: http://www.cumulogic.com/making-the-move-from-iaas-to-iaas/
- Defining why developers like AWS: “It’s the Services”: http://blog.ingineering.it/post/47329908147/why-aws-is-so-far-ahead
There are already companies like Cumulogic (podcast), Transcend Computing and others that are building these IaaS+ layers on top of various IaaS stacks. We’re also seeing projects like OpenStack start to consider this approach with projects like Trove, Marconi, and Savanna.
“To me, the main problem is that IT assumes that VMs “as a service” is enough for developers. But watch AWS, Google and MS. What “cloud services” are they innovating with these days? They’re focused on more than just VMs, that’s for sure. DB services, Queues, CDN, etc… Developers want these higher value services, and they don’t get that from basic IaaS (or even worse, virtualization farms called “clouds”). Until IT departments (1) understand the difference and (2) find a way to deliver these services, the developers will continue to use the clouds that *do* provide them.” - Chip Childers
There are obvious trade-offs between using a platform-centric approach (PaaS) and a decoupled-services-centric approach like IaaS+, especially for the groups that have to operate and maintain the systems. It will be interesting to see which approach accelerates and emerges over the next few years, especially as more of this technology gets adopted by Enterprises for new and modern applications.
Will an OpenStack Leader Emerge?
We’ve all watched the evolution of OpenStack for the past 3+ years, from incubation to massive open-source trend. Sometimes it’s been good and others times less so. Lots of commentary on how much code has been contributed and which vendors are part of the OpenStack Foundation. But will a true leader emerge in 2014, like RedHat did from the Linux wars of the late 1990s? Will we see strategic consolidation (eg. companies like Piston Cloud or Cloudscaling acquired)? Will someone buy the experience of Mirantis? Will an OpenStack powered cloud begin to challenge AWS in services, scale and pricing? Will Oracle or VMware be the catalysts to bring OpenStack to the mainstream Enterprise, or mostly just make announcements?
What Segment of SDN will Emerge?
“SDN” has been through so many evolutions over the last 3-4 years that sometimes it’s hard to even know what the terminology means anymore. Recently articles from various vendors highlight this (here, here). But just like the networking industry, SDN will have applicability in many segments of the network, and many layers of a network stack. Beyond the L2 “virtual switch” in servers, virtual networking still hasn’t established a strong foothold yet. Does it begin to emerge in the L4-L7 services space, or L2-L3 overlays in the datacenter? Do the NFV-driven business motivations of Service Providers occur faster than Enterprise IT changes their operational models? Or will Cisco’s market dominance prevent SDN from gaining significant market traction?
What has a Bigger Impact on Storage – HyperConvergence, Flash or Software-Defined?
Pure Storage gets $150M, Nutanix gets $100M, Nimble has a successful IPO and EMC/VMware launches ViPR, ScaleIO, XtremIO and VSAN. That’s a lot of VC funding for emerging start-up companies, balanced against the market leaders driving disruptive models of both the industry and their own portfolios.
Do Clouds become the new IT Silos?
In the past, Enterprise data centers were filled with application-centric silos. They were consolidated into a small number of locations, but they were still often unique, segmented and isolated. And then IaaS offerings like AWS began showing the market how to build environments that could host many types of applications on shared infrastructure environments. Within the last year, Adobe, Apple, Cisco UC, VMware, Microsoft, SAP and Oracle have all built their own Cloud offerings. Do Enterprise customers continue to build Private Clouds and leverage AWS, or do they flock to the vendor-specific public clouds? Or does better technology emerge to help IT organizations manage the hybrid offerings across all these clouds?
How do we measure PaaS Success?
Only a few years into it’s existence and some people are already calling it “dead” (or not dead), or misdefined. But yet, we’re still seeing large companies publicly put their support behind PaaS initiatives, both in the public cloud and within the Enterprise space. Or maybe PaaS is really old and we should be doing a better job keeping a scorecard of its progress. Will we see PaaS become a trackable market segment in 2014?
Another year, another set of predictions and opinions about Private Cloud. The list below is just a small spectrum of how believe the market is going to vote with their actions. The reality is that this was never really a debate. Even before the term “Private Cloud” got introduced in 2009-2010, that game was over. IT organizations were already using a broad mix of internal and external resources to deliver technology to their business. It was SaaS applications and packaged applications and homegrown applications. Inside the firewall and outside the firewall.
And it’s true that lots of IT organizations either have built or will build a variation on what they believe a Private Cloud should be. For some it’s about efficiency and cost-reductions.For others it’s about agility and ever-growing business demands. But the reality is that no IT organization is going to only have a Private Cloud strategy. They never did before and the competitive world won’t let them going forward.
The reality is we live in a hybrid world. When the CIO sits down at C-suite meetings and looks around the table, they find that every other organization has already adopted a hybrid model. Finance, Marketing, HR, Sales, Engineering. Every one of those organizations has already optimized their segment of the business to leverage the best resources for each departmental requirement, whether they are internal resources or external resources. In essence, they have already build a Hybrid Cloud organization, just not with servers, storage and networks.
The real challenge for IT organizations going forward is how to best manage the Hybrid Clouds they have created.
- How to continue to have visibility across clouds?
- How to continue to reduce friction between business ideas and execution?
- How to create flexibility so that they are at least marginally prepared for the next technology shift (what comes after mobile?)
You hear that?
It’s the sound of frustrated developers and SysAdmins. The groups that are on the shop floors of 21st century bit factories, trying to keep things running more quickly and more smoothly. The groups that have been in the trenches of the evolution known as DevOps. And they are frustrated. Not at the tools that are available to help them do their jobs. Those are plentiful and improving all the time (Github, Jenkins, Vagrant, Docker, Puppet, Chef, Ansible, Zookeeper, etc.). Not at the people that have been in the trenches sharing their experiences (Jez Humble, Patrick Dubois, Gene Kim, etc.). They are frustrated with the marketers that are trying to manipulate DevOps and turn it into another round of Cloudwashing.
It wasn’t great when the great Cloudwashing era of 2009-2012 came over us. Cloud-in-a-Box, Personal Clouds and all the other variations of “whatever we sell is now Cloud”. But at least IT Admins could point to basic elements of virtualization or automation and prove that some things in IT cost less and could be accomplished faster than before. It wasn’t really Cloud, but it was an improvement.
The same thing won’t happen for those trying to push Big DevOps or New DevOps or DevOps is the new Orange. Developers don’t suffer fools. They will ignore you. They will shun you and shame you.
Leave DevOps alone. The developers and sysadmins are making progress, learning on the fly and continuously improving. They don’t need your DevOps training wheels.
Three or four years ago, I had a unique opportunity to spend two weeks touring manufacturing facilities across China as part of a program to study complex systems, international business and patent law (or lack of it). We spent time with many well-known companies looking at their operations processes as well as discussing how they compete with low-cost labor and lack of Intellectual Property (IP) laws in China. While many of the stories can’t be repeated due to NDAs, there was one example that has stuck with me as a unique example of how to compete in a world where the competition plays by very different rules than your company.
[Side note 2: We talked about some of these shifting rules for technology companies on our recent podcast]
The story goes like this. Large (global) automative company wants to sell products throughout China. To compete on price, they have setup manufacturing facilities locally in several Chinese cities. Chinese workers are hired, trained and build automotive products. After a short while, near replicas of those products start showing up on the market with various Chinese brands, at 40-60% of the cost of the global brand.
So what should they do? Do they follow The Innovator’s Dilemma model and get out of those markets because some automotive products are becoming commoditized and let the low-cost competitors take that market? Do they continue to build their existing products, but double-down on marketing to educate the market on why their “value-added” capabilities are worth the price premium? [btw – does this sound familiar in a technology context?]
Actually, they did none of the above. What they did was embrace this new market demand – similar capabilities, but acceptable lower quality – and built a level of expertise around this new approach. The first thing they did was segment a line in the plant to building lower-cost “knock-offs” of their own products. They used lowered cost paint, single-pane instead of double-pane glass, lower lifetime tires, etc. They leveraged the experience they already had in manufacturing and willingly adjusted the product/material quality to satisfy a market that had lower expectations. The margins on these products were lowered (I don’t recall the difference), but it was viewed on getting $0.40-0.50 on the dollar vs. $0 on the dollar to lost business.
But it didn’t stop there. They expanded the number of manufacturing lines for these lower-cost products within the same facilities. They also took the customer feedback and considered it within the higher-end product lines. Ultimately they were able to leverage the expertise they already had (manufacturing, sales, etc.) and augment it with a line of thinking (and products) than satisfied an under-served segment of the market that would pay less for less capabilities, but still wanted some capabilities. Continued »
On-Demand, Self-Service access to IT resources. That’s the promise of Cloud Computing. It’s a reality that has been coming together for the past 5-7 years. At times we marvel at the pace of change that cloud computing has had on the entire IT industry, and other times we grow frustrated that broad adoption and availability of new services aren’t happening fast enough. So as I think about the year ahead, these are some areas where I believe we need to see improvements in 2014 to keep this industry shift moving forward at the pace we’re all expecting.
Move from Building to Buying Private Cloud
We all lived though the “Cloud in a Box” era of HW+SW packaging (sometimes just HW) that was targeting Private Cloud. While this market has continued to grow rapidly (and be renamed “Converged Infrastructure”), it’s only a piece of the puzzle. Those are building blocks. What we need to see is an evolution to “buying blocks”, where the entire solution can be purchased in a way that better aligns to the services that will be consuming those underlying resources. This improvement will require some basic assumptions (or economic changes):
- We may need to have consumption-based pricing of on-premise HW+SW
- We will need to see on-premise resources priced in a way that makes comparisons with public cloud services fairly straight-forwards (hourly, daily, monthly)
Enable the IT Product Manager by Default
While there is still some debate about what the appropriate (or inappropriate) HW+SW to use to build a cloud computing environment, most people agree that making changes to the associated people and process is the more difficult side of the equation. One way to improve this is to enable the IT-as-a-Service “Product Management” function by default. The service catalogs and portals need to be on by default. The basics of the menu need to be delivered as templates, inclusive of the service offering, the pricing models and the basic policies for usage/security/etc. Not every one can become a master chef, but many people can successful learn to make acceptable (and often delicious) meals by giving them a good cookbook filed with recipes. This is also needed for the groups that will be operating the new cloud services. Don’t just give them the tools, but also give them the recipes to enable successful consumption and usage of the services. Continued »
Until a few years ago, you could almost guarantee that every technology vendor pitch would include a claim about how they could help a customer avoid lock-in. It could be through open-standards or open-APIs or an SDK that would allow customization. And then the industry changed and avenues to acquire technology expanded to include public cloud computing and the mainstream use of open-source technologies. These new “vendors” weren’t the same names that had dominated previous generations of IT, so they could use vendor lock-in as FUD to highlight why their new capabilities we worth a look from the IT department (or directly by the lines of business).
Back in the day, technologists would wait until the committees at IETF, IEEE, SNIA, W3C or some other standards body had decided what was officially a standard. Standards were intended as a way to manage interoperability, but also control costs so that no single vendor could dictate a technology and hence hold a market hostage on prices and profit margins. It was the leverage that customers had against the technology vendors. Interoperability was customer’s leverage against lock-in.
Somewhere along the way, may people seem to have forgotten what lock-in really means. They have forgotten that it happens every time any new technology is downloaded, deployed or implemented. Lock-in is the cost of making an initial decision + the cost of learning that new technology + the cost of maintaining that technology + some (potential) future cost of changing that technology. There are other “opportunity costs” that can also be factored in, such CAPEX vs OPEX (eg. public cloud) or agility costs of being able to modify the technology directly (eg. open-source), but those are really just variations on maintenance costs. Continued »
It’s that time of year. Time for that time-honored technology tradition of pontificating bloggers making their brilliant 2014 predictions. The time when we dazzle you with our prognostication skills and crystal clear view of the future.
But the big challenge is we’ve all heard these before. This same problem happened to everyone’s favorite game show, Wheel of Fortune, as people got tired of contestants guessing the same letters in the Final Bonus Round. So to spice things up, they automatically gave the contestants the letters R-S-T-L-N-E, allowing the viewers to bypass the monotony of repetition.
For you, my loyal readers, I’m going to give you the R-S-T-L-N-E of 2014 predictions, so you can bypass all those other blogger predictions. You can thank me later…
R – 2014 will be the Year of VDI. Write it down. Tattoo it on your back. The technology has finally come around and it’s ready to move the ball to the finish line.
S – <Insert_Blogger’s_Company_Focus> will lead the industry in a radical technology shift. In the process, the legacy competitor will have no way to keep up with their world-class technology and overall company nimbleness.
T – Software-Defined OpenAwesomesauce will signal the death of 2005’s leading technology. It’s finally dead. Carve the tombstone.
L – Open Source will win. It’s free, like beer, or puppies, or other free stuff like lollipops at the dry cleaners, and free always wins. And if your company doesn’t open-source all of it’s technology then you will lose.
N - There’s a 50/50 chance that your company will abandon ITIL and embrace DevOps, while converting all of your infrastructure to whitebox hardware and you’ll only hire people in IT that can write production code on Day 1, just like they do at Facebook. All of your internal company meetings will be replaced by hackathons.
E - Adrian Cockcroft (@adrianco, Netflix) will be highlighted (or keynote) at many Cloud Computing events. This will be balanced against Oracle announcing that they are now supporting a technology or partnership that they condemned just 2 years ago. Continued »
After 10+ years of limited change in the networking industry, we entered 2013 with a great amount of fanfare and promise with a new set of networking technologies to address the challenges brought about by server virtualization and web-scale data centers. The $1.2B acquisition of Nicira by VMware, considered one of the leading pioneers of SDN, got people thinking that maybe there was a chance that leading architectures and marketshare could be changing over the next few years. It also signaled that major companies believed in this new networking paradigm and were willing to put significant dollars behind the SDN trend.
Then things got interesting…
Big Switch Networks had attempted to capture the early marketshare by open-sourcing their Floodlight controller, but this was muted when the OpenDaylight project was launched by the Linux Foundation. Major vendors including Brocade, Cisco, Citrix, Ericsson, IBM, Juniper Networks, Microsoft, NEC, Red Hat and VMware joined as founding members. Both the Floodlight and OpenDaylight controller were options, but since then Big Switch has dropped out of the project. Juniper eventually filled the gap of alternative technology when it contributed the OpenContrail project.
And then things got even more interesting… Continued »