Software, software, software….
It’s hard to go anywhere in tech these days without someone highlighting the virtues of software. Software is eating the world. It’s a Software-Defined Economy. It’ll run in your Software-Defined Data Center.
But it feels like there is a gap between the things that developers are using to change the economy, and the tools that fall into the definition of “Software-Defined”. There are the tools of the Devs, and then the tools of the Ops crowd. In the past, the IT community often thought of them as isolated, but that thinking is beginning to change. Examples of “Infrastructure as Code” and “Operations at Scale” are beginning to become more visible and common.
Even the definition of a “developer” is beginning to change. In the past, developer only applied to the people writing the applications that we interacted with as business users. This now applies to infrastructure and operations teams as well. Think of it as the evolution of the SysAdmin, but across all the functional areas of infrastructure and operations.
But there is still opportunity to do better. Making it easier for all kinds of developers to be able to use these new software-defined tools and environments. To eliminate all the friction between them getting the software and making it useful; integrating it into their environments and making it part of their continuous application builds. Just making software free to download is not enough.
- Is the software available via GitHub?
- Are the APIs well documented?
- Does sample code exist across a broad range of languages and frameworks?
- Can a developer run the software locally?
- Can the software be up and running in minutes, using the native deployment tools of the developer?
- Is there an online community to answer questions for developers?
- Is there an online environment that allows the developers to validate that the software can integrate and scale beyond their local machine?
- Does the software creator actively participate in the community, both online and in real life?
Maybe you clicked on this link and thought that I was going to write about the OpenStack Marketplace that was launched during the OpenStack Summit in Atlanta. It’s a natural progression for the OpenStack community to drive awareness of applications and services. But that’s not the marketplace I observed in Atlanta last week.
btw – In case you missed OpenStack Summit, here are all the videos – lots of great technical content and discussions.
The marketplace I observed was the reality of the IT marketplace, with vendors beginning to make moves and announcements which show that OpenStack is no longer a dream about “interoperable open clouds”, but instead is just another set of tools, products and APIs that will vigorously attempt to compete for the hearts, minds and wallets of IT professionals, developers and systems integrators.
This isn’t to say that the OpenStack dream is fading. It was alive and well in keynotes which drew analogies to the Star Wars Rebel alliance, fighting against the Evil Empire and the Death Star. But diving deeper, we saw that there is still division amongst contributors and developers about what should really be considered “Core OpenStack”. Will developers still actively and passionately work on projects if they are not considered “core”? Will some vendors try and claim to be “more core” in their distributions or offerings than others? Continued »
The big trend with Cloud Computing providers these days is a focus on “Enterprise” customers and workloads. We see this from AWS, VMware, EMC, Cisco, Microsoft, IBM, HP and others. Regardless of how each of those companies define “Enterprise”, this means that they are now subject to the rigors of TCO calculations prior to major purchases being made. Welcome to the wonderful world of Enterprise IT.
Now for the fun part. Building a TCO model that properly explains how much the offering will cost over a standard period of time – typically 3-5 years. Having built several of these models in the past, I know that they are always held up to huge amounts of scrutiny. Whether customers think you’re trying to hide or mislead them on costs, or whether customer believe you don’t understand realistic use-cases, no two TCO models are ever the same. They all make assumptions and they all have to make tradeoffs between completeness of information and usability (eg. too many inputs, too many calculations). With that in mind, let’s take a look at a recent attempt at a TCO calculator by AWS, comparing their server vs. traditional data centers.
Let’s begin by asking a few basic questions:
- What type of use-cases or applications does the TCO tool allow to be modeled? (eg. redundancy, security, availability, performance, etc.)
- What assumptions does the TCO tool make about how costs are calculated? (eg. cost of equipment; cost of IT resources; efficiency of IT resources – both people and equipment)
- Does the TCO tool make reasonable comparisons between each model? Too many tools compare one approach’s “best” vs. another approaches “worst” scenario.
- Does the TCO tool surface all the assumptions of both approaches, or just the approach of the tool originator?
With the announcement of the release of OpenStack “Icehouse”, the 9th version of the series of OpenStack projects, it’s now time for the community to focus their attention on the 2014 (Spring) OpenStack Summit in Atlanta, GA. This is an opportunity for the OpenStack Foundation to provide their version of the “State of the Stack”, by highlight customers using the software and interesting trends in the marketplace. It’s also a multi-day set of design sessions for engineers involved in the “Juno” (J) release, which is targeted for Fall 2014. Last but not least, it’s a massive recruiting and networking event, where every company has a vacancy sign out front for anyone with OpenStack-centric skills or experience.
This year, I’ll be interested to see a few things:
The Marquee Customer
There have been rumors for over a year that a Fortune 10 customer would be announced as a huge OpenStack user, for production applications. That name hasn’t emerged yet. I’ll be looking for big names building internal Private Clouds on OpenStack, and at least a few names using Service Provider clouds powered by OpenStack.
OpensStack has long been touted as the alternative to AWS (and often VMware), with the promise of open-clouds for customers. I’ll be looking for progress on overall interoperability between BOTH cloud providers and between distributions of OpenStack. This is still enough of an issue that Red Hat’s General Manager of Open Hybrid Cloud Programs (Alessandro Perilli, @giano) took to Twitter to highlight his belief that start-up OpenStack distributions (or products) should not be trusted by customers because the risk of start-up failure is too high. Indirectly, this is also a statement of the interoperability of various OpenStack distributions, if he’s claiming that switching costs could run into the $$ millions. Continued »
[A better title might be, “How to Drive Yourself Crazy Thinking About Your Career in Tech”]
I wasn’t around when the horse and buggy era ended, but sometimes I feel like I’m having the same conversations that the blacksmiths, saddle makers and stable sweeps had back in the early 1900s. Fast forward a hundred years and this 3rd platform transition has a lot of people wondering what to do next.
Let me step back and put some of this in context. During the late 1990s and early 2000s, the Internet changed everything. Where we work. How we work. The importance of technology for business success. And because of technology, very few of us have the same career path that our parents had before us.
- Companies are no longer centrally located, with their workforces being globally dispersed.
- Companies no longer maintain all their own internal services, often using resources available over the Internet to achieve the best mix of price and service.
- Companies no longer provide long-term pension plans (except some government agencies), so employees rarely stay at the same companies for their entire career.
- Customers (buyers) no longer get all of their knowledge from their suppliers. The Internet allows thousands of voices and opinions to impact their knowledge and buying decisions.
- Customers (buyers) are increasingly looking for new sources and consumption models for the technology that will help their businesses.
- Public Cloud and Open Source Software are moving the margins and control away from vendors and onto parties that can more directly control their future. The supply chain is shifting and with it is the location of value-capture.
So what does this have to do with Cloud Computing? Most people would tend to think that the majority of my days and nights are filled with deep, enlightening conversations about the topic. I wish. In fact, I am spending more and more of my time with people asking what’s next, specifically as it relates to their career in technology.
The Rise of the Rockstars
No phrase in our industry drives me crazier than “He/She is a Rockstar!” Why? A couple reasons:
- Most bands aren’t made up of dozens of people on stage; instead it’s just a few. So most people are either roadies or groupies.
- Rockstars travel. A lot. That’s part of the job. The big money is in being close to the customer. No travel, no rockstar.
- Not all rockstars make the big money. There are plenty that are only headlining at small venues in 2nd or 3rd-tier cities.
- Except for a few rare exceptions, Rockstars burn out in a short period of time.
- Rockstars begin to expect indulgences that lead to unrealistic expectations. Continued »
The “Five Computer” Theory
In 1943, Thomas Watson said, “I think there is a world market for maybe five computers.“ Then in the early 2000s, when the Web 2.0 era was starting GigaOm- resurrected the idea (sorry, can’t find the link, circa 2001-2003), but this time it was five clouds – Google (Search/Ads), Salesforce (Business), Amazon (Retail) and two others. In the first case, it was well before the PC and Smartphone era. In the second case, it was believed that we’d see massive consolidation of web services as companies struggled to figure out business models beyond freemium or backed by Google Ads.
Since then we’ve seen the rise of many “giant computers” – AWS, Google Compute Platform (and all the Google properties – Search, YouTube, Gmail, Maps, etc.), Microsoft/Windows Azure (and their other online properties), Twitter, Facebook, WebEx, Ebay – as well as the rise of many businesses built on top of those computers – Dropbox, SnapChat, Zynga, Box, NetFlix, etc.
We know from the Gartner IaaS MQ that it can be very difficult for smaller companies to catch up to AWS. But this year, I suspect the chart will be somewhat different as Google Compute Platform and VMware vCHS are added, and Microsoft Azure has expanded. Will we begin to see rapid consolidation of the “giant computer” market, with more businesses being built on top of the leaders? Or is still too early to see how the evolution to the 3rd platform will play out?
Opportunities for DevOps Consultants
My friend Jeff Schneider (MomentumSI, Transcend Computing) recently wrote a nice blog on how to evaluate consulting companies that focus on Cloud Computing. More and more, I get questions from my friends working at VARs (and vendors) trying to figure out how to get better prepared for DevOps (tools, methodology, skills, etc.). I believe there is a huge opportunity for the DevOps focused consulting companies to target VARs and existing vendors with training that would accelerate the understanding of this space. Some might argue that this could eventually eliminate the consultants existing business, but I believe the opposite would happen. They’d see an acceleration of companies looking to adopt the skills/tolls/methodologies, and they would continue to be the leading experts. The VARs (and vendors) would bring them into customers at a pace they have never seen before, because of their local presence and existing relationships. The gap between the consulting experts and the average Dev or IT team is huge. There is opportunity there to shrink the gap and significantly expand the market demand.
Can anything slow down AWS?
I get asked this all the time. Watching them continue to innovate and reduce prices, it can often be difficult to image a scenario where this happens. But there are plenty of possibilities: Continued »
The week of March 23rd, 2014 might go down in history as 1 AG (After Google). By launching (or re-launching, who can really keep up), the Google Cloud Platform (GCP) officially created a public utility cloud marketplace.
What? How can that be, considering that there are lots of recognized Cloud providers? While it’s true that all of these companies exist, and many run excellent businesses for specific market segments, most of them aren’t focused on the utility elements of the market.
On the 8th day, the Internet was created. And with it came (basically) seamless ability to access and move information around the big tangled mess of Intertubes. Apparently those TCP/IP inventors were pretty smart. And apparently ever since that day, or whenever Cloud Day 1 was, people have had the belief they could coordinate to do something similar for applications, application-containers and all of their associated data. They like to draw analogies to how email works, or how your mobile phone roams from carrier to carrier. And hence, there is the dream of a Federated InterCloud of Hybridness.
This past week, the market rumbled about how Cisco was planning to spend $1B to create the InterCloud. Since $1B is the new starting point for making cloud announcements or investments, this caught people’s attention. Also, because some media misunderstood it to be that Cisco was actually going to get into the Cloud business and directly compete against AWS, GCP, Azure, Rackspace, etc. That’s not actually the case. This about selling equipment to lots of SPs, with the hope that they will interconnect and allow companies to freely move between them. InterCloud’ing at the network layer, with a means of interconnecting with public cloud APIs.
But Cisco is by no means the first to drive this concept. Let’s take a look at who else has been driving this.
One of the most interesting things about doing The Cloudcast (podcast) is the variety of topics we get to discuss and the perspectives from people across our industry. Over the past few weeks, we’ve done shows focused on the evolution of PaaS, trends with AWS and public cloud, and how large-scale web companies scale. Regardless of whether the perspective was about large Enterprises, using Public Cloud providers or running large-scale operations, every one of the guests repeatedly brought up the growing demand for DevOps skills.
Let me step back and clarify a little bit. Saying a company needs “DevOps skills” is sort of like saying that a company needs to buy “Cloud products”. Both Cloud and DevOps are terms used to describe operational models. And in most cases, they are intertwined. Cloud tends to be more focused on self-service, on-demand resources. DevOps is focused on agile development and agile operational models than are tightly integrated. Both are focused on increasing the velocity that applications can be deployed (and updated) to help a business reach their goals.
There were a few interesting/insightful comments from those shows that got me thinking quite a bit:
“While many companies will use tools like Puppet or NewRelic (or whatever) to deploy to AWS, they have alot of skepticism that their tools companies will be around in a few years. So these companies are encouraging their people to really understand the mechanics of how they deal with operations and deployments. Over the next 2yrs, DevOps skills will be the most highly in demand, especially for anyone that considers themselves in IT Ops today.” – Mat Ellis (Cloudability)
“The problems within an Enterprise are no different than the problems of a start-up, especially if you buy into the notion that every successful company over the next 10yrs will have to have a significant involvement with software (across any industry).” – Mark Imbriaco (GitHub) Continued »
When I got into the technical side of the IT industry back in the 1990s, there was a company just down the street that was starting to gain some traction – RedHat – and they were pushing this new variant of Unix that some of my more experienced colleagues thought was interesting. Most of them already had deep UNIX backgrounds, so they were excited about the idea of a free version that ran on x86 hardware. I was less interested at the time because I thought the things a Cisco IOS box could do were pretty interesting and gave me more than enough to study for my pending CCIE.
Flash forward 15+ years and all that LInux stuff is all over the place as more capabilities that used to be in dedicated devices has now moved out into various LInux-based products and open-source distributions. Whether it’s in the networking space (OpenDayLight, Open vSwitch, OpenStack Neutron, Cumulus Linux, Quagga) or the storage space (CEPH, OpenStack Swift/Cinder, RedHat Gluster, etc.) or the Cloud management space (CloudStack, OpenStack, Cloud Foundry, OpenShift, etc.) or the underpinnings of modern application development, Linux is behind the pace of change in Cloud Computing. Continued »