The Troposphere

May 12, 2009  8:40 PM

Cloud management startups vie for VC dollars

CarlBrooks Carl Brooks Profile: CarlBrooks

Cloud computing is the fresh word on everyone’s lips; the push started in earnest a year or more ago, when Amazon’s Web Services portal came into its own and the reality of instant-on, pay-as-you-go infinitely scalable servers sank in. Now companies that cover the middle ground between your minute-by-minute rental server and the software that actually makes you money are hot topics.

Needless to say, venture capitalists are hungry for cloud companies. Diego Marino of Spanish startup Abiquo says that VC firms have been beating the door down for attention since Under the Radar 2009, April’s startup showboat conference in Redmond, CA. “They are knocking”, he said.

The slump in the US economy has probably done more to stimulate cloud interest than any amount of hype: companies from large to small that were mulling the choice between buying iron or renting cloud space have had to face up to severe bottom-line pressures. The practically zero time and investment costs involved in using cloud and the growing awareness of reliability have tipped the balance for cloud, and big enterprise is slowly but surely hoving over to using cloud technologies in their own data centers to increase utilization and flexibility.

And thus companies who can help manage, deploy and use all these new technologies are all the rage. Here are five shiny new shops to know about, in alphabetical order:

Open source cloud development and management software: Abiquo released AbiCloud in April, has $400k in funding and prospects for more very soon. The company expects to make money giving away the software to developers and charging for support and development. AbiCloud currently supports VirtualBox, VMware ESX & ESXi, Xen and KVM.

Stripped down, intuitive management tools for EC2 instances in a browser format; Cloudkick is angling for admins who want a clever way to keep on top of their cloud. Co-founder Alex Polvi says that they started with $20K in seed money from Y Combinator and coded the project in Python in 5 months. He says the user base is growing fast and “It’s a logical progression” for users to want to buy premium services once they are hooked on the free service. It will be sold just like server instances- by the minute.

Indicative of the current heady state of the marketplace, enStratus started as product development and new features for Valtira, a marketing platform. Valtira honchos David Bagley and George Reese spun enStratus off when it became clear they could use their tools for distributed systems in general and are aimed squarely at companies moving into the cloud. They too, are optimistic about funding.

Eucalyptus Systems:
The bunch who developed EUCALYPTUS, an open source cloud platform out of UC-Santa Barbara are cashing in by offering to help IT shops start build their own data cloud, fully compatible with EC2 and Amazon’s APIs. They have $5.5 million from Benchmark Capital and the support of well established players in cloud, like RightScale. They are head-to-head with Abiquo, but the world is big enough for two free cloud –building software projects, right?

Tap In Systems:
Running from and sold from and used from inside Amazon’s cloud, Tap In is an early example of what will eventually be a crowded ecosystem of businesses inside EC2 space. Built around open source Nagios, Tap In promises support for an increasing number of vendor APIs in the cloud and recently announced partnerships with public cloud providers GoGrid and 3Tera. CEO Peter Loh says they have angel money right now, and they are targeting enterprise users.

May 7, 2009  9:04 PM

Rich Wolski on the difference between data centers and clouds

CarlBrooks Carl Brooks Profile: CarlBrooks

I’m working up an article about EUCALYPTUS and Eucalyptus Systems, cf. my earlier post on’t.

Leaving aside the giggles over nomenclature, I had quite a nice talk with Dr. Rich Wolski, the lead scientist on the original open source project (also with CEO Woody Rollins and the VP of marketing).

Anyway, Wolski had an interesting and quite succinct definition of the differences between a data center employing virtualization in its currently accepted form and a cloud infrastructure, because, on paper, the two share enough common elements that lots of people (and marketers) are happy to fuzz the two together.

I didn’t think that was quite right, and neither does the Good Doctor, (ha! like I’m the expert over here). Otherwise, why get excited over cloud? If that were true, then it’s just re-packaged old news, and nobody needs to do anything but change the badges and maybe dice up the trim package, if I may borrow from the big book of automobile industry metaphors. But cloud is fundamentally something different, and new, and it’s worth knowing why.

He says it’s down to access and the control structure. The major difference between a data center and a cloud is access- a cloud is set up so anyone can drop a penny in the slot and start up a server or six- in a data center, you ask, and someone does it for you, then hands over the steering wheel.

He said something like, pardon the paraphrasing, “in a data center, virtualization is the grease that lubricates resource management” for the admins; it allows the guy in charge to move his resources around- “it’s a reconfiguration mechanism,” but “in a cloud, [virtualization] is a fence.” it separates and protects resources and lets everyone have their own private playground without knocking over the other kids’ toys.

A subtle difference? Wolski says it’s down to a bedrock set of premises and assumptions that drove the development of the cloud model.

“We tried to look at the cloud paradigm from an analytical perspective,” he explains, and “cloud is an ecommerce model-it’s a transactional model [in a] distributed system”.

Did that sink in? At its most basic, cloud is not about computers. It’s about sales. Start with the premise that you have a product (server instances/CPU time/bit buckets), you want to sell them to any and all comers over the internet (ecommerce) and you want to do a lot of it. What you get is “cloud computing”, and logically, it’s no surprise that Amazon pioneered it commercially. They didn’t assume they needed a resource farm and a way to sell access to it- they assumed they had servers and needed to sell those instead.

So that’s interesting to me. Cloud is not a utopian access opium dream- it’s a logical outgrowth of commodity commerce.

UPDATE:Story’s done, look for it soon. One more little nugget from Wolski on defining cloud computing: When he and his crew started thinking about the project in 2007, they found it was “utterly impossible to get a consensus” on what cloud was, so they “decided to sidestep the debate by picking the thing that was demonstrably was a cloud — the one thing no one could say was NOT a cloud.” Their answer? Amazon Web Services.

So there you have it – not sure what an elephant is? Look around; you can’t miss one if it’s in the room with you.

May 6, 2009  1:40 AM

Citrix and Amazon hop into mindbending infinite-loop bed together

CarlBrooks Carl Brooks Profile: CarlBrooks

We’re sensing someone had an office pool going on “alliteration they could get away with” for this slug line: “Citrix Announces Citrix C3 Lab Built on Amazon Web Services to Connect Cloud Computing to the Corporate Datacenter”.

Anyways, what the mess of ess’ and clatter of consonants means is that Citrix, as part of its big show this week is announcing their Citrix’ XenApp software is now available for rent, from AWS, to run Citrix servers on Amazon’s cloud.

C3 stands for Citrix Cloud Center, their management suite for IaaS providers.

That’s the Citrix Citrix Cloud Center Lab, for those who collect poorly thought out product names.

The brain-pain part comes from the fact that Amazon itself runs its cloud on Xen, so a customer of Citrix Citrix would, technically, be using XenApps to use Xen virtual machines to manage and deploy public cloud resources to their customers as a customer using public cloud resources running on Xen virtual machines.

Hey, that may be great and just what C3 users want and it may work just fine, but it’s a hell of a rabbit hole to send your data down, ontologically speaking. We’re not prejudiced, we’re just trying to keep up.

Also, according to the article published today by News Director Alex Barrett, you’ll be able to do it all from your iPhone. I think they should call the app, “DRINK ME”, but no-one reads anymore, they’ll never get it.

This is doubtless targeted at developers who want to play with Citrix distributed computing offerings, and the odd IT shop that wants to offer Citrix AND call itself “cloud” and can’t afford its own data center.

Really, though, it’s just another way of proving that Amazon’s cloud model is still the top of heap. Oracle, IBM, now Citrix, and others are essentially opening Amazon storefronts, only it’s buzzword-friendly grid computing instead of baby clothes and lawn sprinklers. Amazon doesn’t care- they have a two and a half year leap on everyone else wanting to offer public cloud instances, the best distribution channel extant and plenty of headroom. All they have to do is make space, keep the lights on, let their little playfriends from the middleware/manageware classroom play in their pool, and rake in the dough.

I’ll scrape together some time soon and poke around, see how good a job Citrix has done making this work and post some screenshots and an update in a day or so. Just a head’s up, ladies and gentlemen: I may never been seen again.

May 4, 2009  8:58 PM

Cloud Security Alliance in favor of open standards

CarlBrooks Carl Brooks Profile: CarlBrooks

Very interesting conversation with Nils Puhlmann, a co-founder of the Cloud Security Alliance originally about the DTMF initiative for open standards, but went out and about a bit; here are three nuggets which I was interested to hear about:

Security: Puhlmann is of the opinion that a) as much transparency as possible would have direct benefits to cloud providers, since “If you do everything well, why would you not want to show your customer?”.

He feels that customers would actually be more likely to buy into vendors that could show best practice security under some kind of standards model, since it would free customers from glacially slow and costly audits and testing- enterprise could buy into a public cloud without a hitch under the right conditions(read: bomb-proof security standards) and b) “within 12 months we will see many things in cloud security that will have completely failed,” either through backing the wrong horse in terms of security model or through market forces that ebb away from a chosen track.

He thinks that consensus on cloud security will emerge in baby steps as the marketplace learns what works and what doesn’t and what pays off and what doesn’t.

Compliance: Puhlmann says, “What we see for some companies,” that have regulatory oversight, “are compliance rules that rely on the notion that you have complete control over your data,”which, if you are using public clouds, is patently untrue.

But enterprise wants to use public clouds, small and midsize companies that interact with regulated agencies will want to use it; Puhlmann points out that the $19B Electronic Health Records initiative, for instance is “simply not going to happen without cloud” technologies.

FIM: Puhlmann raises hopes of a universal federated identity model, since as data gets more and more distributed, “a good federated identity standard would provide the means to track and control who has access to your data across private enterprise and the public cloud.” and believes it’s a problem that remains unsatisfactorily addressed, cloud or no cloud, but this drive toward IaaS might provoke a more catholic solution.

And the long and short of addressing a lot of these concerns lies in the hands of the agencies that regulate so much of the data about us personally.

Puhlmann thinks that dollars and sense are going to come to a head much quicker than many anticipate, since the poor economic climate is driving an awful lot of fence sitters off the palings and into the clouds, and then,”economic pressure will become so immense that regulators will have a big lobby standing behind them to force them to act,” to catch up to cloud technologies and enact regulations that allow controlled data to exist in public cloud infrastructures.

And who am I to doubt him? A reporter, so of course I doubt him, and this is well-considered analysis, not factual reporting, but from a common-sense perspective, everything he says holds water, and it’s going to be terrifically interesting to see what happens with the cloud rubber meets that federally regulated road.

April 29, 2009  8:47 PM

EUCALYPTUS sprouts business shoots

CarlBrooks Carl Brooks Profile: CarlBrooks

In what is surely a contender for the most-complicated backronym of the ages, the Elastic Utility Computing Architecture for Linking Your Programs to Useful Systems, first developed at Middleware and Applications Yielding Heterogenous Environments for Metacomputing, UC- Santa Barbara (MAYHEM in case you weren’t seeing the silly nomenclature trend) has jumped the public lagoon for the open waters of commerce:

Eucalyptus Systems, formed out of the team that developed EUCALYPTUS has gone commercial to sell help with their open-source software to cloud-minded types.

EUCALYPTUS (fingers…getting…so… tired) is a set of open-source software tools that allow users to interact with and deploy AWS-type clouds. You can use the software to administer your EC2 images or create your own private data cloud along the same model. Pretty neat, and free, but what happens when AWS decides to re-tool their APIs? Questions soon to be answered.

read the press release here.

watch this page for updates.

April 4, 2009  1:27 AM

Amazon’s New Elastic Map Reduce

JohnMWillis John Willis Profile: JohnMWillis

Amazon has offered a new web service that provides Hadoop services called Elastic Map Reduce.  Hadoop is a Java based framework that implements Map Reduce.  Map Reduce is a method of programming that gives a program the capability of breaking a job up into hundreds or even thousands of separate parallel processes.  The idea is that you can take a simple process (like counting the words in a book) and  break it up into multiple running parts (i.e., The Map), then collect them all back into summary counts (i.e., The Reduce).  This allows a programmer to process extremely large data sets in a timely manner.  Hadoop is used by companies like Google, Yahoo,  AOL,  IBM,  Facebook, and to name a few.

The new Amazon Elastic Map Reduce service is another pay-as-you-go service that starts at .015 cents per hour and can go as high as .12 cents per hour.  This is an additional charge on top of your standard EC2 and S3 services.  For example, if you use the Hadoop to read data from S3 and start up 4 EC2 instances, you will be charged the original costs for the EC2 and S3 usage plus an additional charge for the Elastic Map Reduce service.  Amazon is charging you for the setup and administration of Hadoop – as a service.  The Elastic Map Reduce service is extremely simple to set up and run.  Basically, you upload your application and any data you wish to process to an S3 bucket.  You then create a job that includes the location on S3 of your input and output data sets and your map reduce program.  The current implementation supports writing Hadoop Map Reduce  programs in Java, Ruby, Perl, Python, PHP, R, and C++.  You then also configure the number of EC2 instances you want to run for the Map Reduce job.  You can also add some advanced arguments and use more complex processing methods, if you choose.  The AWS Management console has been updated to support a new tab for the Elastic Map Reduce service.  This new service hides all of the system administration complexities in setting up a Hadoop environment which can be quite complex.  A hadoop setup of multiple systems is not a simple task.  Hadoop runs as a cluster of machines with a specific file system called the HDFS (Hadoop File System).  It also has a number of servers called Datanodes (these are the job processing clients) and a master server called the Namenode server.  Here is a diagram of an HDFS environment:

This morning I had an opportunity to test out the new Elastic Map Reduce service using the Amazon Management Console.  I setup one of their sample jobs to do a word count Map Reduce.  There is an excellent video from Amazon on how to get started here.  Here are some screen shots of my testing this morning:

I also had an opportunity to speak to Don Brown, a local Atlanta entrepreneur and founder of, about the significance of the new Amazon Elastic Map Reduce service.  Don  pointed out two significant aspects to the new Amazon Elastic Map Reduce announcement.  First, by Amazon creating this new service,  they have put a higher level of significance on the usage of Map Reduce and Hadoop.  Organizations looking to explore new techniques for large data set processing and/or new ways to do data warehousing, might start hearing about Map Reduce and Hadoop.  With Amazon’s new service these organizations will start getting Google search result lists that show Amazon as a leading player with Hadoop and Map Reduce.  This will add instant credibility to this technology.  In Don’s words, “Amazon is sort of like the new IBM when it comes to cloud computing . . . you can’t go wrong with Amazon!”.  All of this should speed up the adoption of Map Reduce being used as sort of a new data warehouse technology which Don and I both agree is a good thing.  Secondly, Don suggests that the administration it takes an organization to setup Hadoop is a bargain at .015 to .12 cents per hour.  He has done a number of Hadoop consulting engagements and says the cost of a consultant to do a Hadoop setup is not cheap.

We both agree, however, despite all the hoopala regarding this new announcement, that it is still hard to implement a Hadoop solution and there are not that many experts.  Therefore, as exciting as the new Amazon announcement is, it still only gets you half way there.  Learning how to develop code using Map Reduce and Hadoop is a completely different way of thinking than traditional programming paradigms.  Most traditional programming shops will have to re-tool to take advantage of this new paradigm.  On an upside, all freshman CS students at the University of California Berkley are required to learn Hadoop in their freshmen semesters.  All and all this new announcement, in my opinion, puts Amazon in a  class of their own when it comes to “Cloud Computing.”

April 2, 2009  4:48 PM

In 140 characters tell me why map reduce is important?

JohnMWillis John Willis Profile: JohnMWillis

This morning Lance Weatherby, a local Atlanta venture catalyst, asked a simple question on twitter – “In 140 characters tell me why map reduce is important please.” When I saw the question it made me think about the answer. All of us who cover the cloud-o-sphere take terms like map reduce for granted. However, unless you are working directly with cloud computing or you have a BS from Stanford, you probably won’t have a clue why every one is so excited about these terms. My initial one word twitter answer was parallelism. Then I added a second word – multi-core. However, I guess my answers were lazy ones compared to others. Here is a list of great answers posted this morning to Lance’s and my re-tweet question…

March 27, 2009  2:08 PM

Microsoft Was Out of Line

JohnMWillis John Willis Profile: JohnMWillis

Reuven Cohen, the founder of Enomaly, has been one of the brightest voices in the cloud computing revolution. Reuven, a self proclaimed Instigator, has been working tirelessly to get the word out about cloud computing. I first met Reuven on a plane ride from Austin to Chicago last year, and in that time he schooled me on everything “cloud”. I often refer to that trip as my cloud baptism. Since that trip I have stayed in touch and followed his activities. Reuven was responsible for the formation of the extremely popular Cloud Camp events. He also has started the Cloud Computing Interoperability Forum (CCIF) to help pave the way for standards within cloud computing . Most recently he is the Instigator of something called the Cloud Computing Manifesto that will be announced this coming Monday (3/30/09).

A few of us who cover the cloud-o-sphere have been under embargo about this document and its upcoming announcement. I actually had the first opportunity to read this document last week and I had not planned on discussing it until Monday. Actually, I still do not plan on discussing the details of the document until Monday, as I promised; however, what I would like to talk about is the irresponsibility of Microsoft’s actions yesterday regarding this manifesto document. To be clear, Reuven Cohen, first and foremost is an entrepreneur and like all of us he is always looking for angles to create opportunities. Regardless, he has done some great work for the cloud computing community of which myself and software companies like Microsoft have benefited. Therefore, in an effort to coordinate this “Cloud Computing Manifesto”, he needed to share it with hundreds of organizations and numerous individuals, and all he asked the recipients were two questions in return. One, “Are you on board with this?”, and two, “Please keep it under wraps until the announcement date.”. In my opinion Microsoft was totally out of line to pre-announce the manifesto. Millions of analysts, developers, bloggers, and reporters honor Microsoft announcements all of the time and I feel it was completely hypocritical of Microsoft to jump the gun for selfish reasons, and pre-announce this document yesterday. It is obvious they have issues with the document and that is their right; however, they should have voiced their opinions about the document this coming Monday as will I.

March 23, 2009  5:51 PM

The Tale of Three Cloud SLA’s

JohnMWillis John Willis Profile: JohnMWillis

Wikpedia says…

The SLA records a common understanding about services, priorities, responsibilities, guarantees and warranties. Each area of service scope should have the ‘level of service’ defined. The SLA may specify the levels of availability, serviceability, performance, operation, or other attributes of the service such as billing. The ‘level of service’ can also be specified as ‘target’ and ‘minimum’, which allows customers to informed what to expect (the minimum), whilst providing a measurable (average) target value that shows the level of organization performance. In some contracts penalties may be agreed in the case of non compliance of the SLA (but see ‘internal’ customers below).It is important to note that the ‘agreement’ relates to the services the customer receives, and not how the service provider delivers that service.

There has been a lot of hype around the discussion of Service Level Agreements (SLA) in the cloud as of late. SLA hype has been around as long as I can remember and definitely predates the cloud. In the enterprise you will typically hear numbers like 99.999 or terms like “Five Nines”. Five nines is sort of the gold standard in the enterprise, equating to about 5 minutes of outage per year. However, whenever I think of the five nine metric it reminds me what one of my old Six Sigma mentors used to say about five nines – “Five Nines equates to about one commercial plane crash per day for a year.”

The age old problem of negotiating an SLA has always been a difficult task for any client. One of the main contention points in negotiating an SLA is around the outage credits and how they are applied. Does the customer get a reimbursement for the lost services or is the SLA applied to a future credit? In the classic hosting example, a future credit might include hours added after your service contract terminates. The credit for future services is always a suspect model for an SLA. However, with some of the newer pay as you go models in the cloud, it is easier to apply these types of credits (e.g., next month’s bill). However, in any case the test of a great SLA is one that gives a customer a direct reimbursement for lost services. Another area of difficulty when negotiating an SLA is defining the SLA outages. A few times I have been on the provider side of defining an SLA, and it is always in the best interest of the provider to supply extremely clear SLA definitions. Detailed reports are always a good best practice for the provider and the customer. Without clear definitions and documented reports, in most cases, an SLA will be useless. Here are three main areas I typically focus on when discussing an SLA:

  1. Defining the outage.
  2. How does a customer prove an outage to get credit?
  3. How does the credit get applied?

I thought I might take a look at some of the top Cloud providers who provide server instances and see what their SLAs are in relation to the aforementioned three areas.

Amazon Web Services EC2

The Amazon Web Services EC2 SLA can be found at: and describes the details of the AWS EC2cSLA. In the AWS SLA EC2 agreement, Amazon claims a 99.95% SLA. Let’s break down their SLA based on the three areas described above.

Defining the outage

Basically a defined outage in AWS is very confusing at best. It basically means that you can not launch a replacement instance within a 5 minute period while at least two availability zones within the same region are down. I take this to mean that if if two out of three data centers are available and you still can’t launch and/or run any application on your EC2 server, it will not be defined as an outage. To further complicate the matter, AWS calculates their 99.95 based on the previous 365 days. If the customer doesn’t have 365 prior days of service with AWS the prior days are calculated as 100% available. This means if you are a new customer (say 2 months), and a catastrophic event happens to hit two of the three US based data centers and you can’t start an instance for three days, you would get a 10% credit for only one day’s prorated costs for EC2 services. The first two days would not be below the 12 month period 99.95 outage percentage. Also complicating the AWS EC2 SLA is the new reserve instances’ up front fees are not eligible for credits concerning outages. Whoops, they have an exclusion for that scenario described above – “caused by factors outside of our reasonable control, including any force majeure event.”

How does a customer prove an outage to get credit?

In order to receive a credit for a defined AWS EC2 outage a customer has to capture, document, and send a request to Amazon to be processed. In other words, the onus is on the customer to prove the outage. AWS does not provide any interface or report documentation to help the customer define their outages. Furthermore, Amazon requires the customer to document the region, all instance ids, and provide service logs. The customer also is required to cleanse confidential information from the logs and all of this must be done within a 30 day period of the outage.

How does the credit get applied?

First off, the AWS credit gets applied against future credits and is not a reimbursement of lost services. As previously stated it is the customer’s responsibility to provide all of the proof and do it with a 30 day period. If the customer supplies all of the documentation and Amazon approves the outage that qualifies for the below 99.95%, they will then apply a 10 percent discount on the next month’s bill.

SLA Grade “C”

AWS puts a heavy burden on the client to prove the outage. The terms of the SLA are difficult at best to understand.

RackSpace/Mosso Cloud Sites

Cloud Sites was formally called Mosso. Cloud Sites is a service that provides a platform based cloud where users share scalable back end load balancing, web services, and databases clusters. The Cloud Sites SLA can be found here: Let’s break down their SLA based on the three areas described above.

Defining the outage

Based on the agreement, the definition of an outage from Cloud Sites is extremely simple and is described in less than 150 words. The AWS EC2 SLA is over 1000 words. Simply put, if you open an support incident report with Cloud Sites, they will credit you with a 1 day prorated credit for every hour of downtime. Supposedly all you have to do is tell the rep that you have an outage. They should then start the incident and calculate the outage.

How does a customer prove an outage to get credit?

Here is the rub, they don’t tell you about recording the outage when you call. You have to tell them to record the incident as an outage, and then you will have to continuously monitor the situation and call back support to confirm the ending time. Cloud Sites does not do this automatically for you. The reason I know this is because one of my blog sites is hosted on Mosso/Cloud Sites. I have never been given an automatic credit even though I have had at least 5 or 6 outages over the last year. I have also called in at least 5 or six times and an incident report was never discussed. In fact, on most calls they say that a specific cluster is down and that it should be up soon with no mention of a start time or stop time for recording the outage. You have to point this out to them. One of the things that has always annoyed me about Mosso/Cloud Sites is that they never notify you when the outage is fixed, even if you call in and ask about the outage. You have to find that out for yourself. This makes it extremely difficult to document an outage. Another problem with Cloud Sites is that you don’t have access to the servers the way you do with AWS EC2, so it is difficult to gather the appropriate logs to document the outage.

How does the credit get applied?

Cloud Sites outages get applied against future credits and are not a reimbursement of lost services. The Mosso Cloud Sites SLA is a great example of where less is actually less. The brevity of their SLA seems attractive at first, however, they do not have a defined process for requesting a credit. At least AWS has a documented email where you can send your detailed information. There is nothing in the Cloud Sites’ short 58 word SLA that tells you how to go about getting a refund. The client would have to assume they could call Cloud Sites support and request a refund, assuming they actually documented the start and end time of the outage. All and all, it seems like a very confusing process that is supposed to be “as described” very simple. If you can get through all of the above, then Cloud Sites will credit you a one day prorated credit for every 60 minutes of documented downtime.

SLA Grade “B minus- ”

Cloud Sites offers a simple plan; however, they need to be more clear on their SLA. In the SLA, they state that if you open up an incident report they will start the clock. However, they don’t even have a ticketing system for customers to input incidents. You have to call or start a live chat. Also, they do not notify you when the outage is cleared and this makes it difficult for a customer to keep track of their outages.


3Tera recently announced a new 99.999 SLA for their Virtual Private Datacenter (VPDC) customers. The SLA announcement can be found at the following location: Let’s break down their SLA based on the three areas described above.

Defining the outage

According to the 3Tera announcement, the customer does not have to define the outage. 3Tera will automatically detect and calculate outages. The AppLogic Cloud Computing Platform constantly monitors and reports the availability of the system and instantly alerts 3Tera’s operations team of critical issues. Some might think that the 3Tera five nines announcement is the significant part of their SLA compared to AWS 99.95; however, it’s the automatic recording of the outage that is an unprecedented feature of their SLA. While other cloud vendors require the customer to prove the outage times, 3Tera automates this process.

How does a customer prove an outage to get credit?

Short and sweet..automatically with no human intervention.

How does the credit get applied?

3Tera’s credit gets applied to the current month’s bill. VPDC customers automatically receive SLA service credits for any calendar month where availability falls below the targeted 99.999 percent. If availability is anywhere between 99.999 percent and 99.9 percent, a 10 percent credit applies to the whole VPDC service for the entire month. If availability is lower than 99.9 percent, a 25 percent credit applies.

SLA Grade “A minus- ”

I would have otherwise given 3Tera a solid “A”; however, the service has just been announced and is not available yet. When they actually post the SLA page and the actual customer contract, then I will adjust this rating accordingly

March 18, 2009  5:58 PM

Privacy org bashes Google cloud

JoMaitland Jo Maitland Profile: JoMaitland

The EPIC has formally asked the Federal Trade Commission to open an investigation into Google’s cloud computing services — including Gmail, Google Docs, and Picasa — to determine “the adequacy of the privacy and security safeguards.” The petition follows the recent report of a breach of Google Docs.

Here’s a link to the EPIC petition against Google.

Forgot Password

No problem! Submit your e-mail address below. We'll send you an e-mail containing your password.

Your password has been sent to: