Thursday, February 19, 2015

Private cloud IaaS Evolution - Failures were the stepping stones to success



Before we examine cloud computing lets 1st understand some of the factors that precipitated the need for cloud computing.

Cloud computing started with the Web, allowing  Enterprises to promote their information in a form that could be accessed from anywhere (geographically) and from anything (devices) without needing  any client side software (VPN,…).  Realizing the convenience of the Web,  Enterprise users also wanted access to their private Enterprise information through the Web as well (email, CRM & ERP data, HR data, financial data,…).  

All this external access caused a large headache for the Enterprise IT staff where the carefully articulated rules of the intranet simply did not apply to the internet (capacity planning/scaling, security being under a physical lock and key,…). The Enterprise IT staff had to scramble to meet the requirements of their customers through adhoc scripts and home grown solutions/processes designed to scale and secure their current environment to meet the requirements of the internet. 

The formalization of these processes is what I believe to be one of the fundamental requirements that led to today’s Cloud computing.

But that still leaves the question of what does it offer IT departments?  In its simplest form, the Cloud was attempting to solve 3 fundamental IT problems.  Managing Cost, Risk and Service levels.

 

Though there are many technologies that make up  Cloud Computing, there were 2 fundamental technology building blocks that became the foundation for modern clouds.

1.    Virtualization: In this context, virtualization is an abstract representation of a new resource from the underlying physical resources (compute, storage, network, services).   


Note: There is a common misconception that virtualization involves a single large resource (ex: a physical computer) to be partitioned into smaller pieces (host virtualization like Xen, or, as in the case of OS virtualization like LXC). 

Virtualization can also go the other way, where many smaller resources (physical computers) can be aggregated together to also form a larger single resource (ex: shared memory computing system).


2.   Clustering: Allows the above partitions to re-form into dynamic groups with an interdependency between them.   These interdependency's are basic clustering requirements like heartbeats, quorum, cluster membership, fencing, service management frameworks (another future blog of mine),…. of distributed computing.

Having these 2 disparate capabilities of resource optimization and distributed systems coming together on a single platform/system, enabled a flexible and powerful framework; thus setting the stage for cloud computing. 

It took a long time to adopt cloud computing because Enterprises were justifiably wary of the issues exposed by the public cloud (security, performance, failures,..). and so IT providers scrambled by providing private cloud’s (that’s where the money was).  Early efforts in Private cloud IaaS solutions were also not very successful.  There was a fundamental problem that I saw, essentially,...   

What problem were we really solving for the enterprise through this private cloud IaaS?

IaaS gave Enterprises a consistent way to dynamically and efficiently reuse computing resources at the infrastructure level.  However the Enterprises had already solved this problem though various home grown scripts of their own and were opposed to ripping and replacing their familiar well-orchestrated home grown scripts with another vendor’s solution.  Having no consistent standard between various IT providers, IaaS solutions simply added to the controversy.

A lot of this changed with the emergence of initial standardization efforts (Open Stack, Cloud Stack, AWS,..) to provide a consistent approach towards IaaS. 
 
At least 2 things worked in Open Stack’s favor:
    1. AdoptionMany IT vendors abandoned/aligned their private efforts towards Open Stack.
    2. Open development: Unlike the politics of normal standards in the computing industry, Open Stack took a more pragmatic approach by providing Open Source development to everyone in the industry (and then took an aggressive stance in formalizing releases).
Of course, we can always debate the problems with Open Stack (vendors morphying the stack, Open Stack aimed at developer play rather than a deployment ready model, ...), nonetheless Open Stack finally provided the necessary ground swell for the Enterprise data centers to take cloud computing seriously. 

Wednesday, February 4, 2015

Pioneering the cloud - My personal journey





The cloud like everything else (big data, vitualization, SDN), are not “conceptually”, new ideas. Every future “incarnation” of a past idea does however differ in scope, capability and opportunity.  For over a decade people have been mulling on what “cloud” means.  Not wishing to continue beating a dead horse, I want to instead focus on factors that influenced the concepts that led to what we call as cloud today.   Generally a lot of credit is given to Salesforce and Amazon AWS as trailblazers to cloud computing.  True as it maybe, there is always a story behind the story.  This is my story.

I did not invent this incarnation of the cloud, but I am fortunate to have worked with folks like Khalidi, Hisgen, Penkler, Schroepher, and many others, who helped create the initial concepts.  Most of all, I am deeply privileged to have eventually led pioneering efforts that directly influenced/resulted in cloud computing as we know it today. 

My personal story on cloud computing evolved as a result of work on a single system image (multiple machines running a single operating system across them).  The clustering group at Sun, while building an SSI (single system image), had an epiphany.  Why create a single system image when a “perception” of a single system image may be sufficient to achieve high availability goals?  In doing so, we focused on creating this “perception” by virtualizing OS services (like: devices, file system, network, host applications, ..) across multiple machines.  This virtualization layer later influenced the thinking/foundation of a revolutionary project at Sun called N1.

Its at this point in time when we had another epiphany.  Could we extended this concept of “perception” and “virtualization”, from Sun Cluster into systems management, such that we could manage N systems as though they were 1 (N1)?    In doing so, could we then help balance a customer’s costs, risks and service level in a more optimal way?    Sun then embarked on realizing the dream of N1 (the predecessor to what we now call as IaaS – Infrastructure as a Service).

At the time, other system companies, feeling left out, publicly coined similar concepts: IBM with Autonomic Computing, HP with Adaptive infrastructure, even Microsoft got in on the act with Dynamic Systems Interface.   

While N1 and others, took a ground up approach (IaaS) to building what we call as “private clouds”, a parallel top down effort (SaaS) with ASP like capability was quickly evolving into “public cloud” services.  Then with the launch of Salesforce and Amazon’s AWS services, the term “cloud” got cemented in history.

Though this is not a complete list of offerings, I have attempted to sketch out a very rough time line of this evolution.



I thoroughly enjoyed my time in N1.  Under my watch, back in 2005, we had even demonstrated live migration of fully saturated Oracle instances from one physical machine to another in a matter of minutes.  Sadly, it was way ahead of its time and eventually N1 faded from Sun’s history with parts of it being used in various other initiatives.  Though there were many issues to tackle in the evolution of N1, the biggest problems to overcome had to do with human behavior.