Automation and Orchestration: Why What You Think You’re Doing is Less Than Half of What You’re Really Doing

One of the main requirements of the cloud is that most—if not all—of the commodity IT activities in your data center need to be automated (i.e. translated into a workflow) and then those singular workflows strung together (i.e. orchestrated) into a value chain of events that delivers a business benefit. An example of the orchestration of a series of commodity IT activities is the commissioning of a new composite application (an affinitive collection of assets—virtual machines—that represent web, application and database servers as well as the OSes and software stacks and other infrastructure components required) within the environment. The outcome of this commissioning is a business benefit whereas a developer can now use those assets to create an application for either producing revenue, decreasing costs or for managing existing infrastructure better (the holy trinity of business benefits).

When you start to look at what it means to automate and orchestrate a process such as the one mentioned above, you will start to see what I mean by “what you think you’re doing is less than half of what you’re really doing.” Hmm, that may be more confusing than explanatory so let me reset by first explaining the generalized process for turning a series of commodity IT activities into a workflow and by turn, an orchestration and then I think you’ll better see what I mean. We’ll use the example from above as the basis for the illustration.

The first and foremost thing you need to do before you create any workflow (and orchestration) is that you have to pick a reasonably encapsulated process to model and transform (this is where you will find the complexity that you don’t know about…more on that in a bit). What I mean by “reasonably encapsulated” is that there are literally thousands of processes, dependent and independent, going on in your environment right now and based on how you describe them, a single process could be either A) a very large collection of very short process steps, or, Z) a very small collection of very large process steps (and all letters in between). A reasonably encapsulated process is somewhere on the A side of the spectrum but not so far over that there is little to no recognizable business benefit resulting from it.

So, once you’ve picked the process that you want to model (in the world of automation, modeling is what you do before you get to do anything useful ;) ) you then need to analyze all of the processes steps required to get you from “not done” to “done”…and this is where you will find the complexity you didn’t know existed. From our example above I can dive into the physical process steps (hundreds, by the way) that you’re well aware of, but you already know those so it makes no sense to. Instead, I’ll highlight some areas of the process that you might not have thought about.

Aside from the SOPs, the run books and build plans you have for the various IT assets you employ in your environment, there is probably twice that much “required” information that resides in places not easily reached by a systematic search of your various repositories. Those information sources and locations are called “people,” and they likely hold over half of the required information for building out the assets you use, in our example, the composite application. Automating the process steps that are manifested in those locations only is problematic (to say the least), if not for the fact that we haven’t quite solved the direct computer-to-brain interface, but for the fact that it is difficult to get an answer to a question we don’t yet know how to ask.

Well, I should amend that to say “we don’t yet know how to ask efficiently” because we do ask similar questions all the time, but in most cases without context, so the people being asked seldom can answer, at least not completely. If you ask someone how they do their job, or even a small portion of their job, you will likely get a blank stare for a while before they start in how they arrive at 8:45 AM and get a cup of coffee before they start looking at email…well you get the picture. Without context, people rarely can give an answer because they have far too many variables to sort through (what they think you’re asking, what they want you to be asking, why you are asking, who you are, what that blonde in accounting is doing Friday…) before they can even start answering. Now if you give someone a listing or scenario in which they can relate (when do you commission this type of composite application, based on this list of system activities and tools?) they can absolutely tell you what they do and don’t do from the list.

So context is key to efficiently gaining the right amount of information that is related to the subject chain of activities that you are endeavoring to model- but what happens when (and this actually applies to most cases) there is no ready context in which to frame the question? Well, it is then called observation, either self or external, where all process steps are documented and compiled. Obviously this is labor intensive and time inefficient, but unfortunately it is the reality because probably less than 50% of systems are documented or have recorded procedures for how they are defined, created, managed and operated…instead relying on institutional knowledge and processes passed from person to person.

The process steps in your people’s heads, the ones that you don’t know about—the ones that you can’t get from a system search of your repositories—are the ones that will take most of the time documenting, which is my point, (“what you think you’re doing is less than half of what you’re really doing”) and where a lot of your automation and orchestration efforts will be focused, at least initially.

That’s not to say that you shouldn’t automate and orchestrate your environment—you absolutely should—just that you need to be aware that this is the reality and you need to plan for it and not get discouraged on your journey to the cloud.

Blackbaud Acquires Convio

Blackbaud, Inc. today announced that it has completed its acquisition of Convio, Inc., a leading provider of on-demand constituent engagement solutions. Under the terms of the merger agreement, Blackbaud paid an aggregate purchase price of approximately $325 million. Blackbaud financed the deal through a combination of cash and borrowings from its credit facility.

“This is an exciting day for the Blackbaud and Convio teams. Together, we can build better solutions for nonprofits, and that’s what drives us,” said Marc Chardon, Blackbaud’s chief executive officer. “Convio’s strengths in online and social marketing, and subscription and cloud-based offerings complement ours, and will accelerate our ability to deliver more to the nonprofit sector.”

“Having worked with both Convio and Blackbaud over the past few years, we are pleased to see this new development,” said Major George Hood, National Community Relations Secretary for The Salvation Army. “Being able to work with one vendor across multiple channels of engagement will be a benefit to The Salvation Army, and we are confident Blackbaud will continue to help us more effectively engage with our donors and supporters.”

Originally announced on January 17, 2012, the acquisition followed the completion of the tender offer Blackbaud made through its wholly owned subsidiary, Caribou Acquisition Corporation, for all the outstanding shares of Convio common stock for $16 per share, net to the seller in cash, without interest and less any applicable withholding taxes. Immediately prior to the merger, Caribou Acquisition Corporation held approximately 90.4% of Convio’s outstanding common stock. As a result, Blackbaud was able to complete a “short-form” merger under Delaware law where all outstanding shares of Convio common stock that were not previously tendered (other than shares held by Caribou Acquisition Corporation or Convio stockholders that properly exercise appraisal rights under Delaware law) were converted into the right to receive the same consideration paid to stockholders in the tender offer. Blackbaud assumed Convio equity awards that were unvested as of closing. Convio’s common stock has ceased trading on the Nasdaq Global Select Market.

Blackbaud plans to support Convio’s current offerings, and the companies’ combined research and development (R&D) teams will work with customers to improve and extend current products and build new offerings. Blackbaud plans to keep Convio’s current office structure, adding key offices in the Bay Area and Austin. Gene Austin, Convio’s former president and CEO, will lead the enterprise customer business unit at Blackbaud, reporting to Marc Chardon. “We are excited to work together to bring nonprofits the technology they need at a faster pace than either of us could have separately,” said Austin.

Easter Seals, an organization focused on providing exceptional services, education, outreach, and advocacy to people living with autism and other disabilities, has worked closely with both Blackbaud and Convio for nearly a decade. “In that time, each company has developed core, and in many cases distinct strengths, in the fundraising and marketing arenas,” said Steve Bergman, Easter Seals’ CIO. “Blackbaud’s acquisition of Convio will hasten the creation of new products that could enhance the effectiveness of nonprofits like Easter Seals.”Mexico, the Netherlands, and the United Kingdom.


The Definition of a Converged Infrastructure

There’s been a cacophony of hyperbole and at times marketing fluff from vendors and analysts with regards to Reference Architectures and Converged Infrastructures. As IBM launched PureSystems, NetApp & Cisco decided it was also a good time to reiterate their strong partnership with FlexPod. In the midst of this, EMC decided to release their new and rather salaciously titled VSPEX. From the remnants and ashes of all these new product names and fancy launch conferences, the resultant war blogs and Twitterati battles ensued. As I poignantly watched on from the trenches in an almost Siegfried Sassoon moment, it was quickly becoming evident that there was now an even more ambiguous understanding of what distinguishes a Converged Infrastructure from a Reference Architecture, what it’s relation was with the Private Cloud and more importantly whether you, the end user should even care.

read more

Using an In-Memory Data Grid to Scale Cloud-Based Apps at Cloud Expo NY

The elastic resources offered by cloud computing have created an exciting opportunity for applications to handle very large workloads. However, writing applications that span an elastic pool of virtual servers creates huge challenges for developers. How can these virtual servers easily and efficiently share application data while avoiding scaliability bottlenecks? The answer lies in using in-memory data grids (IMDGs) to provide a powerful, easy-to-use, and highly scalable storage layer.
IMDGs unlock the potential for application scaling while eliminating the cost and performance impact of using blob stores and database servers. IMDGs also provide a powerful platform for fast data analysis and enable transparent data migration between on-premise sites and the cloud.

read more

Xerox’s Cloud Computing Capabilities to Aid Airline Safety

“Airlines can now leapfrog to the cloud to expedite their communications and do so at costs much lower than maintaining existing mainframe systems,” said Ken Stephens, senior vice president of cloud services, Xerox, as Xerox and AvFinity announced and agreement to provide seamless transmission of flight-critical communication.

“There is no room for error in ensuring safety in the skies,” Stephens added.

read more

Multi Community Cloud Services – Moving Beyond the Community Cloud

Community cloud services are services that are shared by multiple members of a community. They centralize the functions and can provide specific mission requirements. The mission requirements can be related to specific policies and compliance. Community clouds leverage the benefits of a public cloud but can satisfy specific requirements. These are perfect for communities with specific focus areas such as performance, auditing or policies that need to be applied across the community.
A major benefit of community clouds is that many organizations can come together and pitch in for the cost of the cloud. This sharing results in significant savings compared to each organization setting up and supporting their own services. Community cloud is a term that is discussed a lot, but I have a new term for the cloud that combines multiple communities together, “Multi community cloud”. Multi Community cloud services span across communities with similar functions and interests. For example, payment cloud services can be used by multiple communities (such as banking, finance, insurance etc) to make payments to individuals and companies, this would fall under multi community cloud services. These services have to be broad enough to encompass the functions of the multiple communities. With the rapid growth of the community cloud, multi community clouds will evolve to a higher level as communities identify and realize the benefits of leveraging many similar capabilities, even though some policies may be different.

read more

Canada Cloud Network – A Triple Helix Design

The idea behind the Canada Cloud Network has been to identify, review and then utilize a number of Innovation Best Practices to design how (and why) this forum works.

In particular one key goal is to implement a ‘Next Generation Cluster‘, a thought leadership piece from Cisco on how the original cluster model from Michael Porter can be upgraded through new Cloud 2.0 technologies.

read more

Cloud Expo New York Speaker Profile: Mårten Mickos – Eucalyptus Systems

With Cloud Expo 2012 New York (10th Cloud Expo) now five weeks away, what better time to introduce you in greater detail to the distinguished individuals in our incredible Speaker Faculty for the technical and strategy sessions at the conference…

We have technical and strategy sessions for you every day from June 11 through June 14 dealing with every nook and cranny of Cloud Computing and Big Data, but what of those who are presenting? Who are they, where do they work, what else have they written and/or said about the Cloud that is transforming the world of Enterprise IT, side by side with the exploding use of enterprise Big Data – processed in the Cloud – to drive value for businesses…?

read more

Fall is a Season Not the Future of SaaS

I recently read a TechCrunch article titled: The Rise of Big Data Apps and the Fall of SaaS written by Raj De Datta.

I do agree with the prefix. As far as the suffix is concerned, I can only explain why Raj De Datta’s opinion is wrong.

I usually tend not to agree with statements about premature End of Life of technologies or architectures or trends.

In most cases, it easier to kill a concept in writing, than to persuade a large number of users to change their habits.

read more

Cloud Management – Various Solutions and Standards

The growing use of Infrastructure as a Service to run custom applications, SaaS, and PaaS offerings is increasing the demand of management solutions to leverage the automation offered by IaaS.  One of the biggest advantages of IaaS is the ability to automate tasks which were not possible before.  E.g. a crashed server can be restored with an API call by launching a new instance.

read more