Welcome!

Agile Computing Authors: XebiaLabs Blog, Elizabeth White, Yeshim Deniz, Carmen Gonzalez, Pat Romanski

Related Topics: Containers Expo Blog, @CloudExpo

Containers Expo Blog: Blog Feed Post

The Cloud Metastructure Hubub

How Infrastructure 2.0 might leverage publish-subscribe technology like PubSubHubub to enable portability of applications

Pieter_Bruegel_TowerBabel
Tower of Babel by Pieter Bruegel the Elder.
One of the topics surrounding cloud computing that continues to rear its ugly head is the problem of portability across clouds. Avoiding vendor lock-in has been problematic since the day the first line of proprietary code was written and cloud computing does nothing to address this. If anything, cloud makes this worse because one of its premises is that users (that’s you, IT staff) need not concern themselves with the underlying infrastructure. It’s a service, right, so you just use it and don’t worry about it.

Let’s assume for a moment that you can easily move applications from data center to cloud to cloud. Plenty of folks are working on that, but very few of them address the “rest of the story”: the metastructure.

Metastructure contains the metadata that describes the network, application network, and security infrastructure providing all those “don’t worry about” services cloud providers offer. Load balancing, firewalls, IPS, IDS, application acceleration, secure remote access. If you’ve spent time with your cloud provider tweaking those services – or configuring them yourself – then moving to a new cloud provider is not only a huge investment in time, it’s actually going to be painful because you’re essentially going to have to recreate every metastructure configuration again.

Yes, you’ve done this inside your own data center for years. Every forklift replacement or upgrade of infrastructure has come with its own load of baggage in the configuration arena. Switching out vendor equipment – especially core components – can be extremely painful, especially when configurations need to essentially be “translated” between them. But cloud makes this worse because technically speaking you don’t even have access to the existing configurations. You can’t see them, you can’t have them, and you can’t run them through whatever “upgrade” or “migration” script your new vendor offers to ease the process.

Are you depressed yet?

There’s been some talk of including metastructure data with the virtual machine, but the problem with this is that it almost always requires that the meta data be wrapped up using a proprietary API, such as is provided by VMware. That’s okay if you restrict yourself to only cloud providers that use the same virtualization technology, but not okay if you want to be able to make a move from one technology to another. It also assumes that the metadata is specific to the infrastructure, which is even more unlikely when moving between cloud providers.


HOW ABOUT A CLOUD-BASED CMDB (Configuration Management Database)?


There are several ongoing efforts to address this very scenario because it is so painful. Most of them would, if adopted, require vendors to implement support for a specific standard so that configurations can be managed and exchanged in that standard format. That makes sense, that’s how we’ve always handled translation of data between disparate systems that don’t speak the same language. In the application world we call the process of mapping one format to another “integration” and you can easily evoke a look of terror on a co-worker’s face just by saying the word within their range of hearing. Go ahead, try it. Just make sure they aren’t carrying anything heavy that can be easily thrown at you when you do.

CMDB (Configuration Management Database) technology is another method of addressing the problem of, well, managing configurations. These solutions store configuration of a wide variety of infrastructure solutions – from routers and switches to web and application servers to application delivery controllers. They do a great job of managing configuration and can even “push” configuration out to devices if so desired. But the configurations stored and managed in a CMDB are product-specific, not generic, so they can’t adequately today address the problem of portability.

You can probably see where this is going: a cross between CMDB and a nice, industry-wide standard would probably do the trick, wouldn’t it? And if it was public (in the sense that any application or service is public on the network – that is, accessible via the Internet to any cloud provider or customer site) then cloud providers and organizations alike could take advantage of that configuration management mechanism and use it to their advantage. Portability becomes possible rather than fantasy.


PUBSUBHUBUB


Cloud providers and organizations alike are likely to stop right there. Sharing configuration of infrastructure and core components is just asking for trouble. If ever such a cloud-based CMDB were compromised, well…let’s just say it would be A Very Bad Thing.

But what if the actual metadata, the configuration information, were stored either in the enterprise or the cloud provider (or both), and merely pushed and pulled via a public mechanism on-demand?  Configuration isn’t changed all that often and if an organization is moving between clouds they certainly know when they’re doing it. If there was some mechanism through which metastructure could be published and to which infrastructure could subscribe then when changes were made or providers changed that metastructure data could be easily grabbed from the public cloud-CMDB system (cloud catalog, anyone?) and interpreted into product-specific configuration by the products themselves.

Think of  it like SOA clients pulling WSDL (Web Services Description Language) from a UDDI (Universal Description, Discovery, and Integration) server. The SOA client pulls the WSDL, which describes the service(s), configures itself appropriately, and then is able to make use of those services. The intent of introducing UDDI was a service-catalog that could be polled on-demand to provide the latest information about the service and describe it in an abstract, vendor-neutral way such that any client could access any service, regardless of implementation language or environment. Sounds a lot like what we want for infrastructure portability, doesn’t it?

PUBSUBHUBHUBThat’s where PubSubHubub comes in. While this draft standard for a publish-subscribe system is generally being leveraged by software developers to enable faster sharing of information across the Internet, it is also a fine example of a system that could be used by infrastructure 2.0 solutions to share metastructure. Consider the existence of a public PubSubhubub Hub, like Google’s public PubSubHubub Hub, and how it might be leveraged to share metastructure between clouds or the organization and the cloud.

Note that XMPP is used today by at least one cloud provider to enable distributed cloud management in a nature very similar to that of PubSubhubhub.

In any case, the specific implementation of the configuration “hub” is relatively unimportant; what’s important is that (a) customers can publish a vendor-neutral metastructure to an isolated channel that communicates their specific infrastructure needs and (b) providers can subscribe, at will, to customer topics and retrieve metastructure in a way that allows their infrastructure to in turn configuration itself (or be configured by the provider’s system, as is required by the provider’s implementation).

Early on it would be necessary for the cloud provider to provide the “translation” and configuration services simply because even if a metastructure standard existed today (and it doesn’t) it would take months and possibly years before all the possible infrastructure vendors were able to update their systems to interpret the standard. If the provider implements a configuration “gateway”, however, he can immediately take advantage of such a standard and use existing skills and knowledge gained from its automation and orchestration of its cloud to configure the infrastructure appropriately based on the metastructure. This has the added advantage of “hiding” the infrastructure implementation from the outside world, which for some providers is a very important thing to do.


SOME CONFIGURATIONS ARE INHERENTLY VENDOR SPECIFIC


That’s okay for two reasons: first, we ensure that the metadata description is XML-based, because it’s extensible. If we build into the standard a way to extend it naturally such as is provided with XML the interpreters (configuration “gateways”) can either (a) translate if it can or (b) ignore.

Consider the use of OVF (Open Virtualization Format) to further describe what is called a Virtual Machine Contract (VMC):

For each virtual system, the associated metadata is described in a set of specific sections. The VirtualHardwareSection describes the virtual hardware required including the amount of memory, number of CPUs, information about network interfaces, etc. The OperatingSystemSection describes the guest operating system that will run in the virtual system. The ProductSection provides basic information such as the name and vendor of the appliance and can also specify a set of properties that can be used to customize the appliance.

While VMC is very basic at this point, it’s a good start at providing the foundation for building out a more complete, standards-based description of the metastructure necessary to configure an infrastructure to deploy a specific application in a virtual machine format. Using this as the basis for metadata exchange – when fully described – via a public hub could alleviate most of the issues with sharing infrastructure metadata (metastructure) across clouds in a generally vendor non-specific manner. In other words, portability of both the virtual machine and the specific infrastructure configurations necessary to optimally execute and deliver the application to the end user in the most fast and secure manner possible.

We’re nowhere near this point, by the way. VMC needs to be fleshed out as far as standard metadata goes for infrastructure (perhaps a good chore for the SRI Infrastructure 2.0 Working Group) and vendors would need to adopt and extend out the ProductSection of VMC for product specific configuration that isn’t included in the base format. And PubSubHubub would need to be proven to be a secure method of exchanging the metastructure across clouds. What is likely is that as we move forward trying to extend the plateau of collaboration down the stack toward the core infrastructure is that a new set of tools, products, solutions, and services will emerge to fill the unavoidable gaps in the standards, e.g. a service-based cloud configuration hub offering translation of proprietary metastructure data to some other proprietary metastructure data.

Perhaps there’s a better way overall, and OVF/VMC and PubSubHubub will simply remain in our memories as the catalyst and template for a different set of standards providing portability across clouds. But there is a way to provide this level of portability and collaboration across clouds, across the infrastructure and the application. The need – and perhaps more importantly the belief that it’s necessary to address the need – is growing.

UPDATE: Christofer Hoff pointed out that vCloud has been submitted to the DMTF for standardization, technically making it "open" rather than "proprietary." It is still only implemented by VMware technologies, so for the time being it might as well be proprietary, but this may change in the future.

Follow me on Twitter View Lori's profile on SlideShare friendfeedicon_facebook AddThis Feed Button Bookmark and Share

Related blogs & articles:

More Stories By Lori MacVittie

Lori MacVittie is responsible for education and evangelism of application services available across F5’s entire product suite. Her role includes authorship of technical materials and participation in a number of community-based forums and industry standards organizations, among other efforts. MacVittie has extensive programming experience as an application architect, as well as network and systems development and administration expertise. Prior to joining F5, MacVittie was an award-winning Senior Technology Editor at Network Computing Magazine, where she conducted product research and evaluation focused on integration with application and network architectures, and authored articles on a variety of topics aimed at IT professionals. Her most recent area of focus included SOA-related products and architectures. She holds a B.S. in Information and Computing Science from the University of Wisconsin at Green Bay, and an M.S. in Computer Science from Nova Southeastern University.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
Buzzword alert: Microservices and IoT at a DevOps conference? What could possibly go wrong? In this Power Panel at DevOps Summit, moderated by Jason Bloomberg, the leading expert on architecting agility for the enterprise and president of Intellyx, panelists peeled away the buzz and discuss the important architectural principles behind implementing IoT solutions for the enterprise. As remote IoT devices and sensors become increasingly intelligent, they become part of our distributed cloud enviro...
WebRTC sits at the intersection between VoIP and the Web. As such, it poses some interesting challenges for those developing services on top of it, but also for those who need to test and monitor these services. In his session at WebRTC Summit, Tsahi Levent-Levi, co-founder of testRTC, reviewed the various challenges posed by WebRTC when it comes to testing and monitoring and on ways to overcome them.
In his General Session at 16th Cloud Expo, David Shacochis, host of The Hybrid IT Files podcast and Vice President at CenturyLink, investigated three key trends of the “gigabit economy" though the story of a Fortune 500 communications company in transformation. Narrating how multi-modal hybrid IT, service automation, and agile delivery all intersect, he will cover the role of storytelling and empathy in achieving strategic alignment between the enterprise and its information technology.
"A lot of times people will come to us and have a very diverse set of requirements or very customized need and we'll help them to implement it in a fashion that you can't just buy off of the shelf," explained Nick Rose, CTO of Enzu, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
DevOps is being widely accepted (if not fully adopted) as essential in enterprise IT. But as Enterprise DevOps gains maturity, expands scope, and increases velocity, the need for data-driven decisions across teams becomes more acute. DevOps teams in any modern business must wrangle the ‘digital exhaust’ from the delivery toolchain, "pervasive" and "cognitive" computing, APIs and services, mobile devices and applications, the Internet of Things, and now even blockchain. In this power panel at @...
WebRTC services have already permeated corporate communications in the form of videoconferencing solutions. However, WebRTC has the potential of going beyond and catalyzing a new class of services providing more than calls with capabilities such as mass-scale real-time media broadcasting, enriched and augmented video, person-to-machine and machine-to-machine communications. In his session at @ThingsExpo, Luis Lopez, CEO of Kurento, introduced the technologies required for implementing these idea...
Every successful software product evolves from an idea to an enterprise system. Notably, the same way is passed by the product owner's company. In his session at 20th Cloud Expo, Oleg Lola, CEO of MobiDev, will provide a generalized overview of the evolution of a software product, the product owner, the needs that arise at various stages of this process, and the value brought by a software development partner to the product owner as a response to these needs.
The WebRTC Summit New York, to be held June 6-8, 2017, at the Javits Center in New York City, NY, announces that its Call for Papers is now open. Topics include all aspects of improving IT delivery by eliminating waste through automated business models leveraging cloud technologies. WebRTC Summit is co-located with 20th International Cloud Expo and @ThingsExpo. WebRTC is the future of browser-to-browser communications, and continues to make inroads into the traditional, difficult, plug-in web co...
Internet of @ThingsExpo, taking place June 6-8, 2017 at the Javits Center in New York City, New York, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. @ThingsExpo New York Call for Papers is now open.
Who are you? How do you introduce yourself? Do you use a name, or do you greet a friend by the last four digits of his social security number? Assuming you don’t, why are we content to associate our identity with 10 random digits assigned by our phone company? Identity is an issue that affects everyone, but as individuals we don’t spend a lot of time thinking about it. In his session at @ThingsExpo, Ben Klang, Founder & President of Mojo Lingo, discussed the impact of technology on identity. Sho...
Technology vendors and analysts are eager to paint a rosy picture of how wonderful IoT is and why your deployment will be great with the use of their products and services. While it is easy to showcase successful IoT solutions, identifying IoT systems that missed the mark or failed can often provide more in the way of key lessons learned. In his session at @ThingsExpo, Peter Vanderminden, Principal Industry Analyst for IoT & Digital Supply Chain to Flatiron Strategies, will focus on how IoT depl...
Data is an unusual currency; it is not restricted by the same transactional limitations as money or people. In fact, the more that you leverage your data across multiple business use cases, the more valuable it becomes to the organization. And the same can be said about the organization’s analytics. In his session at 19th Cloud Expo, Bill Schmarzo, CTO for the Big Data Practice at Dell EMC, introduced a methodology for capturing, enriching and sharing data (and analytics) across the organization...
With all the incredible momentum behind the Internet of Things (IoT) industry, it is easy to forget that not a single CEO wakes up and wonders if “my IoT is broken.” What they wonder is if they are making the right decisions to do all they can to increase revenue, decrease costs, and improve customer experience – effectively the same challenges they have always had in growing their business. The exciting thing about the IoT industry is now these decisions can be better, faster, and smarter. Now ...
WebRTC is about the data channel as much as about video and audio conferencing. However, basically all commercial WebRTC applications have been built with a focus on audio and video. The handling of “data” has been limited to text chat and file download – all other data sharing seems to end with screensharing. What is holding back a more intensive use of peer-to-peer data? In her session at @ThingsExpo, Dr Silvia Pfeiffer, WebRTC Applications Team Lead at National ICT Australia, looked at differ...
The cloud market growth today is largely in public clouds. While there is a lot of spend in IT departments in virtualization, these aren’t yet translating into a true “cloud” experience within the enterprise. What is stopping the growth of the “private cloud” market? In his general session at 18th Cloud Expo, Nara Rajagopalan, CEO of Accelerite, explored the challenges in deploying, managing, and getting adoption for a private cloud within an enterprise. What are the key differences between wh...
"ReadyTalk is an audio and web video conferencing provider. We've really come to embrace WebRTC as the platform for our future of technology," explained Dan Cunningham, CTO of ReadyTalk, in this SYS-CON.tv interview at WebRTC Summit at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
In 2014, Amazon announced a new form of compute called Lambda. We didn't know it at the time, but this represented a fundamental shift in what we expect from cloud computing. Now, all of the major cloud computing vendors want to take part in this disruptive technology. In his session at 20th Cloud Expo, John Jelinek IV, a web developer at Linux Academy, will discuss why major players like AWS, Microsoft Azure, IBM Bluemix, and Google Cloud Platform are all trying to sidestep VMs and containers...
IoT is at the core or many Digital Transformation initiatives with the goal of re-inventing a company's business model. We all agree that collecting relevant IoT data will result in massive amounts of data needing to be stored. However, with the rapid development of IoT devices and ongoing business model transformation, we are not able to predict the volume and growth of IoT data. And with the lack of IoT history, traditional methods of IT and infrastructure planning based on the past do not app...
The many IoT deployments around the world are busy integrating smart devices and sensors into their enterprise IT infrastructures. Yet all of this technology – and there are an amazing number of choices – is of no use without the software to gather, communicate, and analyze the new data flows. Without software, there is no IT. In this power panel at @ThingsExpo, moderated by Conference Chair Roger Strukhoff, Dave McCarthy, Director of Products at Bsquare Corporation; Alan Williamson, Principal ...
WebRTC has had a real tough three or four years, and so have those working with it. Only a few short years ago, the development world were excited about WebRTC and proclaiming how awesome it was. You might have played with the technology a couple of years ago, only to find the extra infrastructure requirements were painful to implement and poorly documented. This probably left a bitter taste in your mouth, especially when things went wrong.