Welcome!

IBM Cloud Authors: Elizabeth White, Pat Romanski, Liz McMillan, Yeshim Deniz, Stefan Bernbo

Related Topics: Machine Learning , IBM Cloud, @CloudExpo

Machine Learning : Blog Post

Cloud Computing and Virtual Images

A look at several different cloud computing solutions will reveal a technological enabler present in almost each one

Cloud Expo on Ulitzer

A look at several different cloud computing solutions will reveal a technological enabler present in almost each one. The enabler I'm talking about is the use of virtual images. I cannot think of many, if any, cloud computing solutions that provide software elements (i.e. more than just servers, storage, memory, etc.) that do not use virtual images in some form or fashion.

Of course, one of the reasons virtual images form the backbone of many cloud solutions is obvious. Virtual images provide the benefits of server virtualization. We can activate many virtual images on the same physical machine, thus allowing us to achieve multi-tenancy (multiple operating systems and software stacks installed on the same physical machine). Besides driving higher hardware utilization rates, it also provides us the capability to run heterogeneous software environments on the same piece of hardware. This both enables and encourages the creation of a shared pool of compute resources which is a key characteristic of cloud computing environments.

Server virtualization may be the first thing that comes to mind when we think about virtual images, but at least in the context of cloud computing, I do not believe this is the most important benefit. If we are looking at cloud computing as a means to quickly and consistently provision software environments, then I think virtual images provide us a capability and benefit more valuable than server virtualization. In this respect, virtual images provide us a medium through which we can templatize the configuration of our software environments.

Consider the case of a fairly basic application serving environment. In this environment, you are likely to install an operating system, application server, and probably some type of load balancing solution. Each of these typically requires a different piece of software, different installation procedures, and finally integration with the other components. Installing these into a typical environment, without the use of virtual images, means that you either have scripts that account for each different piece of software and finally integration of the different components, or it means that a person manually installs and integrates the pieces each time you need an environment. Either process can be time-consuming and costly to maintain over time.

Enter the use of virtual images. With a virtual image, you can install and integrate all three components ONE time, and then capture the resultant environment as a virtual image. At this point, when an application environment is needed, the virtual image can simply be activated on top of a hypervisor platform. The application environment is typically available in a much more timely fashion than if manually installed or integrated because the installation and configuration have already been captured in the virtual image.

From what I described above, you may have caught on to what would seem like a drawback of using virtual images to templatize software environments. Specifically, it may seem that you need a distinct virtual image for every unique configuration of your application environment. If this were the case, management of your virtual image library would soon become a nightmare and the resulting cost (in both resource and time) would likely outweigh the original benefits. However, thanks to a relatively new standards-based approach to virtual images, this is not necessarily a problem.

The standard I'm talking about is the Open Virtualization Format (OVF) standard which has been published by the Distributed Management Task Force (DMTF). According to the OVF standard, it "describes an open, secure, portable, efficient and extensible format for the packaging and distribution of software to be run in virtual machines." In particular to our discussion here, there is a part of the standard that describes the use of an ovf-env.xml file within the virtual image.

This file is essentially a key-value style XML file that describes desired aspects of the environment. Keys and values can be supplied during image activation, and configurations scripts that run during virtual image activation can read information from the file and react appropriately. Thus, instead of supplying N different virtual images for N different software environments, you can supply 1 virtual image and utilize the ovf-env.xml file in conjunction with configuration scripts within the image to produce N different environments. The use of this mechanism with virtual images delivers the capability to templatize software environments without sacrificing flexibility or encouraging unsustainable virtual image proliferation.

In WebSphere, we utilize the model outlined in the OVF standard when packaging our WebSphere Application Server Hypervisor Edition virtual images. This allows our WebSphere CloudBurst product to provision these images and create many different types of WebSphere Application Server environments from a single virtual image (read this article for more information). I expect the use of this standard and the mechanisms it provides will become pretty prevalent in the near future. Now if we could just get to the point where virtual disk formats are standardized, but that's an entirely different topic.

More Stories By Dustin Amrhein

Dustin Amrhein joined IBM as a member of the development team for WebSphere Application Server. While in that position, he worked on the development of Web services infrastructure and Web services programming models. In his current role, Dustin is a technical specialist for cloud, mobile, and data grid technology in IBM's WebSphere portfolio. He blogs at http://dustinamrhein.ulitzer.com. You can follow him on Twitter at http://twitter.com/damrhein.

IoT & Smart Cities Stories
The graph represents a network of 1,329 Twitter users whose recent tweets contained "#DevOps", or who were replied to or mentioned in those tweets, taken from a data set limited to a maximum of 18,000 tweets. The network was obtained from Twitter on Thursday, 10 January 2019 at 23:50 UTC. The tweets in the network were tweeted over the 7-hour, 6-minute period from Thursday, 10 January 2019 at 16:29 UTC to Thursday, 10 January 2019 at 23:36 UTC. Additional tweets that were mentioned in this...
Artificial intelligence, machine learning, neural networks. We're in the midst of a wave of excitement around AI such as hasn't been seen for a few decades. But those previous periods of inflated expectations led to troughs of disappointment. This time is (mostly) different. Applications of AI such as predictive analytics are already decreasing costs and improving reliability of industrial machinery. Pattern recognition can equal or exceed the ability of human experts in some domains. It's devel...
The term "digital transformation" (DX) is being used by everyone for just about any company initiative that involves technology, the web, ecommerce, software, or even customer experience. While the term has certainly turned into a buzzword with a lot of hype, the transition to a more connected, digital world is real and comes with real challenges. In his opening keynote, Four Essentials To Become DX Hero Status Now, Jonathan Hoppe, Co-Founder and CTO of Total Uptime Technologies, shared that ...
The Japan External Trade Organization (JETRO) is a non-profit organization that provides business support services to companies expanding to Japan. With the support of JETRO's dedicated staff, clients can incorporate their business; receive visa, immigration, and HR support; find dedicated office space; identify local government subsidies; get tailored market studies; and more.
After years of investments and acquisitions, CloudBlue was created with the goal of building the world's only hyperscale digital platform with an increasingly infinite ecosystem and proven go-to-market services. The result? An unmatched platform that helps customers streamline cloud operations, save time and money, and revolutionize their businesses overnight. Today, the platform operates in more than 45 countries and powers more than 200 of the world's largest cloud marketplaces, managing mo...
At CloudEXPO Silicon Valley, June 24-26, 2019, Digital Transformation (DX) is a major focus with expanded DevOpsSUMMIT and FinTechEXPO programs within the DXWorldEXPO agenda. Successful transformation requires a laser focus on being data-driven and on using all the tools available that enable transformation if they plan to survive over the long term. A total of 88% of Fortune 500 companies from a generation ago are now out of business. Only 12% still survive. Similar percentages are found throug...
When Enterprises started adopting Hadoop-based Big Data environments over the last ten years, they were mainly on-premise deployments. Organizations would spin up and manage large Hadoop clusters, where they would funnel exabytes or petabytes of unstructured data.However, over the last few years the economics of maintaining this enormous infrastructure compared with the elastic scalability of viable cloud options has changed this equation. The growth of cloud storage, cloud-managed big data e...
Your applications have evolved, your computing needs are changing, and your servers have become more and more dense. But your data center hasn't changed so you can't get the benefits of cheaper, better, smaller, faster... until now. Colovore is Silicon Valley's premier provider of high-density colocation solutions that are a perfect fit for companies operating modern, high-performance hardware. No other Bay Area colo provider can match our density, operating efficiency, and ease of scalability.
At CloudEXPO Silicon Valley, June 24-26, 2019, Digital Transformation (DX) is a major focus with expanded DevOpsSUMMIT and FinTechEXPO programs within the DXWorldEXPO agenda. Successful transformation requires a laser focus on being data-driven and on using all the tools available that enable transformation if they plan to survive over the long term. A total of 88% of Fortune 500 companies from a generation ago are now out of business. Only 12% still survive. Similar percentages are found throug...
As you know, enterprise IT conversation over the past year have often centered upon the open-source Kubernetes container orchestration system. In fact, Kubernetes has emerged as the key technology -- and even primary platform -- of cloud migrations for a wide variety of organizations. Kubernetes is critical to forward-looking enterprises that continue to push their IT infrastructures toward maximum functionality, scalability, and flexibility. As they do so, IT professionals are also embr...