Welcome!

IBM Cloud Authors: Elizabeth White, Yeshim Deniz, Pat Romanski, Liz McMillan, Carmen Gonzalez

Related Topics: @CloudExpo

@CloudExpo: Blog Feed Post

A Formula for Just in Time Provisioning in the Cloud

Traditional Architectures and Deployment Models is Superior to Cloud Computing in Provisioning

imageOne of the ways in which traditional architectures and deployment models is actually superior (yes, I said superior) to cloud computing is in provisioning.  Before you label me a cloud heretic, let me explain. In traditional deployment models capacity is generally allocated based on anticipated peaks in demand. Because the time to acquire, deploy, and integrate hardware into the network and application infrastructure this process is planned for and well-understood, and the resources required are in place before they are needed. In cloud computing, the benefit is that the time required to acquire those resources is contracted to virtually nothing, making capacity planning much more difficult. The goal is just-in-time provisioning – resources are not provisioned until you are sure you’re going to need them because part of the value proposition of cloud and highly virtualized infrastructure is that you don’t pay for resources until you need them. But it’s very hard to provision just-in-time and sometimes the result will end up being almost-but-not-quite-in-time. Here’s a cute [whale | squirrel | furry animal] to look at until service is restored.

While fans of Twitter’s fail whale are loyal and everyone will likely agree its inception and subsequent use bought Twitter more than a bit of patience with its often times unreliable service, not everyone will be as lucky or have customers as understanding as Twitter. We’d all really rather prefer not to see the Fail Whale, regardless of how endearing he (she? it?) might be.

But we also don’t want to overprovision and potentially end up spending more money than we need to. So how can these two needs be balanced?


THE VARIABLES

The first thing we need to do is know, in a given cloud, how long it will take to provision capacity and put it into the rotation. It would be nice if cloud providers offered a service devops could query to get the “current wait time” (a la customer service queues) but until then this timing will certainly need to be obtained by simply keeping track yourself.

The other “constant” (if there is such a thing in a elastic environment) is the capacity of the instances you are using. We’ll consider this a constant at this point because honestly, we’re not ready to move to the higher levels of enlightenment (and programmability) required to dynamically determine this value – though that will most certainly be the subject of a future, future post. Capacity needs to be in units measurable by the solution aggregating requests (a strategic point of control). This is almost certainly a Load balancer or application delivery controller of some kind, as these components are what enable elastic scalability and basically make cloud work. Typical units might be RPS (requests per second) but because of differences in the way different types of requests consume resources it may be easier and more consistent across applications to use connections, as in “concurrent open connections” as it is one of the limiting factors on capacity of application services.

The other two variables we need are only available at run-time, dynamically. You need to know the existing load – in the same units as capacity – and the current resource consumption rate. The resource consumption rate should be in the same units as capacity and in the same time unit as time to provision. If that’s minutes, use minutes. If that’s seconds, use seconds, and so on. It should be noted that the resource consumption rate is the harder of the two to obtain, requiring access to the historical performance statistics of the aggregating component (the load balancer).

But let’s assume you can and do have all these variables. How is that useful?


THE FORMULA

image

The formula is actually a fairly simple one once you’ve got the variables. You’re trying to figure out how much time you have before capacity is depleted (and hoping that the answer is smaller than the time to provision). Just-in-time provisioning, as the term implies, is an attempt to formulaically determine when to start the provisioning process such that capacity always meets demand without over-provisioning. Now, you’re always hedging your bets that a high resource consumption rate will continue in the next “time to provision”. It may be the case that the “spike” is over before the new instance is provisioned, but in this case you’re better safe than sorry, right? Unless your customers like seeing a [whale | squirrel | furry animal] and don’t mind the wait.

Consider the following example:

Total capacity right now is 1000 connections. The existing load is 800 connections. Connections are currently being consumed at a rate of 200 per minute. Provisioning more capacity takes 5 minutes. 1000image – 800 = 200 / 200 per minute = 1 minute of capacity left.

Provisioning should have begun at least 4 minutes ago, and optimally 9 minutes ago (too many years developing software for me – fudge factor included) to ensure capacity was available. In this situation, someone is getting a picture of a [whale | squirrel | furry animal].

The trick for devops is to tune the threshold at which the provisioning process begins. Too soon and you might be wasting resources (and money), too late and you end up with timeouts and angry users. Devops needs a way to programmatically evaluate the results and decide, based on the application (it may be more sensitive to failure than others) and the business significance of the transaction (purchase processes may need more warning than search or general browsing), when it is appropriate to start provisioning in such a way as to ensure availability without incurring a lot of cost overhead.


DYNAMIC INFRASTRUCTURE can ENABLE THIS TODAY

Now I’m sure this sounds like something out of science fiction, but it’s not. The variables can be obtained, if not easily, and the formula can easily be codified into scripts or management applications that enable this entire process to be automated. At a minimum, it should be possible for any skilled devop (developer or operations focused) to create a script/application/widget/gadget that gathers the data required and displays an alert when it’s time to provision – I suggest a nice HTML interface that encloses the entire page in BLINK tags, because nothing says FIX THIS NOW than BLINKING TEXT, right?

Regardless of how it’s actually to put to use, just-in-time provisioning is the goal of agile operations. How that happens is by leveraging cloud computing and highly virtualized data centers and combining that flexibility with the agility of a dynamic infrastructure. Remember, Infrastructure 2.0 isn’t just about configuration through automation. That’s nice, but it’s not the whole enchilada. It’s also about dynamism and flexibility at run-time, in providing actionable data and capabilities that allow elastic scalability to be truly elastic.

Read the original blog entry...

More Stories By Lori MacVittie

Lori MacVittie is responsible for education and evangelism of application services available across F5’s entire product suite. Her role includes authorship of technical materials and participation in a number of community-based forums and industry standards organizations, among other efforts. MacVittie has extensive programming experience as an application architect, as well as network and systems development and administration expertise. Prior to joining F5, MacVittie was an award-winning Senior Technology Editor at Network Computing Magazine, where she conducted product research and evaluation focused on integration with application and network architectures, and authored articles on a variety of topics aimed at IT professionals. Her most recent area of focus included SOA-related products and architectures. She holds a B.S. in Information and Computing Science from the University of Wisconsin at Green Bay, and an M.S. in Computer Science from Nova Southeastern University.

@ThingsExpo Stories
In his session at @ThingsExpo, Dr. Robert Cohen, an economist and senior fellow at the Economic Strategy Institute, presented the findings of a series of six detailed case studies of how large corporations are implementing IoT. The session explored how IoT has improved their economic performance, had major impacts on business models and resulted in impressive ROIs. The companies covered span manufacturing and services firms. He also explored servicification, how manufacturing firms shift from se...
The question before companies today is not whether to become intelligent, it’s a question of how and how fast. The key is to adopt and deploy an intelligent application strategy while simultaneously preparing to scale that intelligence. In her session at 21st Cloud Expo, Sangeeta Chakraborty, Chief Customer Officer at Ayasdi, will provide a tactical framework to become a truly intelligent enterprise, including how to identify the right applications for AI, how to build a Center of Excellence to ...
SYS-CON Events announced today that Massive Networks will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Massive Networks mission is simple. To help your business operate seamlessly with fast, reliable, and secure internet and network solutions. Improve your customer's experience with outstanding connections to your cloud.
SYS-CON Events announced today that CAST Software will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 - Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. CAST was founded more than 25 years ago to make the invisible visible. Built around the idea that even the best analytics on the market still leave blind spots for technical teams looking to deliver better software and prevent outages, CAST provides the software intelligence that matter ...
Everything run by electricity will eventually be connected to the Internet. Get ahead of the Internet of Things revolution and join Akvelon expert and IoT industry leader, Sergey Grebnov, in his session at @ThingsExpo, for an educational dive into the world of managing your home, workplace and all the devices they contain with the power of machine-based AI and intelligent Bot services for a completely streamlined experience.
Because IoT devices are deployed in mission-critical environments more than ever before, it’s increasingly imperative they be truly smart. IoT sensors simply stockpiling data isn’t useful. IoT must be artificially and naturally intelligent in order to provide more value In his session at @ThingsExpo, John Crupi, Vice President and Engineering System Architect at Greenwave Systems, will discuss how IoT artificial intelligence (AI) can be carried out via edge analytics and machine learning techn...
SYS-CON Events announced today that Datera, that offers a radically new data management architecture, has been named "Exhibitor" of SYS-CON's 21st International Cloud Expo ®, which will take place on Oct 31 - Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Datera is transforming the traditional datacenter model through modern cloud simplicity. The technology industry is at another major inflection point. The rise of mobile, the Internet of Things, data storage and Big...
In his session at @ThingsExpo, Arvind Radhakrishnen discussed how IoT offers new business models in banking and financial services organizations with the capability to revolutionize products, payments, channels, business processes and asset management built on strong architectural foundation. The following topics were covered: How IoT stands to impact various business parameters including customer experience, cost and risk management within BFS organizations.
SYS-CON Events announced today that CA Technologies has been named "Platinum Sponsor" of SYS-CON's 21st International Cloud Expo®, which will take place October 31-November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. CA Technologies helps customers succeed in a future where every business - from apparel to energy - is being rewritten by software. From planning to development to management to security, CA creates software that fuels transformation for companies in the applic...
SYS-CON Events announced today that Pulzze Systems will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place October 31-November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Pulzze Systems Inc, provides the software product "The Interactor" that uniquely simplifies building IoT, Web and Smart Enterprise Solutions. It is a Silicon Valley startup funded by US government agencies, NSF and DHS to bring innovative solutions to market.
From 2013, NTT Communications has been providing cPaaS service, SkyWay. Its customer’s expectations for leveraging WebRTC technology are not only typical real-time communication use cases such as Web conference, remote education, but also IoT use cases such as remote camera monitoring, smart-glass, and robotic. Because of this, NTT Communications has numerous IoT business use-cases that its customers are developing on top of PaaS. WebRTC will lead IoT businesses to be more innovative and address...
WebRTC is great technology to build your own communication tools. It will be even more exciting experience it with advanced devices, such as a 360 Camera, 360 microphone, and a depth sensor camera. In his session at @ThingsExpo, Masashi Ganeko, a manager at INFOCOM Corporation, will introduce two experimental projects from his team and what they learned from them. "Shotoku Tamago" uses the robot audition software HARK to track speakers in 360 video of a remote party. "Virtual Teleport" uses a...
When shopping for a new data processing platform for IoT solutions, many development teams want to be able to test-drive options before making a choice. Yet when evaluating an IoT solution, it’s simply not feasible to do so at scale with physical devices. Building a sensor simulator is the next best choice; however, generating a realistic simulation at very high TPS with ease of configurability is a formidable challenge. When dealing with multiple application or transport protocols, you would be...
SYS-CON Events announced today that CA Technologies has been named “Platinum Sponsor” of SYS-CON's 21st International Cloud Expo®, which will take place October 31-November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. CA Technologies helps customers succeed in a future where every business – from apparel to energy – is being rewritten by software. From planning to development to management to security, CA creates software that fuels transformation for companies in the applic...
SYS-CON Events announced today that Elastifile will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 - Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Elastifile Cloud File System (ECFS) is software-defined data infrastructure designed for seamless and efficient management of dynamic workloads across heterogeneous environments. Elastifile provides the architecture needed to optimize your hybrid cloud environment, by facilitating efficient...
SYS-CON Events announced today that Golden Gate University will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Since 1901, non-profit Golden Gate University (GGU) has been helping adults achieve their professional goals by providing high quality, practice-based undergraduate and graduate educational programs in law, taxation, business and related professions. Many of its courses are taug...
Recently, WebRTC has a lot of eyes from market. The use cases of WebRTC are expanding - video chat, online education, online health care etc. Not only for human-to-human communication, but also IoT use cases such as machine to human use cases can be seen recently. One of the typical use-case is remote camera monitoring. With WebRTC, people can have interoperability and flexibility for deploying monitoring service. However, the benefit of WebRTC for IoT is not only its convenience and interopera...
When shopping for a new data processing platform for IoT solutions, many development teams want to be able to test-drive options before making a choice. Yet when evaluating an IoT solution, it’s simply not feasible to do so at scale with physical devices. Building a sensor simulator is the next best choice; however, generating a realistic simulation at very high TPS with ease of configurability is a formidable challenge. When dealing with multiple application or transport protocols, you would be...
WebRTC is the future of browser-to-browser communications, and continues to make inroads into the traditional, difficult, plug-in web communications world. The 6th WebRTC Summit continues our tradition of delivering the latest and greatest presentations within the world of WebRTC. Topics include voice calling, video chat, P2P file sharing, and use cases that have already leveraged the power and convenience of WebRTC.
An increasing number of companies are creating products that combine data with analytical capabilities. Running interactive queries on Big Data requires complex architectures to store and query data effectively, typically involving data streams, an choosing efficient file format/database and multiple independent systems that are tied together through custom-engineered pipelines. In his session at @BigDataExpo at @ThingsExpo, Tomer Levi, a senior software engineer at Intel’s Advanced Analytics ...