Deploying APM in the Enterprise Part 3 – Getting Started With APM

Welcome back to my series on Deploying APM in the Enterprise. In Part 2 we discussed APM maturity from an unusual perspective as maturity models go. In this blog we are going to discuss the process of getting what you want and need in the way of an APM product.

Warning – This is an astronomically long blog post so make sure you have plenty of time to read it through.

Let’s assume for a minute that you actually read Part 2 – APM Maturity As You’ve Probably Never Seen It Before and have an idea where you and your organization currently are on the maturity scale. To advance to a higher level of maturity you can work on process and people until the cows come home but eventually you are going to want a tool that provides insight your people and processes never will. So that is where I am going to focus the rest of this series while touching on the people and process aspect throughout.

If you have decided you need or want to buy a new software tool here is my tried and proven method for making it happen. Just to make things a little more interesting each step in the process is a song title that relates somehow and that I actually like. So feel free to comment on the blog post and my taste in music if you feel like it.

Throughout this blog post I will share some key lessons learned from my own experience dealing with software vendors. This information will be highlighted in bold red text so it is easier to pick out.

It Hurts So Good – John Cougar

Nobody will agree to spend money on a tool unless there is some problem putting the hurt on your business (lost revenue, productivity impact, customer satisfaction, etc…). Find and document a tangible problem. Preferably an issue with a business/mission critical application like your e-commerce platform, online trading, payment gateway, risk calculation, settlement system, etc… Find some application or service that is impacting your business in a meaningful way due to poor performance and/or downtime and document the following:

Number of issues and severity level

Mean Time To Repair (MTTR – usually the average amount of time from first impact to problem resolution)

Quantifiable measure of impact on business (dollars lost per minute, potential customers lost, trades lost per minute, etc…)

Average number of employees involved in troubleshooting each issue

Root cause of each incident

You will use this data in your evaluation document and your business justification down the road.

The World I Know – Collective Soul

“So I walk up on high / And I step to the edge / To see my world below” If you haven’t already done it, you need to take inventory of what you already own and document your findings. You will use this information for years to come as long as you keep it up to date.

What tools exist and what category should they be put in? (Database Monitoring, Network Monitoring, OS Monitoring, Desktop Monitoring, etc…)

How many licenses do we have and are the current?

What are they good at?

What are they not good at?

What would be classified as an APM tool?

If I already have an APM tool why is it not being used properly?

Put labels on your existing tools and understand what they do!

Down in a Hole – Alice In Chains

Now that you have the overall landscape of your monitoring ecosystem laid out you need to see if there are any gaping holes. AppDynamics is an APM company so we suggest you compare your existing tools to the Gartner definition of APM to see what you might be missing. Here it is paraphrased in my own words:

End User Experience Monitoring: Measuring the response time of your application all the way to the end user. It’s not good enough to just understand how fast your application runs within the confines of the data center(s).

Application Topology Mapping: Automatic detection and display of all components involved in the delivery of your application. You need to know what application components are in use at any given time, but especially when there is an issue impacting your users.

Business Transaction Profiling: Detecting and measuring the response time of all application component activity initiated by a single user request. This is not the same as measuring the response time of a web page!!!

Deep Application Diagnostics: Detecting and measuring the run time code execution within your application containers. If your current or prospective solution does not load into the application container you will NOT have this important capability.

Analytics: Intelligence applied to data which provides you with actionable information. This is not the same as reporting and analytics can (and should) be a key differentiator between competing solutions.

You can do this same type of gap analysis for other types of monitoring but you will need to figure out the main aspects of each type for yourself (sorry).

So What’cha Want – The Beastie Boys

So lets assume that you need an APM solution to fill that large void in your monitoring capabilities and to solve that pesky problem which has been lingering for the past 6 months. You know which problem I’m referring to, it’s the one that crops up for a few hours every other week, it wreaks havoc on your business, then it magically disappears after half of your IT staff is on a massive conference call trying to figure out what’s going on. It’s the problem that your boss has to answer for in the daily update calls and can never explain. It’s the problem that will eventually get someone fired or “re-organized” and if you can fix it you will be a hero or a rockstar (you get to choose which term you prefer since you fixed the problem).

In order to pick the right APM solution (picking the wrong one can turn that rockstar potential completely upside down) you need to develop a method for comparing different solutions. First you need to narrow down the crowded list of vendors to a group of 2 or 3 at most to do a Proof Of Concept (POC). This initial narrowing of the field is usually done through feature comparisons (vendor websites), references from people you know, phone calls with vendors, phone calls with analysts (Gartner, Forrester, etc…), crystal balls, voodoo rituals, and possibly animal sacrifice.

Use your favorite spreadsheet program (I always wonder how many people don’t actually use Excel at work) and create a matrix with all of the products you might be interested in (AppDynamics should be first on the list, wink, wink) and compare each product to the level of support they have for each of your requirements. Your requirements at this point should be pretty generic as the detailed requirements are best suited for the POC phase. Here are some requirements to get you started:

Open-ness of vendor (did they skirt around my questions?, did they talk in circles?)

Many more requirements you should add…

One you have this matrix built it should be easy to narrow your choices (hopefully AppDynamics made the short list, wink, wink).

One piece of advice before moving along to the POC requirements section… Be very explicit and detailed with your questions as they apply to your environment. If you get a partial answer from a vendor or a response that is not quite what you are looking for make sure you dig deeper right away to figure out if they are blowing smoke or not. If they keep talking in circle you can be pretty sure they can’t really do what you are asking.

CIO, CTO & Developer Resources

Little Lies – Fleetwood Mac

No, this is not the point where I tell you to start making stuff up so you can get your software purchased. This is the part where I warn you that vendors can get very “creative” with their marketing. Some vendors are worse about this than others but it can be very difficulty to wade through the enormous amounts of BS that are splattered across many vendors websites.

Remember this key principle when dealing with all vendors … Don’t trust anything a vendor tells you, make them show you in a live demo!!!

I’m not saying that vendors are evil liars, but they will give you their best answer to your questions and it might not align with the intent of your question. This has happened to me many times before and I learned the hard way that although a vendors answer was technically true, the answer was far from reality when compared with the intent of my question.

And now back to our regularly scheduled broadcast.

3 is the Magic Number – De La Soul

The POC demands it own set of much more detailed requirements. This is the time to really look under the covers at a product and see what it can do versus what you have already been told. Again, using your favorite spreadsheet tool create a new sheet for each tool in the POC. I prefer to use a weighting and grading scale to help differentiate between tools.

Weighting: Some requirements will be more important to you than others. I use a Low, Medium, and High scale with associated values of 1, 2, and 3. This is your multiplier for the grade you give each requirement

Grading: Some tools have better support for a requirement than others. I use a grading scale of None, Poor, Adequate, and Excellent with associated values of 0, 1, 2, and 3.

I also like to break up the POC requirements into categories. This helps keep me organized, ensures I don’t miss testing something on my list, and helps me write the analysis document at the end of the POC. Some examples of categories and requirements are shown below.

Automated intelligence to ensure instrumentation does not use excessive overhead

Your list should be much longer and more detailed. This requirements list is the basis of the POC and all follow up documentation so make sure it is thorough and not slanted towards a particular vendor. Think of it as your Christmas list when you were a kid, don’t be afraid to ask for things that might seem impossible but that could be really useful.

Another key lesson learned coming your way… Don’t let the vendor control the POC. You define the environment (Dev, Test, and even Prod if it can be done safely), you do the installation, you do the configuration, you do everything related to the POC. You are the one who has to use the tool after you buy it so be sure to personally do everything during the POC.

Be sure you provide the same playing field for each vendor so that your results are really comparable. Having a successful POC using a team of 5 vendor engineers versus another successful POC using just 1 engineer is comparing apples and oranges.

We Are the Champions – Queen

After you wrap up all of your POCs you should have enough data to pick a winner. Hopefully you built your spreadsheets so that they automatically add up the numbers related to all of the requirements. You will have a statistical winner based purely upon you spreadsheet data and that usually aligns with the overall feeling you have after the POC is complete but sometimes it might not. You may have a vendor/product that is statistically the winner but your gut is telling you to choose a different vendor. In this case you need to figure out exactly why you feel this way. You cannot justify vendor selection based upon a gut instinct with no explanation. Maybe one vendor was just a royal pain to deal with, maybe the solution worked well but was agonizingly painful to deploy. The key here is being able to express why your instinct is pushing you a certain direction and quantify the potential impact. Something similar to the following statement can be used as justification of your position:

“Even though Vendor X Product statistically scored highest in the evaluation there is one overriding factor that prohibits selecting them as the overall winner. Deployment and configuration of Vendor X Product is difficult and time consuming. Based upon the observed deployment and configuration time of 2 weeks for 1 application during the POC it would take approximately 19 years to configure monitoring for the 500 applications which are in scope. Deploying Vendor X Product does not make sense in our environment.”

The information you have been building throughout this entire process should be used to create an overall evaluation document. This document should have the following information at a minimum:

Description of problem

Description of proposed solution

Vendors/Products evaluated

Evaluation Criterial (Requirements)

Evaluation Results

Recommendation

Next steps

It can also be helpful to create a short (3–10 pages) presentation to accompany the evaluation document which you can use to brief management on your findings. This presentation should contain only the most important facts since you have all the detail anyone should ever want in the full evaluation document.

Another important document you can create is the business justification. The business justification cuts out all the technical details related to the product you want to purchase and gets right down to the economics of the matter. I am not going to dive deep into how to write a business justification but to help you get started you should make sure you get a ROI (Return On Investment) calculator from each vendor that participates in your POC. Vendors want to help you buy their solution and have a wealth of information available to help you build your business justification so just ask them for help if you need it. Just make sure that everything in the business justification is factually based and relevant to your business.

Lastly it is really helpful to have broad support for your initiative. Seek out people across your organization that will support your recommendation of product or who will validate the problem that you are trying to solve. If you have full support of an Application Owner whose business is being directly impacted this greatly increases your chances of success.

If you read this entire blog post please accept my gratitude! I know it was a long post but it’s a huge topic to cover. Hopefully you picked up some good information and will return for my next post about deploying the product you just worked so hard to get in the door. As always, your comments are welcome. I’d love to hear your tips for getting new solutions approved or any war stories from your vendor evaluations.

In high-production environments where release cycles are measured in hours or minutes — not days or weeks — there's little room for mistakes and no room for confusion. Everyone has to understand what's happening, in real time, and have the means to do whatever is necessary to keep applications up and running optimally.

DevOps is a high-stakes world, but done well, it delivers the agility and performance to significantly impact business competitiveness.

The industrial software market has treated data with the mentality of “collect everything now, worry about how to use it later.” We now find ourselves buried in data, with the pervasive connectivity of the (Industrial) Internet of Things only piling on more numbers. There’s too much data and not enough information.
In his session at @ThingsExpo, Bob Gates, Global Marketing Director, GE’s Intelligent Platforms business, to discuss how realizing the power of IoT, software developers are now focu...

SYS-CON Events announced today that Dyn, the worldwide leader in Internet Performance, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY.
Dyn is a cloud-based Internet Performance company. Dyn helps companies monitor, control, and optimize online infrastructure for an exceptional end-user experience. Through a world-class network and unrivaled, objective intelligence into Internet conditions, Dyn ensures...

As organizations shift toward IT-as-a-service models, the need for managing and protecting data residing across physical, virtual, and now cloud environments grows with it. CommVault can ensure protection &E-Discovery of your data – whether in a private cloud, a Service Provider delivered public cloud, or a hybrid cloud environment – across the heterogeneous enterprise.
In his session at 16th Cloud Expo, Randy De Meno, Chief Technologist - Windows Products and Microsoft Partnerships, will disc...

Skytap Inc., has appointed David Frost as vice president of professional services. David joins Skytap from Deloitte Consulting where he served as Managing Director leading SAP, Cloud, and Advanced Technology Services. At Skytap, David will head the company's professional services organization, and spearhead a new consulting practice that will guide IT organizations through the adoption of DevOps best practices. David's appointment comes on the heels of Skytap's recent $35 million Series D fundin...

Business and IT leaders today need better application delivery capabilities to support critical new innovation. But how often do you hear objections to improving application delivery like, “I can harden it against attack, but not on this timeline”; “I can make it better, but it will cost more”; “I can deliver faster, but not with these specs”; or “I can stay strong on cost control, but quality will suffer”? In the new application economy, these tradeoffs are no longer acceptable. Customers will ...

Hadoop as a Service (as offered by handful of niche vendors now) is a cloud computing solution that makes medium and large-scale data processing accessible, easy, fast and inexpensive.
In his session at Big Data Expo, Kumar Ramamurthy, Vice President and Chief Technologist, EIM & Big Data, at Virtusa, will discuss how this is achieved by eliminating the operational challenges of running Hadoop, so one can focus on business growth. The fragmented Hadoop distribution world and various PaaS soluti...

Red Hat has launched the Red Hat Cloud Innovation Practice, a new global team of experts that will assist companies with more quickly on-ramping to the cloud. They will do this by providing solutions and services such as validated designs with reference architectures and agile methodology consulting, training, and support.
The Red Hat Cloud Innovation Practice is born out of the integration of technology and engineering expertise gained through the company’s 2014 acquisitions of leading Ceph s...

Even as cloud and managed services grow increasingly central to business strategy and performance, challenges remain. The biggest sticking point for companies seeking to capitalize on the cloud is data security. Keeping data safe is an issue in any computing environment, and it has been a focus since the earliest days of the cloud revolution. Understandably so: a lot can go wrong when you allow valuable information to live outside the firewall. Recent revelations about government snooping, along...

Operational Hadoop and the Lambda Architecture for Streaming Data
Apache Hadoop is emerging as a distributed platform for handling large and fast incoming streams of data. Predictive maintenance, supply chain optimization, and Internet-of-Things analysis are examples where Hadoop provides the scalable storage, processing, and analytics platform to gain meaningful insights from granular data that is typically only valuable from a large-scale, aggregate view. One architecture useful for capturing...

SYS-CON Events announced today that Vitria Technology, Inc. will exhibit at SYS-CON’s @ThingsExpo, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY.
Vitria will showcase the company’s new IoT Analytics Platform through live demonstrations at booth #330. Vitria’s IoT Analytics Platform, fully integrated and powered by an operational intelligence engine, enables customers to rapidly build and operationalize advanced analytics to deliver timely business outcomes ...

The Internet of Things (IoT) promises to evolve the way the world does business; however, understanding how to apply it to your company can be a mystery. Most people struggle with understanding the potential business uses or tend to get caught up in the technology, resulting in solutions that fail to meet even minimum business goals.
In his session at @ThingsExpo, Jesse Shiah, CEO / President / Co-Founder of AgilePoint Inc., showed what is needed to leverage the IoT to transform your business. ...

DevOps is about increasing efficiency, but nothing is more inefficient than building the same application twice. However, this is a routine occurrence with enterprise applications that need both a rich desktop web interface and strong mobile support. With recent technological advances from Isomorphic Software and others, it is now feasible to create a rich desktop and tuned mobile experience with a single codebase, without compromising performance or usability.

Advanced Persistent Threats (APTs) are increasing at an unprecedented rate. The threat landscape of today is drastically different than just a few years ago. Attacks are much more organized and sophisticated. They are harder to detect and even harder to anticipate. In the foreseeable future it's going to get a whole lot harder. Everything you know today will change. Keeping up with this changing landscape is already a daunting task. Your organization needs to use the latest tools, methods and ex...

HP and Aruba Networks on Monday announced a definitive agreement for HP to acquire Aruba, a provider of next-generation network access solutions for the mobile enterprise, for $24.67 per share in cash. The equity value of the transaction is approximately $3.0 billion, and net of cash and debt approximately $2.7 billion. Both companies' boards of directors have approved the deal.
"Enterprises are facing a mobile-first world and are looking for solutions that help them transition legacy investme...

Containers and microservices have become topics of intense interest throughout the cloud developer and enterprise IT communities.
Accordingly, attendees at the upcoming 16th Cloud Expo at the Javits Center in New York June 9-11 will find fresh new content in a new track called PaaS | Containers & Microservices
Containers are not being considered for the first time by the cloud community, but a current era of re-consideration has pushed them to the top of the cloud agenda. With the launch ...

The Workspace-as-a-Service (WaaS) market will grow to $6.4B by 2018. In his session at 16th Cloud Expo, Seth Bostock, CEO of IndependenceIT, will begin by walking the audience through the evolution of Workspace as-a-Service, where it is now vs. where it going.
To look beyond the desktop we must understand exactly what WaaS is, who the users are, and where it is going in the future. IT departments, ISVs and service providers must look to workflow and automation capabilities to adapt to growing ...

SYS-CON Events announced today Arista Networks will exhibit at SYS-CON's DevOps Summit 2015 New York, which will take place June 9-11, 2015, at the Javits Center in New York City, NY.
Arista Networks was founded to deliver software-driven cloud networking solutions for large data center and computing environments. Arista’s award-winning 10/40/100GbE switches redefine scalability, robustness, and price-performance, with over 3,000 customers and more than three million cloud networking ports depl...

The speed of software changes in growing and large scale rapid-paced DevOps environments presents a challenge for continuous testing. Many organizations struggle to get this right. Practices that work for small scale continuous testing may not be sufficient as the requirements grow.
In his session at DevOps Summit, Marc Hornbeek, Sr. Solutions Architect of DevOps continuous test solutions at Spirent Communications, will explain the best practices of continuous testing at high scale, which is r...

Disruptive macro trends in technology are impacting and dramatically changing the "art of the possible" relative to supply chain management practices through the innovative use of IoT, cloud, machine learning and Big Data to enable connected ecosystems of engagement. Enterprise informatics can now move beyond point solutions that merely monitor the past and implement integrated enterprise fabrics that enable end-to-end supply chain visibility to improve customer service delivery and optimize sup...

The speed of product development has increased massively in the past 10 years. At the same time our formal secure development and SDL methodologies have fallen behind. This forces product developers to choose between rapid release times and security.
In his session at DevOps Summit, Michael Murray, Director of Cyber Security Consulting and Assessment at GE Healthcare, examined the problems and presented some solutions for moving security into the DevOps lifecycle to ensure that we get fast AND ...

An anatomy of startup ventures for the Internet of Things market. Like GE describes in their white paper Pushing the Boundaries of Mind and Machine, this is basically a process of innovating through more intelligent machines to reinvent workflow models.
For a useful overview as to what constitutes an ‘IoT startup’, check out one example for some key characteristics: Hutgrip. Hutgrip is a SaaS solution that replaces VPNs with the Cloud and real time analytics, with the headline points being:
Clear description of the business benefit the new technology will bring – Smarter automation of bi...

The Internet of Things has emerged as the universally accepted term for the ‘next big thing’ wave, not replacing but building upon the Cloud Computing cycle, which itself built upon SaaS and ASPs.
There are many technology aspects to this trend, which will be covered extensively throughout this guide and ongoing series, but overall our goal is to describe the associated startup venture opportunities.
Indeed it’s not limited to startups, the IoT represents a new product innovation platform for any and all businesses, and this is the overall theme of this paper.

Containers and microservices have become topics of intense interest throughout the cloud developer and enterprise IT communities.
Accordingly, attendees at the upcoming 16th Cloud Expo at the Javits Center in New York June 9-11 will find fresh new content in a new track called PaaS | Containers & Microservices
Containers are not being considered for the first time by the cloud community, but a current era of re-consideration has pushed them to the top of the cloud agenda. With the launch of Docker's initial release in March of 2013, interest was revved up several notches. Then late last...

Over the last couple of years I have talked to numerous enterprise customers, analysts, industry pundits, and others interested in cloud technologies, and one thing is abundantly clear – Platform-as-a-Service (PaaS) seems to mean different things to different people. But the term PaaS is irrelevant – it's just noise. What is relevant, and what is important, is what PaaS does: enable applications. That's what enterprises care about. They want to accelerate application development to get products to market faster and into users' hands sooner.

RealTime Medicare Data analyzes huge volumes of Medicare data and provides analysis to their many customers on the caregiver side of the healthcare sector using HP Vertica.
Here to explain how they manage such large data requirements for quality, speed, and volume, we're joined by Scott Hannon, CIO of RealTime Medicare Data and he's based in Birmingham, Alabama. The discussion is moderated by me, Dana Gardner, Principal Analyst at Interarbor Solutions.

Welcome to my four-part series on what I’m going to call the Art of DevOps. We will embark on a mission to reveal the extremely valuable intelligence that’s been collected about a unique strategy to continuously deliver assets to the operational battleground safely, securely and quickly. This strategy drives optimal monitoring of the frontlines and enhanced communications with the troops supporting the initial development. I do make the assumption that most of you are battle-worthy veterans in one or more of the environments that I’ll review.
Beyond this introduction blog, I’ll brief you on ...

Creating global change that is actually good for the entire world is a mammoth task. With a population of almost 7 Billion people as of 2015, the planet is taking a toll with surviving the brunt of keeping the works going. What role can Cloud Computing play in making it easier for all of us?

Although programmers for decades have been starting their careers by writing a “Hello World” program, the phenomenon has remained as unknown to the general public as the song of the same name by Lady Antebellum. But the former is now about to change.
Programming is in fact the real “new way of working”. For years education for the general public was largely confined to learning how to use computers. Thus we created operators rather than programmers. And today operators of all kinds are rapidly being replaced by automation. Just look at the big banks where people who “worked with computers” ...

One important differentiator between what passed for digital back in the dot-com days and today’s notion of digital is the role mobile plays. Yes, this company had a mobile site, and they had what the AVP claimed was a “mobile first” plan for their web content, but as yet they had yet to roll out any responsive design. In the final analysis, their digital effort up to this point boiled down to little more than better brochureware, a la 1990s web redesigns.
But more significantly, what was entirely missing from their digital achievements (although the AVP did indicate that it was a roadmap i...

In September, Macy’s announced that they will invest $1 Billion into their omni-channel strategy. When spending so much money the question that immediately comes up is how to measure the success? Key questions such as “Are conversion rates increasing as planned?” and “How good is the User Experience for each channel?” need answers. Since my first blog about omni-channel monitoring, I came across a variety of customer implementations. Today, I want to share with you how Solocal Group, provider of the French Yellow Pages (PagesJaunes), brings all key metrics together on a single dashboard....

CIOs are increasingly concerned about cloud security. And they should be: with the recent outbreak of visible breaches at high-profile organizations like Target, Anthem, and others, and the subsequent damage they cause, corporations are scrambling to make sure their cloud applications, whether on private, public, or hybrid clouds, are safe.
But cloud security is complex. With ephemeral applications and services springing up around multiple data centers, with dozens or hundreds of independent microservices each with their own access mechanisms, with the widespread adoption of virtualization an...

One of the neat things about microservices is the ability to segment functional actions into scalability domains. Login, browsing, and checkout are separate functional domains that can each be scaled according to demand. While one hopes that checkout is similarly in demand, it is unlikely to be as popular as browsing, after all, and the days of wasting expensive money on idle compute resources went out when the clouds descended.
In that same vein comes the ability to also create performance domains. After all, if you're scaling out a specific functional service domain you can also specify p...

It is no surprise to anyone that service providers need to find new sources of revenue and increase profitability. The digital, cloud and as-a-service revolution provides a silver lining.
As IT organizations feel the tension that comes from a combination of aging legacy B2B infrastructure, changing business mandates and rapidly evolving e-commerce requirements, they are increasingly looking at digital services and outsourcing to trusted providers. They want a trusted partner to deliver connected digital services; including mobile, cloud and M2M/IoT.
The pressure is on for businesses to thin...

When people talk about the Internet of Things (IoT) they tend to think about big data technologies like Hadoop where petabyte size datasets are store and analyzed for both known and unknown patterns. What many people don’t realize is that many IoT use cases only require small datasets.

I attended a Meetup yesterday in Mountain View, hosted by The Hive group on the subject of Lambda Architecture. Since I had never heard about this new phrase, my curiosity took me there. There was a panel discussion and panelists came from Hortonworks, Cloudera, MapR, Teradata, etc.
Lambda Architecture is a useful framework to think about designing big data applications. Nathan Marz designed this generic architecture addressing common requirements for big data based on his experience working on distributed data processing systems at Twitter. Some of the key requirements in building this archi...

The competition among public cloud providers is red hot, private cloud continues to grab increasing shares of IT budgets, and hybrid cloud strategies are beginning to conquer the enterprise IT world.

Big Data is driving dramatic leaps in resource requirements and capabilities, and now the Internet of Things promises an exponential leap in the size of the Internet and Worldwide Web.

The world of SDX now encompasses Software-Defined Data Centers (SDDCs) as the technology world prepares for the Zettabyte Age.

Add the key topics of WebRTC and DevOps into the mix, and you have three days of pure cloud computing that you simply cannot miss.

Cloud Expo - the world's most established event - offers a vast selection of 130+ technical and strategic Industry Keynotes, General Sessions, Breakout Sessions, and signature Power Panels. The exhibition floor features 100+ exhibitors offering specific solutions and comprehensive strategies. The floor also features two Demo Theaters that give delegates the opportunity to get even closer to the technology they want to see and the people who offer it.

Attend Cloud Expo. Craft your own custom experience. Learn the latest from the world's best technologists. Find the vendors you want and put them to the test.