Deploying APM in the Enterprise Part 3 – Getting Started With APM

Welcome back to my series on Deploying APM in the Enterprise. In Part 2 we discussed APM maturity from an unusual perspective as maturity models go. In this blog we are going to discuss the process of getting what you want and need in the way of an APM product.

Warning – This is an astronomically long blog post so make sure you have plenty of time to read it through.

Let’s assume for a minute that you actually read Part 2 – APM Maturity As You’ve Probably Never Seen It Before and have an idea where you and your organization currently are on the maturity scale. To advance to a higher level of maturity you can work on process and people until the cows come home but eventually you are going to want a tool that provides insight your people and processes never will. So that is where I am going to focus the rest of this series while touching on the people and process aspect throughout.

If you have decided you need or want to buy a new software tool here is my tried and proven method for making it happen. Just to make things a little more interesting each step in the process is a song title that relates somehow and that I actually like. So feel free to comment on the blog post and my taste in music if you feel like it.

Throughout this blog post I will share some key lessons learned from my own experience dealing with software vendors. This information will be highlighted in bold red text so it is easier to pick out.

It Hurts So Good – John Cougar

Nobody will agree to spend money on a tool unless there is some problem putting the hurt on your business (lost revenue, productivity impact, customer satisfaction, etc…). Find and document a tangible problem. Preferably an issue with a business/mission critical application like your e-commerce platform, online trading, payment gateway, risk calculation, settlement system, etc… Find some application or service that is impacting your business in a meaningful way due to poor performance and/or downtime and document the following:

Number of issues and severity level

Mean Time To Repair (MTTR – usually the average amount of time from first impact to problem resolution)

Quantifiable measure of impact on business (dollars lost per minute, potential customers lost, trades lost per minute, etc…)

Average number of employees involved in troubleshooting each issue

Root cause of each incident

You will use this data in your evaluation document and your business justification down the road.

The World I Know – Collective Soul

“So I walk up on high / And I step to the edge / To see my world below” If you haven’t already done it, you need to take inventory of what you already own and document your findings. You will use this information for years to come as long as you keep it up to date.

What tools exist and what category should they be put in? (Database Monitoring, Network Monitoring, OS Monitoring, Desktop Monitoring, etc…)

How many licenses do we have and are the current?

What are they good at?

What are they not good at?

What would be classified as an APM tool?

If I already have an APM tool why is it not being used properly?

Put labels on your existing tools and understand what they do!

Down in a Hole – Alice In Chains

Now that you have the overall landscape of your monitoring ecosystem laid out you need to see if there are any gaping holes. AppDynamics is an APM company so we suggest you compare your existing tools to the Gartner definition of APM to see what you might be missing. Here it is paraphrased in my own words:

End User Experience Monitoring: Measuring the response time of your application all the way to the end user. It’s not good enough to just understand how fast your application runs within the confines of the data center(s).

Application Topology Mapping: Automatic detection and display of all components involved in the delivery of your application. You need to know what application components are in use at any given time, but especially when there is an issue impacting your users.

Business Transaction Profiling: Detecting and measuring the response time of all application component activity initiated by a single user request. This is not the same as measuring the response time of a web page!!!

Deep Application Diagnostics: Detecting and measuring the run time code execution within your application containers. If your current or prospective solution does not load into the application container you will NOT have this important capability.

Analytics: Intelligence applied to data which provides you with actionable information. This is not the same as reporting and analytics can (and should) be a key differentiator between competing solutions.

You can do this same type of gap analysis for other types of monitoring but you will need to figure out the main aspects of each type for yourself (sorry).

So What’cha Want – The Beastie Boys

So lets assume that you need an APM solution to fill that large void in your monitoring capabilities and to solve that pesky problem which has been lingering for the past 6 months. You know which problem I’m referring to, it’s the one that crops up for a few hours every other week, it wreaks havoc on your business, then it magically disappears after half of your IT staff is on a massive conference call trying to figure out what’s going on. It’s the problem that your boss has to answer for in the daily update calls and can never explain. It’s the problem that will eventually get someone fired or “re-organized” and if you can fix it you will be a hero or a rockstar (you get to choose which term you prefer since you fixed the problem).

In order to pick the right APM solution (picking the wrong one can turn that rockstar potential completely upside down) you need to develop a method for comparing different solutions. First you need to narrow down the crowded list of vendors to a group of 2 or 3 at most to do a Proof Of Concept (POC). This initial narrowing of the field is usually done through feature comparisons (vendor websites), references from people you know, phone calls with vendors, phone calls with analysts (Gartner, Forrester, etc…), crystal balls, voodoo rituals, and possibly animal sacrifice.

Use your favorite spreadsheet program (I always wonder how many people don’t actually use Excel at work) and create a matrix with all of the products you might be interested in (AppDynamics should be first on the list, wink, wink) and compare each product to the level of support they have for each of your requirements. Your requirements at this point should be pretty generic as the detailed requirements are best suited for the POC phase. Here are some requirements to get you started:

Open-ness of vendor (did they skirt around my questions?, did they talk in circles?)

Many more requirements you should add…

One you have this matrix built it should be easy to narrow your choices (hopefully AppDynamics made the short list, wink, wink).

One piece of advice before moving along to the POC requirements section… Be very explicit and detailed with your questions as they apply to your environment. If you get a partial answer from a vendor or a response that is not quite what you are looking for make sure you dig deeper right away to figure out if they are blowing smoke or not. If they keep talking in circle you can be pretty sure they can’t really do what you are asking.

CIO, CTO & Developer Resources

Little Lies – Fleetwood Mac

No, this is not the point where I tell you to start making stuff up so you can get your software purchased. This is the part where I warn you that vendors can get very “creative” with their marketing. Some vendors are worse about this than others but it can be very difficulty to wade through the enormous amounts of BS that are splattered across many vendors websites.

Remember this key principle when dealing with all vendors … Don’t trust anything a vendor tells you, make them show you in a live demo!!!

I’m not saying that vendors are evil liars, but they will give you their best answer to your questions and it might not align with the intent of your question. This has happened to me many times before and I learned the hard way that although a vendors answer was technically true, the answer was far from reality when compared with the intent of my question.

And now back to our regularly scheduled broadcast.

3 is the Magic Number – De La Soul

The POC demands it own set of much more detailed requirements. This is the time to really look under the covers at a product and see what it can do versus what you have already been told. Again, using your favorite spreadsheet tool create a new sheet for each tool in the POC. I prefer to use a weighting and grading scale to help differentiate between tools.

Weighting: Some requirements will be more important to you than others. I use a Low, Medium, and High scale with associated values of 1, 2, and 3. This is your multiplier for the grade you give each requirement

Grading: Some tools have better support for a requirement than others. I use a grading scale of None, Poor, Adequate, and Excellent with associated values of 0, 1, 2, and 3.

I also like to break up the POC requirements into categories. This helps keep me organized, ensures I don’t miss testing something on my list, and helps me write the analysis document at the end of the POC. Some examples of categories and requirements are shown below.

Automated intelligence to ensure instrumentation does not use excessive overhead

Your list should be much longer and more detailed. This requirements list is the basis of the POC and all follow up documentation so make sure it is thorough and not slanted towards a particular vendor. Think of it as your Christmas list when you were a kid, don’t be afraid to ask for things that might seem impossible but that could be really useful.

Another key lesson learned coming your way… Don’t let the vendor control the POC. You define the environment (Dev, Test, and even Prod if it can be done safely), you do the installation, you do the configuration, you do everything related to the POC. You are the one who has to use the tool after you buy it so be sure to personally do everything during the POC.

Be sure you provide the same playing field for each vendor so that your results are really comparable. Having a successful POC using a team of 5 vendor engineers versus another successful POC using just 1 engineer is comparing apples and oranges.

We Are the Champions – Queen

After you wrap up all of your POCs you should have enough data to pick a winner. Hopefully you built your spreadsheets so that they automatically add up the numbers related to all of the requirements. You will have a statistical winner based purely upon you spreadsheet data and that usually aligns with the overall feeling you have after the POC is complete but sometimes it might not. You may have a vendor/product that is statistically the winner but your gut is telling you to choose a different vendor. In this case you need to figure out exactly why you feel this way. You cannot justify vendor selection based upon a gut instinct with no explanation. Maybe one vendor was just a royal pain to deal with, maybe the solution worked well but was agonizingly painful to deploy. The key here is being able to express why your instinct is pushing you a certain direction and quantify the potential impact. Something similar to the following statement can be used as justification of your position:

“Even though Vendor X Product statistically scored highest in the evaluation there is one overriding factor that prohibits selecting them as the overall winner. Deployment and configuration of Vendor X Product is difficult and time consuming. Based upon the observed deployment and configuration time of 2 weeks for 1 application during the POC it would take approximately 19 years to configure monitoring for the 500 applications which are in scope. Deploying Vendor X Product does not make sense in our environment.”

The information you have been building throughout this entire process should be used to create an overall evaluation document. This document should have the following information at a minimum:

Description of problem

Description of proposed solution

Vendors/Products evaluated

Evaluation Criterial (Requirements)

Evaluation Results

Recommendation

Next steps

It can also be helpful to create a short (3–10 pages) presentation to accompany the evaluation document which you can use to brief management on your findings. This presentation should contain only the most important facts since you have all the detail anyone should ever want in the full evaluation document.

Another important document you can create is the business justification. The business justification cuts out all the technical details related to the product you want to purchase and gets right down to the economics of the matter. I am not going to dive deep into how to write a business justification but to help you get started you should make sure you get a ROI (Return On Investment) calculator from each vendor that participates in your POC. Vendors want to help you buy their solution and have a wealth of information available to help you build your business justification so just ask them for help if you need it. Just make sure that everything in the business justification is factually based and relevant to your business.

Lastly it is really helpful to have broad support for your initiative. Seek out people across your organization that will support your recommendation of product or who will validate the problem that you are trying to solve. If you have full support of an Application Owner whose business is being directly impacted this greatly increases your chances of success.

If you read this entire blog post please accept my gratitude! I know it was a long post but it’s a huge topic to cover. Hopefully you picked up some good information and will return for my next post about deploying the product you just worked so hard to get in the door. As always, your comments are welcome. I’d love to hear your tips for getting new solutions approved or any war stories from your vendor evaluations.

In high-production environments where release cycles are measured in hours or minutes — not days or weeks — there's little room for mistakes and no room for confusion. Everyone has to understand what's happening, in real time, and have the means to do whatever is necessary to keep applications up and running optimally.

DevOps is a high-stakes world, but done well, it delivers the agility and performance to significantly impact business competitiveness.

The deluge of IoT sensor data collected from connected devices and the powerful AI required to make that data actionable are giving rise to a hybrid ecosystem in which cloud, on-prem and edge processes become interweaved. Attendees will learn how emerging composable infrastructure solutions deliver the adaptive architecture needed to manage this new data reality. Machine learning algorithms can better anticipate data storms and automate resources to support surges, including fully scalable GPU-c...

Machine learning has taken residence at our cities' cores and now we can finally have "smart cities." Cities are a collection of buildings made to provide the structure and safety necessary for people to function, create and survive. Buildings are a pool of ever-changing performance data from large automated systems such as heating and cooling to the people that live and work within them. Through machine learning, buildings can optimize performance, reduce costs, and improve occupant comfort by ...

The explosion of new web/cloud/IoT-based applications and the data they generate are transforming our world right before our eyes. In this rush to adopt these new technologies, organizations are often ignoring fundamental questions concerning who owns the data and failing to ask for permission to conduct invasive surveillance of their customers. Organizations that are not transparent about how their systems gather data telemetry without offering shared data ownership risk product rejection, regu...

René Bostic is the Technical VP of the IBM Cloud Unit in North America. Enjoying her career with IBM during the modern millennial technological era, she is an expert in cloud computing, DevOps and emerging cloud technologies such as Blockchain. Her strengths and core competencies include a proven record of accomplishments in consensus building at all levels to assess, plan, and implement enterprise and cloud computing solutions.
René is a member of the Society of Women Engineers (SWE) and a m...

Poor data quality and analytics drive down business value. In fact, Gartner estimated that the average financial impact of poor data quality on organizations is $9.7 million per year. But bad data is much more than a cost center. By eroding trust in information, analytics and the business decisions based on these, it is a serious impediment to digital transformation.

Predicting the future has never been more challenging - not because of the lack of data but because of the flood of ungoverned and risk laden information. Microsoft states that 2.5 exabytes of data are created every day. Expectations and reliance on data are being pushed to the limits, as demands around hybrid options continue to grow.

Digital Transformation and Disruption, Amazon Style - What You Can Learn. Chris Kocher is a co-founder of Grey Heron, a management and strategic marketing consulting firm. He has 25+ years in both strategic and hands-on operating experience helping executives and investors build revenues and shareholder value. He has consulted with over 130 companies on innovating with new business models, product strategies and monetization. Chris has held management positions at HP and Symantec in addition to ...

Enterprises have taken advantage of IoT to achieve important revenue and cost advantages. What is less apparent is how incumbent enterprises operating at scale have, following success with IoT, built analytic, operations management and software development capabilities - ranging from autonomous vehicles to manageable robotics installations. They have embraced these capabilities as if they were Silicon Valley startups.

As IoT continues to increase momentum, so does the associated risk. Secure Device Lifecycle Management (DLM) is ranked as one of the most important technology areas of IoT. Driving this trend is the realization that secure support for IoT devices provides companies the ability to deliver high-quality, reliable, secure offerings faster, create new revenue streams, and reduce support costs, all while building a competitive advantage in their markets. In this session, we will use customer use cases...

The deluge of IoT sensor data collected from connected devices and the powerful AI required to make that data actionable are giving rise to a hybrid ecosystem in which cloud, on-prem and edge processes become interweaved. Attendees will learn how emerging composable infrastructure solutions deliver the adaptive architecture needed to manage this new data reality. Machine learning algorithms can better anticipate data storms and automate resources to support surges, including fully scalable GPU-centric compute for the most data-intensive applications. Hyperconverged systems already in place can...

The explosion of new web/cloud/IoT-based applications and the data they generate are transforming our world right before our eyes. In this rush to adopt these new technologies, organizations are often ignoring fundamental questions concerning who owns the data and failing to ask for permission to conduct invasive surveillance of their customers. Organizations that are not transparent about how their systems gather data telemetry without offering shared data ownership risk product rejection, regulatory scrutiny and increasing consumer lack of trust in technology in general.

René Bostic is the Technical VP of the IBM Cloud Unit in North America. Enjoying her career with IBM during the modern millennial technological era, she is an expert in cloud computing, DevOps and emerging cloud technologies such as Blockchain. Her strengths and core competencies include a proven record of accomplishments in consensus building at all levels to assess, plan, and implement enterprise and cloud computing solutions.
René is a member of the Society of Women Engineers (SWE) and a member of the Society of Information Management (SIM) Atlanta Chapter. She received a Business and Ec...

Cloud computing budgets worldwide are reaching into the hundreds of billions of dollars, and no organization can survive long without some sort of cloud migration strategy. Each month brings new announcements, use cases, and success stories.