SYS-CON MEDIA Authors: Liz McMillan, Carmen Gonzalez, Zakia Bouachraoui, Roger Strukhoff, David Linthicum

Blog Feed Post

Protecting SaaS Revenue Through SLA Monitoring

One of the biggest nightmares for any service provider is to find themselves in SLA hell due to poor performance. An issue that negatively impacts end users’ experience is inevitably going to have an effect on a company’s business metrics, and when that happens, they’re going to be looking for someone to blame, and more importantly, and compensate them for that lost revenue.

The reasoning behind having comprehensive SLAs in place is not a difficult concept to grasp. Protection of one’s brand image and revenue stream(s) is obviously of paramount importance. Yet as the landscape of digital architecture grows more and more complex, companies are forced to outsource more functionality to third-party vendors, which in turn creates additional places where performance can go bad.

An SLA is designed to mitigate the risk of that outsourcing by holding vendors financially accountable for any performance degradations that affects the end users through objective SLA monitoring, grading, and governance. According to the 2017 State of SaaS report conducted by Tech Target, over 25 percent of respondents acknowledged that they had incurred financial penalties for failing to meet their SLAs, with the average amount in penalties rising above $350K.

With that much money on the line, the simple truth is that vendors cannot afford to be the cause of their customers’ poor performance.

To make matters worse, more than 10 percent of the respondents admitted that service disruptions led to the loss of a customer, illustrating how much poor performance can erode the trust that’s necessary for a customer-vendor relationship. No business can afford to allow their brand to be harmed by poor customer experiences, so having strict SLAs in place along with diligent SLA monitoring practices becomes an absolute necessity.

The latter part of that strategy – diligent SLA monitoring practices – is dependent upon having a powerful synthetic monitoring solution in place that can replicate the end user experience while measuring from both backbone and last mile locations. The backbone tests, which eliminate noise that is out of the vendor’s control (e.g. local ISP or user hardware issues), are the most valuable for SLA monitoring and validation, while last mile and real user measurements provide additional context by showing the actual end-user experience.

A Two-Pronged Approach to Monitoring

Meanwhile, SaaS vendors themselves must also have end user experience monitoring strategies in place, with a two-pronged approach: one is to ensure the health of their digital supply chain, and the other is to validate their SLA requirements by proving that they are not the cause of any disruptions in their clients’ customer experiences. These two complementary goals ultimately serve the underlying purpose of SLA monitoring – that is to minimize the amount of money penalties that a vendor must pay their customers in penalties.

This is the approach taken by Zscaler, the world’s largest cloud security platform, which helps some of the biggest companies and government agencies around the world securely transform their networks and applications. Given their service offering, Zscaler’s security applications obviously must be placed in the path between the end users and whatever application they’re using (i.e. video conferencing software, banking software, etc.). This means that should Zscaler’s own digital supply chain suffer a service disruption, it will likely cause a negative digital experience for the end user as well.

The Need for Synthetic SLA Monitoring

The prevalence of both first- and third-party services within everyone’s digital supply chain emphasizes the need for complete outside-in view of the end user experience; viewing solely from within one’s own network is incomplete, and only relying on real user monitoring will still leave gaps in visibility when trying to determine the root cause of the issue (i.e. who ultimately bears responsibility for the disruption).

By being able to synthetically test every step of the digital supply chain, a SaaS vendor such as Zscaler is able to spot potential performance degradations before they have an impact on the end user experience, and then drill down into the analytics to pinpoint the root cause of the issue and troubleshoot a solution. This aspect of SLA monitoring is crucial, as it allows Zscaler to head off any problems before they trigger an SLA breach. After all, the best way to avoid paying penalties on your performance is to always have great performance.

There are a number of different ways that Zscaler obtains the real-time, actionable insights that allow them to detect and fix issues as quickly as possible. One crucial aspect is testing from as close as possible to the physical location of the end user(s).

http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-node-map-3... 300w, http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-node-map-7... 768w, http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-node-map.png 1036w" sizes="(max-width: 640px) 100vw, 640px" />

Many performance degradations are localized in specific geographies due to problems with single servers or datacenters, or peering issues with local networks and ISPs. When that’s the case, a performance test run from a different country or on a different ISP isn’t going to give you data that you can act on. Therefore, a testing infrastructure that provides a wide array of locations, ISPs, and cloud networks is vital to ensuring the end user experience.

Another important aspect for diagnosing and fixing performance issues is to have access to a wide range of test types and metrics. Once a performance alert goes off, an IT Ops/SRE must then drill deeper into the data to pinpoint the root cause, often by running different test types depending on the nature of the issue; for example, when an API fails, an API-specific test is in order; to pinpoint a network peering issue, a traceroute test is required.

http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-Traceroute... 300w, http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-Traceroute... 768w" sizes="(max-width: 640px) 100vw, 640px" />

However, effective SLA monitoring is about more than just ensuring that your own services are performing up to standards – it’s also about proving that you’re not responsible for other people’s failures.

SLA Monitoring Through Validation

Anyone who grew up with at least one sibling knows the value of passing the buck when something breaks. You know your little brother was the one who broke that lamp, but of course he doesn’t want to be punished, so he’s going to go out of his way to push the blame onto you. And unless you can prove it, it’s your word against his.

The same principle applies to business and digital performance, albeit with consequences much more severe than an early bedtime. When a company suffers a performance issue that results in loss of revenue and/or brand prestige, they’re naturally going to look for the culprit that’s responsible and tie it to an SLA breach in order to recoup some of that money. They’re going to be armed with data in these attempts, so vendors must be equally armed as well through their own SLA monitoring efforts. The name of the game, as it was when you were a kid, is to prove that it wasn’t your fault.

Once again, the answer lies with deployment of a thorough synthetic monitoring solution that can clearly and definitively articulate the root cause(s) of any performance problems during the post-mortem analysis.

When a vendor such as Zscaler is tasked with proving that they were not the source of a performance problem, one of the most important aspects is to be able to do so through data and charts that are easy to share and understand. Remember that these analyses and the business decisions that result are often being performed by people who don’t have the technical proficiency of a network engineer or SRE, so clear and obvious visual evidence is crucial.

http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-Waterfall-... 300w, http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-Waterfall-... 768w" sizes="(max-width: 640px) 100vw, 640px" />

Another helpful tactic for SLA monitoring is the ability to isolate first- and third-party content, and to be able to identify exactly who is responsible for the performance of all those third-parties. For example, if social sharing tag causes excessive delays in the loading of a website page, your synthetic monitoring solution should be able to pinpoint exactly what the tag is, who hosts it, and how much of a delay it caused.

http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-Zones-300x... 300w, http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-Zones-768x... 768w, http://blog.catchpoint.com/wp-content/uploads/2018/02/Zscaler-Zones.png 1173w" sizes="(max-width: 640px) 100vw, 640px" />

Finally, the ability to filter out extraneous noise through synthetic tests is vital to ensure accurate SLA monitoring. The simple fact is that some performance degradations are out of our hands; they can be caused by a weak home WiFi network, a damaged ISP cable, or something as simple as inclement weather that disrupts a mobile network. Here again, we see the importance of a synthetic “clean-room environment” that just looks at the customer-critical elements in the digital supply chain.

Don’t Get Blamed for Someone Else’s Mistake

The ultimate goal behind any vendor’s SLA monitoring strategy is to ensure that that you minimize the amount of penalties that you have to pay to your clients. With a strong synthetic monitoring platform in place, you should be able to catch issues as soon as they arise and fix them quickly, and demonstrate the root cause of issues that lie beyond your control and for which you are therefore not responsible. This two-pronged approach to SLA monitoring will save your company money in both the short- and long-term, and protect your brand’s prestige at the same time.

The post Protecting SaaS Revenue Through SLA Monitoring appeared first on Catchpoint's Blog - Web Performance Monitoring.

Read the original blog entry...

More Stories By Mehdi Daoudi

Catchpoint radically transforms the way businesses manage, monitor, and test the performance of online applications. Truly understand and improve user experience with clear visibility into complex, distributed online systems.

Founded in 2008 by four DoubleClick / Google executives with a passion for speed, reliability and overall better online experiences, Catchpoint has now become the most innovative provider of web performance testing and monitoring solutions. We are a team with expertise in designing, building, operating, scaling and monitoring highly transactional Internet services used by thousands of companies and impacting the experience of millions of users. Catchpoint is funded by top-tier venture capital firm, Battery Ventures, which has invested in category leaders such as Akamai, Omniture (Adobe Systems), Optimizely, Tealium, BazaarVoice, Marketo and many more.

Latest Stories
The platform combines the strengths of Singtel's extensive, intelligent network capabilities with Microsoft's cloud expertise to create a unique solution that sets new standards for IoT applications," said Mr Diomedes Kastanis, Head of IoT at Singtel. "Our solution provides speed, transparency and flexibility, paving the way for a more pervasive use of IoT to accelerate enterprises' digitalisation efforts. AI-powered intelligent connectivity over Microsoft Azure will be the fastest connected pat...
There are many examples of disruption in consumer space – Uber disrupting the cab industry, Airbnb disrupting the hospitality industry and so on; but have you wondered who is disrupting support and operations? AISERA helps make businesses and customers successful by offering consumer-like user experience for support and operations. We have built the world’s first AI-driven IT / HR / Cloud / Customer Support and Operations solution.
ScaleMP is presenting at CloudEXPO 2019, held June 24-26 in Santa Clara, and we’d love to see you there. At the conference, we’ll demonstrate how ScaleMP is solving one of the most vexing challenges for cloud — memory cost and limit of scale — and how our innovative vSMP MemoryONE solution provides affordable larger server memory for the private and public cloud. Please visit us at Booth No. 519 to connect with our experts and learn more about vSMP MemoryONE and how it is already serving some of...
Darktrace is the world's leading AI company for cyber security. Created by mathematicians from the University of Cambridge, Darktrace's Enterprise Immune System is the first non-consumer application of machine learning to work at scale, across all network types, from physical, virtualized, and cloud, through to IoT and industrial control systems. Installed as a self-configuring cyber defense platform, Darktrace continuously learns what is ‘normal' for all devices and users, updating its understa...
Codete accelerates their clients growth through technological expertise and experience. Codite team works with organizations to meet the challenges that digitalization presents. Their clients include digital start-ups as well as established enterprises in the IT industry. To stay competitive in a highly innovative IT industry, strong R&D departments and bold spin-off initiatives is a must. Codete Data Science and Software Architects teams help corporate clients to stay up to date with the mod...
As you know, enterprise IT conversation over the past year have often centered upon the open-source Kubernetes container orchestration system. In fact, Kubernetes has emerged as the key technology -- and even primary platform -- of cloud migrations for a wide variety of organizations. Kubernetes is critical to forward-looking enterprises that continue to push their IT infrastructures toward maximum functionality, scalability, and flexibility. As they do so, IT professionals are also embr...
Platform9, the leader in SaaS-managed hybrid cloud, has announced it will present five sessions at four upcoming industry conferences in June: BCS in London, DevOpsCon in Berlin, HPE Discover and Cloud Computing Expo 2019.
At CloudEXPO Silicon Valley, June 24-26, 2019, Digital Transformation (DX) is a major focus with expanded DevOpsSUMMIT and FinTechEXPO programs within the DXWorldEXPO agenda. Successful transformation requires a laser focus on being data-driven and on using all the tools available that enable transformation if they plan to survive over the long term. A total of 88% of Fortune 500 companies from a generation ago are now out of business. Only 12% still survive. Similar percentages are found throug...
When you're operating multiple services in production, building out forensics tools such as monitoring and observability becomes essential. Unfortunately, it is a real challenge balancing priorities between building new features and tools to help pinpoint root causes. Linkerd provides many of the tools you need to tame the chaos of operating microservices in a cloud native world. Because Linkerd is a transparent proxy that runs alongside your application, there are no code changes required. I...
In his general session at 21st Cloud Expo, Greg Dumas, Calligo’s Vice President and G.M. of US operations, discussed the new Global Data Protection Regulation and how Calligo can help business stay compliant in digitally globalized world. Greg Dumas is Calligo's Vice President and G.M. of US operations. Calligo is an established service provider that provides an innovative platform for trusted cloud solutions. Calligo’s customers are typically most concerned about GDPR compliance, application p...
Modern software design has fundamentally changed how we manage applications, causing many to turn to containers as the new virtual machine for resource management. As container adoption grows beyond stateless applications to stateful workloads, the need for persistent storage is foundational - something customers routinely cite as a top pain point. In his session at @DevOpsSummit at 21st Cloud Expo, Bill Borsari, Head of Systems Engineering at Datera, explored how organizations can reap the bene...
"NetApp's vision is how we help organizations manage data - delivering the right data in the right place, in the right time, to the people who need it, and doing it agnostic to what the platform is," explained Josh Atwell, Developer Advocate for NetApp, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
Druva is the global leader in Cloud Data Protection and Management, delivering the industry's first data management-as-a-service solution that aggregates data from endpoints, servers and cloud applications and leverages the public cloud to offer a single pane of glass to enable data protection, governance and intelligence-dramatically increasing the availability and visibility of business critical information, while reducing the risk, cost and complexity of managing and protecting it. Druva's...
Kubernetes as a Container Platform is becoming a de facto for every enterprise. In my interactions with enterprises adopting container platform, I come across common questions: - How does application security work on this platform? What all do I need to secure? - How do I implement security in pipelines? - What about vulnerabilities discovered at a later point in time? - What are newer technologies like Istio Service Mesh bring to table?In this session, I will be addressing these commonly asked ...
BMC has unmatched experience in IT management, supporting 92 of the Forbes Global 100, and earning recognition as an ITSM Gartner Magic Quadrant Leader for five years running. Our solutions offer speed, agility, and efficiency to tackle business challenges in the areas of service management, automation, operations, and the mainframe.