Resources

FILTERS

Resources

Blog Post 5 min read

KPI Analysis: the Secret to Tracking and Monitoring Your Goals Lies in AI Anomaly Detection

KPIs are the Cliff’s Notes of business metrics: a handpicked selection of the many measurable quantities which a business can - or does - collect. They can provide vital clues as to how well the business is doing as a whole, for instance which marketing initiatives are and aren’t working, and thus they provide feedback and can guide decisions on how resources are expended, for instance discontinuing a promotional offer that’s not driving any new sales. KPIs are considered “key” often because managers deem them directly related to profit, which is why metrics such as year over year same-store sales are such a common KPI. In order to tell a clear and convincing story, a given KPI needs to be transformed by analysis and visualization from raw data into business insight, and the correct way to do so isn’t always obvious. This is why traditional KPI analysis requires human data science knowledge and is very iterative, two aspects which prevent discovering actionable insights in real time. Traditional KPI Analysis In order to understand how Anodot’s full-stack solution makes dashboard-based KPI analysis tools obsolete, it’s helpful to review the typical KPI workflow which relies on those tools. First, the underlying data for a KPI needs to be pulled in from whatever files or databases they live in. Then, that data needs to be structured or formatted in a way that makes analysis smooth. This step is called “data cleaning” and is often required if the files containing the data for the KPI have rows that serve as table titles or headers, or if data from two sources needs to be combined before analysis. Finally, the data is imported into the dashboarding tool, where it is analyzed and visualized. Usually, the analysis and visualization isn’t a single step, but rather multiple iterations of reviewing, giving feedback and implementing changes. This is due to the fact that a whole team of analysts and data scientists often work on a single KPI, and the multiple iterations are needed before everyone on the team is confident that the visualization of the analyzed data will meet the requirements of management. Although there are obvious benefits to this collaborative approach, it is very time-consuming and takes data scientists and analysts - of which companies have a limited supply - off of other tasks. Furthermore, even after a KPI visualization is created and added to a dashboard, it may be modified or even replaced by another one when it is reevaluated as time goes on and business goals are (or aren’t) met. Once created, the dashboard visualization of the KPI is then monitored either manually or semi-manually via the use of static thresholds and alerts. Drawbacks to traditional KPI analysis, tracking & monitoring We’ve already mentioned the problems of tying up limited data science and analyst personnel as well as the fact that KPI analysis via dashboards is far too time-intensive to alert you to business incidents in real-time, but there are other drawbacks to this approach as well. Since KPIs tend to be aggregated metrics (like averages or totals), they can hide isolated yet significant problems or opportunities, as we've seen in temporary ecommerce glitches, such as listing a top-brand headset as free, rather than at its normal price, even for just a number of hours. Also, this KPI analysis workflow can very easily miss important signals in your data if you choose the wrong KPIs for monitoring to begin with. Lastly, the low number and aggregate nature of KPIs results in data that isn’t very granular. Granularity, as discussed in our previous post, is required to actually understand and fix the specific problems discovered on the dashboard. Monitoring your KPIs with Anodot: a comprehensive tracking solution for all your metrics As we mentioned above, Anodot provides a “full-stack” real time AI analytics solution. Not only can it accurately detect anomalies in all your metrics in real-time - eliminating the need to select a handful of KPIs - but Anodot’s correlation of related anomalies gives your analysts the benefits of granularity (specific actionable insights) with the conciseness you’d expect from a traditional dashboard tool. By giving you built-in data science, the majority of manual KPI analysis can be replaced by intelligently correlated and collated anomalies which point directly at business incidents. Anodot is scalable to millions of metrics, far beyond the manual, iterative approach that relies on traditional BI dashboard tools. That scalability enables Anodot to track all your metrics, thus giving you the granularity you need to find the important signals in all of your data, not just the specific KPIs you think you should be looking at. The more pixels, the sharper the image, and the key advantage of AI analytics is the ability to learn both what’s important and what’s related, not only showing you what’s there in the image, but also what it means.
Documents 1 min read

Case Study: Payoneer Discovers a ‘Game Changer’ in Pairing Coralogix with Anodot

The Production Services team at leading fintech Payoneer uses Anodot Autonomous Detection integrated with the Coralogix Log Analysis Platform to accurately identify system issues with real-time alerts and to accelerate remediation time.
Blog Post 5 min read

Expectations are High for Advanced Anomaly Detection

Recently Gartner released their latest Hype Cycle report for Data Science and Machine Learning, advising Data and analytics leaders to use this report to better understand the data science and machine learning landscape and evaluate their current capabilities and technology adoption prospects. The report states that "to succeed with data and analytics initiatives, enterprises must develop a holistic view of critical technology capabilities." One of the technologies noted that is 'On the Rise' side of the curve is Advanced Anomaly Detection. Advanced Anomaly Detection is Accelerating Expecting Advanced Anomaly Detection to accelerate to mainstream over the next ten years, the Gartner report explained that "Anomaly detection is relevant in situations where the existence, nature and extent of disruptions cannot be predicted. Systems that use advanced anomaly detection are more effective than those that use simpler techniques. They can detect subtle anomalies that might otherwise escape notice, provide earlier warning of impending problems or more time to capitalize on emerging opportunities, reduce the human effort required to develop a monitoring or measuring application, and reduce the time to solution for implementing complicated anomaly detection systems." In the report, Gartner senior analyst Peter Krensky and Gartner VP and distinguished analyst W. Roy Schulte provide their insight for how Advanced Anomaly Detection will impact companies, and the benefits to be gained. They say that “virtually every company has some aspects of operations for which it is important to distinguish between routine conditions and matters that require extra attention. Much of advanced anomaly detection is a competitive advantage today, but we expect most of the current technology driving it will be widespread and even taken for granted within 10 years because of its broad applicability and benefits. Advanced anomaly detection is already being embedded in many enterprise applications.” Forces Driving the Need for Advanced Anomaly Detection We find the main drivers that are creating the need for Advanced Anomaly Detection solutions come as businesses are having to manage increasing amounts of data and tracking more metrics. An unavoidable consequence of the scale and speed of today’s business, especially when lines of code sets are changed in seconds, are costly glitches. Moreover, as many processes happen simultaneously, many businesses monitor activities by a different person or team. Anomalies in one area often affect performance in other areas, but it is difficult for the association to be made when departments operate independently of one another. With businesses constantly generating new data, they need a solution that can analyze metrics on a holistic level and give people the insights they need to know what is going on. Applying Advanced Anomaly Detection Looking at how Advanced Anomaly Detection has already been applied, the analysts note that "its use is increasing in applications as diverse as enterprise security, unified monitoring and log analysis, application and network performance monitoring, business activity monitoring (including business process monitoring), Internet of Things predictive equipment maintenance, supply chain management, and corporate performance." High Business Impact of Advanced Anomaly Detection Advanced Anomaly Detection received a maturity level of ‘Emerging’, meaning that it is expected to grow and evolve. The strong momentum that Advanced Anomaly Detection has gained has drawn the attention of innovative vendors. The report lists some sample vendors who offer Advanced Anomaly Detection solutions, such as Anodot. Our Take on the Gartner Hype Cycle on Data Science and Machine Learning Each year, Gartner produces Hype Cycle reports to provide a graphic representation of the maturity and adoption of technologies and applications, and how they are potentially relevant to solving real business problems and exploit new opportunities. In our case, the challenge with big data is how to find the right insight, or really what are the right business incidents in the data. While Gartner outlines a clear need and drive for Advanced Anomaly Detection for data driven businesses, we believe that mainstream adoption will appear faster than projected in the report, because businesses that implement it will enjoy clear competitive advantage, driving competitors in their segment to move quickly to add Advanced Anomaly Detection capabilities. There are many systems that try to find events and data that is not normal. Yet, often these systems fail, identifying too many anomalies (false positives) or not enough (false negatives). Applying machine learning to many rapidly changing environments means constantly updating and retraining models, to prevent false positives. Our AI analytics solution offers autonomous analytics. Instead of having to ask the many questions and carry out complex data analysis, Anodot does the work, and provides the answers to understand why an incident happened. Next Step Click here to download our white paper that helps you explore and weigh the issues around the Build or Buy Dilemma for Advanced Anomaly Detection. All statements in this report attributable to Gartner represent Anodot’s interpretation of data, research opinion or viewpoints published as part of a syndicated subscription service by Gartner, Inc., and have not been reviewed by Gartner. Each Gartner publication speaks as of its original publication date (and not as of the date of this report). The opinions expressed in Gartner publications are not representations of fact, and are subject to change without notice. Gartner does not endorse any vendor, product or service depicted in its research publications, and does not advise technology users to select only those vendors with the highest ratings or other designation. Gartner research publications consist of the opinions of Gartner's research organization and should not be construed as statements of fact. Gartner disclaims all warranties, expressed or implied, with respect to this research, including any warranties of merchantability or fitness for a particular purpose. Gartner, Hype Cycle for Data Science and Machine Learning, 2017, 28 July 2017
Documents 1 min read

Mission-critical monitoring that scales with rapid growth

"Anodot is a game-changer. It connects the dots all together quickly, in a single email. We've reduced the number of false positives to almost zero."
Blog Post 5 min read

We’re Only Human, After All: How Can Effective AI Metrics Monitoring Uncover the Opportunities You’re Missing

In this first post of our new series on overlooked incidents and AI analytics, we’ll discuss how intelligent monitoring of your metrics can open the door when opportunity knocks…before it walks away. “The main thing is to keep the main thing the main thing” - Stephen Covey Or at least that’s the intent when starting off. While it’s technically true that revenue and cost are only two key performance indicators (KPIs), it’s also true that both of those metrics are in fact aggregates - or more accurately, summaries. At such a high level of aggregation, there’s zero actionable insights because your team can’t directly alter either of those two metrics, but rather the myriad of little decisions which can tune the performance of the complex engine which is your business. Without access to much more granular data, you’re forced to make specific actions with incredibly vague information. The solution, of course, is to monitor a much higher number of much more specific metrics. This approach, however, can turn into an embarrassment of riches if your monitoring solution can’t keep up all the signals that will now need monitoring. Metrics monitoring can easily become an overwhelming task There’s an example we like to use which really drives home this problem of scale: Let’s say you’re an analyst at an ecommerce company and at the moment you’re monitoring only two KPIs - the number of purchases, and revenue. You’re a smaller e-tailer, with only 50 products total spread out over 10 categories, and you’re focused exclusively on the United States and want state-level data. As a company that does business on the web, you’ve already been bitten by platform and version-specific problems, and thus you want statistics from eight operating systems (four major Windows versions, 2 desktop Mac OSs, and one version each for both iOS and Android) to help you identify and fix future problems much faster. So how many total metrics will you be monitoring? The answer is: 2 X 50 X 10 X 50 X 8 = 400,000 As you get more granular, the number of permutations increases rapidly; in the off chance that Puerto Rico becomes the 51st state - that number will increase by 8,000. 400,000 metrics is far too many for manual metrics monitoring via traditional BI dashboards, alerts and teams of data scientists and analysts to be practical. Fortunately, AI-powered automated machine learning solutions are able to take human eyeballs off the dashboards because these real-time analytics solutions are able to accurately and automatically detect the anomalies in all that time series data. Those anomalies are the real signals in your data upon which you need to act. Accurate and real-time anomaly detection, coupled with the ability to correlate related anomalies across multiple data sources, is ushering in a new time in business intelligence when no data-driven organization should be surprised by an unexpected business incident ever again. The anomalies in the data which point to the opportunities in the market can now be found. Think of it this way: if you monitor everything, you can detect anything, especially events you didn’t know you had to look for. This is the real power of AI analytics. Absolutely no metric is overlooked Perhaps for a better perspective, let’s discuss a real-life example of how machine learning-based anomaly detection can help a business gain actionable business insights. When a celebrity endorses a product on Instagram, the free positive buzz can really drive up sales, but only if the reaction is in time. A large apparel conglomerate learned that the hard way when their BI team discovered the endorsement…two days later. If they discovered the sharp uptick in sales for that product and the rapidly dwindling inventory of that product in one of their regional warehouses in real time, they could have capitalized on the opportunity by increasing the price or replenishing the inventory to keep the customer demand fed. Now, that same Fortune 500 heavyweight is an Anodot customer and hot opportunities like that don’t slip by anymore. According to the Data Analytics Director, “With Anodot, we get real time alerts for sales spikes…or when an impending snow storm causes a decline in in-store purchases in the Midwest.” Switching to much more effective metrics monitoring has obviously added to their bottom line. When you have a data scientist in the cloud, any user can easily and automatically gain actionable business insights. Whether it’s a celebrity endorsement or signs of premature equipment failure from a swarm of IoT sensor devices, AI analytics tools are data agnostic and can find the signal hidden in your time series data. With these tools, you can extract not only actionable insights, but ultimately increased revenue from the business incidents discovered in your data. And isn’t that supposed to be the main thing?   Read our next post about "KPI Analysis with AI Anomaly Detection"
Documents 1 min read

Case Study: Vimeo Uses Anodot to Tap Into User Experience and Optimize Internal Operations

Data is a treasure trove for Vimeo, and certainly something it sees as a differentiator and competitive advantage. Anodot helps the company find those nuggets of insight that would otherwise be overlooked.
Blog Post 3 min read

Anodot’s AI Analytics Deliver Valuable Business Insights to Media Giant PMC

Facing significant delays in discovering important incidents in their active, online business, PMC’s data science team needed a better way to stay on top of business incidents. PMC had been relying on Google Analytics’ alert function. However, the problem was that they had to know what they’re looking for in order to set the alerts in Google Analytics. This was really time consuming and some things tended to get missed, especially with millions of users across dozens of professional publications. A Bit About Penske Media Corporation (PMC) PMC is a leading digital media and information services company. Its owned and operated brands reach more than 179 million visitors monthly and Penske Media was recently named one of the Top 100 Private Companies in the United States and North America.  PMC is one of the largest digital media companies in the world, publishing more than 20 digital media brands including PMC Studios, WWD, Footwear News, M, Deadline.com, Variety magazine, Beauty Inc, Movieline, and more. PMC additionally owns and produces more than 80 annual events, summits, award shows and conferences while housing a dynamic research and emerging data business. Finding Issues Even Without Knowing What to Look For And then came Anodot. Anodot’s AI Analytics started by tracking PMC’s Google Analytics activity, identifying anomalous behavior in impressions and click-through-rates for advertising units. Analyzing the Google Analytics data, Anodot identified a new trend where a portion of the traffic to one of PMC’s media properties came from a “bad actor” - referral spam that was artificially inflating visitor statistics. For PMC’s analytics team, spotting this issue would have meant the impossible - that they already knew what they were looking for in advance.  After discovering this by using Anodot, PMC was able to block the spam traffic, and free up critical resources for legitimate visitors. PMC could then accurately track the traffic that mattered the most, enabling PMC executives to make more informed decisions.   If a story in one of our publications goes wild, we see spikes in our data which Anodot catches. Our editorial team can use the information about successful content themes derived from Anodot’s analytics. Andrew Maguire, PMC’s Head Data Scientist Moving Forward: PMC uses Anodot for Intelligent Alerting PMC plans to apply Anodot to provide an intelligent alerting system driven by machine learning with hardly any direction required from business users in terms of complex rules or triggers. PMC will incorporate Anodot into more core sources of data across the business and implement even more nuanced custom tracking on top of Google Analytics so that they can track key metrics that matter. Click here to download the full case study and find out how Anodot’s AI Analytics integrated with Google Analytics is helping PMC to prevent revenue loss, remedying urgent problems quickly, and capturing opportunities fast.
Blog Post 5 min read

Small Glitches, Big Trouble: How Checking for Potential Outliers in Time Series Data is a Necessity in eCommerce

Just before we get into how Anodot extracts actionable insights from time series data, it’s worthwhile to recap what exactly a time series is and how businesses typically generate them. First, companies take a certain metric (a value or quantity) which is considered important, most often it’s one of the usual key performance indicators (KPIs): revenue, profit, or cost. Then, the company decides how often they’re going to sample (update) that number, and then they pull in data samples at that interval. Lastly, those two-item data points then go into some designated data bucket, such as a database. Analytics tools like dashboards then retrieve the data as a set, generate a plot and update it as each data point comes in. Depending on the type of data, the amount of noise and sampling rate - the actual data distribution (and thus appearance of the plotted time series) can vary widely. It’s important to note that the data bucket is for the benefit of later in-depth analysis, not for Anodot’s machine-learning powered outlier detection. This is because Anodot uses “online” algorithms - computational processes which learn and update its models with each incoming data point, without having to go back to all the data points in that time series before it. All the previous data points are “encoded” into the models already. Examples of time series data Time series data sets are very useful when monitoring a given value over time. Since this is a near-universal need in many industries, it’s no surprise that all of today’s fastest growing (and data-intensive) industries use time series data sets. In ad tech, for example, there are metrics such as cost per lead, impression share, cost per click, bounce rate, page views and click-through rate. Common metrics in ecommerce include conversion rate, revenue per click, number of transactions, and average order value. However, the actual time series data which are measured are far more specific because each of those above examples are often broken down by geographic region (e.g. North America or Asia), or operating system - this is especially true for mobile app metrics, since revenue-sapping compatibility problems are often OS specific. This level of granularity allows companies to spot very specific anomalies, especially those which would get smoothed out and thus left unnoticed by more encompassing metrics, especially averages and company-wide totals. How Anodot checks for different types of potential outliers In the first installment of this series, we discussed the three different categories of outliers: global (or point), contextual (also called conditional) and collective outliers. All three of these can occur in time series data and all three can be detected. Take for example, a large spike in transaction volume at an ecommerce company which reaches a value never before seen in the data, thus making it a textbook example of a global outlier. This can be a great thing, since more sales usually means more revenue. Well, usually. A large spike in sales volume can also indicate that you have a stampede of online shoppers taking advantage of a pricing glitch. In such a case, your average revenue per transaction might actually be dipping down a little, depending on the ratio of glitch sales to normal sales. This slight dip might actually be completely normal at other times of year (like the normal retail slow periods which occur outside of the holiday season or back to school shopping), but not when you’re running a promotion. In this case, the low values of average revenue per transaction would be considered a contextual outlier. Hmmm, perhaps the promotional sale price for those TVs was entered as $369 instead of $639. Anodot is able to detect both types of outliers thanks to its ability to account for any and all seasonal patterns in a time series metric, thus catching the contextual outliers, and accurately determine whether a data point falls far outside the natural variance of that time series, thus catching global outliers. Anodot’s first layer - univariate outlier detection - is all about identifying global and contextual outliers in individual metrics. A second layer then focuses on what are called collective outliers (a subset of data which, as a collection, deviates from rest of the data it’s found in). This second layer uses multivariate outlier detection to group related anomalies together. This two-layer approach provides analysts both granularity and concise alerts at the same time. The advantage of automated outlier detection for finding potential outliers Human beings are quite good at spotting outliers visually on a time series plot, because we benefit from possessing the most sophisticated neural network we know of. Our “wetware”, however, can’t do this instantly at the scale of millions of metrics in real time. Under those constraints, not only are recall and precision important, but also detection time. In the case of our hypothetical price glitch above, each second that glitch persists unfixed means thousands of dollars lost. Automated anomaly detection simplifies the detection and correlation of outliers in your data, giving you real-time insights and real-world savings. Considering building a machine learning anomaly detection system? Download this white paper.
Blog Post 5 min read

Unexpected plot twist: understanding what are data outliers and how their detection can eliminate business latency

Business metric data is only as useful as the insights that can be extracted from them, and that extraction is ultimately limited by the tools employed. One of the most basic and commonly used data analysis and visualization tools is the time series: a two-dimensional plot of some metric’s value at many sequential moments in time. Each data point in that time series is a record of one facet of your business at that particular instant. When plotted over time, that data can reveal trends and patterns that indicate the current state of that particular metric. The basics: understanding what data outliers are A time series plot shows what is happening in your business. Sometimes, that can diverge from what you expect should happen. When that divergence is outside the usual bounds of variance, it’s an outlier. In Anodot’s outlier detection system, the expectations which set those bounds are derived from a continuous examination of all the data points for that metric. In many situations, data outliers are errant data which can skew averages, and thus are usually filtered out and excluded by statisticians and data analysts before they attempt to extract insights from the data. The rationale is that those outliers are due to reporting errors or some other cause they needn’t worry about. Also, since genuine outliers are relatively rare, they aren’t seen as indicating a deeper, urgent problem with the system being monitored. Outliers, however, can be significant data points in and of themselves when reporting or when other sources of error aren’t suspected. An outlier in one of your metrics could reflect a one-off event or a new opportunity, like an unexpected increase in sales for a key demographic you’ve been trying to break into. Outliers in time series data mean something has changed. Significant changes can first manifest as outliers when only a few events serve as an early harbinger of a much more widespread issue. A large ecommerce company, for example, may see a larger than usual number of payment processing failures from a specific but rarely used financial institution. The failures were due to the fact that they updated their API to incorporate new regulatory standards for online financial transactions. This particular bank was merely the first to be compliant with the new industry-wide standard. If these failures get written off as inconsequential outliers and not recognized as the canary in a coal mine, the entire company may soon not be able to accept any payments, as every bank eventually adopts the new standard. Outliers to the rescue At Anodot, we learned firsthand that an entire metric for a particular object may be an outlier, compared to that identical metric from other similar objects. This is a prime example of how outlier detection can be a powerful tool for optimization: by spotting a single underperforming component, the performance of the whole system can be dramatically improved. For us, it was a degradation of the performance from a single Cassandra node. For your business, it could be a CDN introducing unusually high latency, causing web page load times to rise, becoming unbearable for your visitors as they click away and fall into someone else’s funnel. Anodot’s outlier detection compares aspects that are supposed to behave similarly and identifies the ones that are behaving differently: a single data point which is unexpectedly different from the previous ones, or a metric from a particular aspect which deviates from that same metric from other identical aspects. Context requires intelligent data monitoring Anomalous data points are classified in the context of all the data points which came before. The significance of detected anomalies is then quantified in the context of their magnitude and persistence. Finally, concise reporting of those detected significant anomalies is informed by the context of other anomalies in related metrics. Context requires understanding… understanding gleaned from learning. Machine learning, that is. Even though there are several tests for outliers which don’t involve machine learning, they almost always assume a standard Gaussian distribution (the iconic bell curve), which real data often doesn’t exhibit. But there’s another kind of latency which outlier detection can remove: business latency. One example of business latency is the lag between a problem’s occurrence and its discovery. Another is the time delay between the discovery of the problem and when an organization possesses the actionable insights to quickly fix it. Anodot’s outlier detection system can remove both: the former by accurate real-time anomaly detection, the latter by concise reporting of related anomalies. Solving the problem of business latency is a priority for all companies in the era of big data, and it’s a much harder problem to solve with traditional business intelligence (BI) tools. Traditional BI: high latency, less results Traditional BI is not designed for real time big data, but rather for analyzing historical data. In addition, they simply visualize the given data, rather than surface issues that need to be considered. Therefore, analysts cannot rely on BI solutions to find what they are looking for, as they first must understand what they need to find. Using traditional BI, analysts may identify issues late, if at all, which leads to loss of revenue, quality, and efficiency. Speed is what’s needed – that one essential component for successful BI alerts and investigations. And that speed can make your business an outlier - way above your competition.