Troubleshooting machine learning anomaly detection and frequently asked questions
editTroubleshooting machine learning anomaly detection and frequently asked questions
editUse the information in this section to troubleshoot common problems and find answers for frequently asked questions.
How to restart failed anomaly detection jobs
editIf an anomaly detection job fails, try to restart the job by following the procedure described below. If the restarted job runs as expected, then the problem that caused the job to fail was transient and no further investigation is needed. If the job quickly fails after the restart, then the problem is persistent and needs further investigation. In this case, find out which node the failed job was running on by checking the job stats on the Job management pane in Kibana. Then get the logs for that node and look for exceptions and errors where the ID of the anomaly detection job is in the message to have a better understanding of the issue.
If an anomaly detection job has failed, do the following to recover from failed
state:
-
Force stop the corresponding datafeed by using the Stop datafeed API with the
force
parameter beingtrue
. For example, the following request force stops themy_datafeed
datafeed.POST _ml/datafeeds/my_datafeed/_stop { "force": "true" }
-
Force close the anomaly detection job by using the Close anomaly detection job API with the
force
parameter beingtrue
. For example, the following request force closes themy_job
anomaly detection job:POST _ml/anomaly_detectors/my_job/_close?force=true
- Restart the anomaly detection job on the Job management pane in Kibana.
What machine learning methods are used for anomaly detection?
editFor detailed information, refer to the paper Anomaly Detection in Application Performance Monitoring Data by Thomas Veasey and Stephen Dodson, as well as our webinars on The Math behind Elastic Machine Learning and Machine Learning and Statistical Methods for Time Series Analysis.
Further papers cited in the C++ code:
What are the input features used by the model?
editAll input features are specified by the user, for example, using diverse statistical functions like count or mean over the data of interest.
Does the data used by the model only include customers' data?
editYes. Only the data specified in the anomaly detection job configuration are used for detection.
What does the model output score represent? How is it generated and calibrated?
editThe ensemble model generates a probability value, which is then mapped to an anomaly severity score between 0 and 100. The lower the probability of observed data, the higher the severity score. Refer to this advanced concept doc for details. Calibration (also called as normalization) happens on two levels:
-
Within the same metric/partition, the scores are re-normalized “back in time”
within the window specified by the
renormalization_window_days
parameter. This is the reason, for example, that bothrecord_score
andinitial_record_score
exist. - Over multiple partitions, scores are renormalized as described in this blog post.
Is the model static or updated periodically?
editIt’s an online model and updated continuously. Old parts of the model are pruned
out based on the parameter model_prune_window
(usually 30 days).
Is the performance of the model monitored?
editThere is a set of benchmarks to monitor the performance of the anomaly detection algorithms and to ensure no regression occurs as the methods are continuously developed and refined. They are called "data scenarios" and consist of 3 things:
- a dataset (stored as an Elasticsearch snapshot),
- a machine learning config (anomaly detection, {dfanalysis}, transform, or inference),
- an arbitrary set of static assertions (bucket counts, anomaly scores, accuracy value, and so on).
Performance metrics are collected from each and every scenario run and they are persisted in an Elastic Cloud cluster. This information is then used to track the performance over time, across the different builds, mainly to detect any regressions in the performance (both result quality and compute time).
On the customer side, the situation is different. There is no conventional way to monitor the model performance as it’s unsupervised. Usually, operationalization of the model output include one or several of the following steps:
- Creating alerts for influencers, buckets, or records based on a certain anomaly score.
- Use the forecasting feature to predict the development of the metric of interest in the future.
- Use one or a combination of multiple anomaly detection jobs to identify the significant anomaly influencers.
How to measure the accuracy of the unsupervised machine learning model?
editFor each record in a given time series, anomaly detection models provide an anomaly severity score, 95% confidence intervals, and an actual value. This data is stored in an index and can be retrieved using the Get Records API. With this information, you can use standard measures to assess prediction accuracy, interval calibration, and so on. Elasticsearch aggregations can be used to compute these statistics.
The purpose of anomaly detection is to achieve the best ranking of periods where an anomaly happened. A practical way to evaluate this is to keep track of real incidents and see how well they correlate with the predictions of anomaly detection.
Can the anomaly detection model experience model drift?
editElasticsearch’s anomaly detection model continuously learns and adapts to changes in the time series. These changes can take the form of slow drifts as well as sudden jumps. Therefore, we take great care to manage the adaptation to changing data characteristics. There is always a fine trade-off between fitting anomalous periods (over-fitting) and not learning new normal behavior. The following are the main approaches Elastic uses to manage this trade-off:
- Learning the optimal decay rate based on measuring the bias in the forecast and the moments of the error distribution and error distribution moments.
- Allowing continuous small drifts in periodic patterns. This is achieved by continuously minimizing the mean prediction error over the last iteration with respect to a small bounded time shift.
- If the predictions are significantly wrong over a long period of time, the algorithm tests whether the time series has undergone a sudden change. Hypothesis Testing is used to test for different types of changes, such as scaling of values, shifting of values, and large time shifts in periodic patterns such as daylight saving time.
- Running continuous hypothesis tests on time windows of various lengths to test for significant evidence of new or changed periodic patterns, and update the model if the null hypothesis of unchanged features is rejected.
- Accumulating error statistics on calendar days and continuously test whether predictive calendar features need to be added or removed from the model.
What is the minimum amount of data for an anomaly detection job?
editElastic machine learning needs a minimum amount of data to be able to build an effective model for anomaly detection.
-
For sampled metrics such as
mean
,min
,max
, andmedian
, the minimum data amount is either eight non-empty bucket spans or two hours, whichever is greater. - For all other non-zero/null metrics and count-based quantities, it’s four non-empty bucket spans or two hours, whichever is greater.
-
For the
count
andsum
functions, empty buckets matter and therefore it is the same as sampled metrics - eight buckets or two hours. -
For the
rare
function, it’s typically around 20 bucket spans. It can be faster for population models, but it depends on the number of people that interact per bucket.
Rules of thumb:
- more than three weeks for periodic data or a few hundred buckets for non-periodic data
- at least as much data as you want to forecast
Are there any checks or processes to ensure data integrity?
editThe Elastic machine learning algorithms are programmed to work with missing and noisy data and use denoising and data reputation techniques based on the learned statistical properties.