2020
DOI: 10.1142/s0218194020500114
|View full text |Cite
|
Sign up to set email alerts
|

Log-Based Anomaly Detection with the Improved K-Nearest Neighbor

Abstract: Logs play an important role in the maintenance of large-scale systems. The number of logs which indicate normal (normal logs) differs greatly from the number of logs that indicate anomalies (abnormal logs), and the two types of logs have certain differences. To automatically obtain faults by K-Nearest Neighbor (KNN) algorithm, an outlier detection method with high accuracy, is an effective way to detect anomalies from logs. However, logs have the characteristics of large scale and very uneven samples, which wi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(6 citation statements)
references
References 18 publications
0
5
0
1
Order By: Relevance
“…The results with Auto-LSTM, IKNN and nLSALog for the BGL data set are given in Table 2 a. The precision, recall and F-measure results for negative logs are better than the 92%, 91% and 92%, respectively, with the improved K-nearest neighbors (IKNN) supervised algorithm [ 36 ]. The precision, recall and F-measure results for negative logs are also better than the 82.5%, 94.7% and 88.2%, respectively, with the nLSALog algorithm [ 40 ].…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The results with Auto-LSTM, IKNN and nLSALog for the BGL data set are given in Table 2 a. The precision, recall and F-measure results for negative logs are better than the 92%, 91% and 92%, respectively, with the improved K-nearest neighbors (IKNN) supervised algorithm [ 36 ]. The precision, recall and F-measure results for negative logs are also better than the 82.5%, 94.7% and 88.2%, respectively, with the nLSALog algorithm [ 40 ].…”
Section: Resultsmentioning
confidence: 99%
“…A decision tree model was considered in [32] to detect faults using log messages. An improved supervised K-nearest neighbors (IKNN) method was employed in [36] to detect anomalies in log messages. However, using supervised methods is not always possible because of the lack of labeled data.…”
Section: Introductionmentioning
confidence: 99%
“…The F-measure is 99.6% for both negative and positive logs. The precision, recall, and F-measure with oversampling for negative logs are better than the 96%, 96%, and 96%, respectively, with the Improved K-Nearest Neighbor algorithm [23].…”
Section: Thunderbirdmentioning
confidence: 87%
“…Liberty veri setinin K-en yakın komşular makine öğrenimi algoritması ile log tabanlı anomali tespiti için 2020 yılına ait bir çalışma kullanıldığı görülmektedir [15]. 2021 yılında gerçekleştirilen başka bir araştırmada log verisindeki anomalilerin taksonomisi konusu araştırılmış ve Thunderbird, Spirit ve BGL veri setlerinin kullanıldığı görülmüştür [16]. Blue Gene/L, Thunderbird, Redstorm, Liberty ve Spirit veri setlerinin oluşturulmaya başlanma tarihleri oldukça eskidir [17].…”
Section: İlgili çAlışmalarunclassified