No full text
Paper published in a book (Scientific congresses, symposiums and conference proceedings)
Efficient Log-based Anomaly Detection with Knowledge Distillation
Nguyen, Huy-Trung; Nguyen, Lam-Vien; LE, Van Hoang et al.
2024In Chang, Rong N. (Ed.) Proceedings - 2024 IEEE International Conference on Web Services, ICWS 2024
Peer reviewed
 

Files


Full Text
No document available.

Send to



Details



Keywords :
Deep learning; Knowledge distillation; Log-based anomaly detection; Anomaly detection; Computational resources; Detection accuracy; Large models; Learning models; Regular operations; Resourceconstrained devices; Information Systems; Computer Science Applications; Computer Networks and Communications; Information Systems and Management; Artificial Intelligence
Abstract :
[en] Logs are produced by many systems for troubleshooting purposes. Detecting abnormal events is crucial to maintaining regular operations and securing the security of systems. Despite the achievements of deep learning models on anomaly detection, it remains challenging to apply these deep learning models in some scenarios; one popular case is deploying on resource-constrained scenarios such as IoT devices due to the limitation of computational resources on these devices. We identify two main problems of adopting these deep learning models in practice, including (1) they cannot deploy on resource-constrained devices because of the size of large models and the time needed to analyze data with the models, and (2) they cannot achieve satisfactory detection accuracy with simple models. In this work, we proposed a novel lightweight anomaly detection method from system logs, DistilLog, to overcome these problems. DistilLog utilizes a pretrained word2vec model to represent log event templates as semantic vectors, incorporated with the PCA dimensionality reduction algorithm to minimize computational and storage burden. The Knowledge Distillation technique is applied to reduce the size of the detection model while maintaining high detection accuracy. The experimental results show that DistilLog can achieve high F-measures of 0.964 and 0.961 on HDFS and BGL datasets while maintaining the minimized model size and fastest detection speed. This effectiveness and efficiency demonstrate the potential for widespread use in most scenarios by showing the ability to deploy the proposed model on resource-constrained systems.
Disciplines :
Computer science
Author, co-author :
Nguyen, Huy-Trung;  People's Security Academy, Viet Nam
Nguyen, Lam-Vien;  People's Security Academy, Viet Nam
LE, Van Hoang  ;  University of Newcastle, Australia
Zhang, Hongyu;  Chongqing University, China
Le, Manh-Trung;  Viet Nam
External co-authors :
yes
Language :
English
Title :
Efficient Log-based Anomaly Detection with Knowledge Distillation
Publication date :
2024
Event name :
2024 IEEE International Conference on Web Services (ICWS)
Event place :
Hybrid, Shenzhen, Chn
Event date :
07-07-2024 => 13-07-2024
Main work title :
Proceedings - 2024 IEEE International Conference on Web Services, ICWS 2024
Editor :
Chang, Rong N.
Publisher :
Institute of Electrical and Electronics Engineers Inc.
ISBN/EAN :
9798350368550
Peer reviewed :
Peer reviewed
Funders :
Australian Research Council
Funding text :
Van-Hoang Le and Hongyu Zhang are supported by Australian Research Council (ARC) Discovery Projects (DP200102940, DP220103044). We also thank anonymous reviewers for their insightful and constructive comments, which significantly improve this paper.
Available on ORBilu :
since 26 January 2026

Statistics


Number of views
3 (0 by Unilu)
Number of downloads
0 (0 by Unilu)

Scopus citations®
 
2
Scopus citations®
without self-citations
2
OpenCitations
 
0
OpenAlex citations
 
1

Bibliography


Similar publications



Contact ORBilu