Article (Scientific journals)
Few-Shot Anomaly Detection via Category-Agnostic Registration Learning.
Huang, Chaoqin; Guan, Haoyan; Jiang, Aofan et al.
2025In IEEE Transactions on Neural Networks and Learning Systems, 36 (7), p. 13527 - 13541
Peer Reviewed verified by ORBi
 

Files


Full Text
2406.08810v2.pdf
Author preprint (6.69 MB)
Request a copy

All documents in ORBilu are protected by a user license.

Send to



Details



Keywords :
Anomaly detection (AD); few-shot learning (FSL); registration; self-supervised learning; Anomaly detection; Anomaly detection methods; Few-shot learning; Artificial Intelligence; Feature extraction; Training; Location awareness; Robustness; Defect detection
Abstract :
[en] Most existing anomaly detection (AD) methods require a dedicated model for each category. Such a paradigm, despite its promising results, is computationally expensive and inefficient, thereby failing to meet the requirements for real-world applications. Inspired by how humans detect anomalies, by comparing a query image to known normal ones, this article proposes a novel few-shot AD (FSAD) framework. Using a training set of normal images from various categories, registration, aiming to align normal images of the same categories, is leveraged as the proxy task for self-supervised category-agnostic representation learning. At test time, an image and its corresponding support set, consisting of a few normal images from the same category, are supplied, and anomalies are identified by comparing the registered features of the test image to its corresponding support image features. Such a setup enables the model to generalize to novel test categories. It is, to our best knowledge, the first FSAD method that requires no model fine-tuning for novel categories: enabling a single model to be applied to all categories. Extensive experiments demonstrate the effectiveness of the proposed method. Particularly, it improves the current state-of-the-art (SOTA) for FSAD by 11.3% and 8.3% on the MVTec and MPDD benchmarks, respectively. The source code is available at https://github.com/Haoyan-Guan/CAReg.
Disciplines :
Computer science
Author, co-author :
Huang, Chaoqin ;  Shanghai Jiao Tong University, Cooperative Medianet Innovation Center, Shanghai, China ; National University of Singapore, Department of Electrical and Computer Engineering, Singapore ; National University of Defense Technology, College of Information Communications, Wuhan, China
Guan, Haoyan ;  King’s College London, Department of Informatics, London, United Kingdom
Jiang, Aofan;  Shanghai Jiao Tong University, Cooperative Medianet Innovation Center, Shanghai, China ; Shanghai Artificial Intelligence Laboratory, Shanghai, China
Zhang, Ya ;  Shanghai Jiao Tong University, Cooperative Medianet Innovation Center, Shanghai, China ; Shanghai Artificial Intelligence Laboratory, Shanghai, China
SPRATLING, Michael  ;  University of Luxembourg > Faculty of Humanities, Education and Social Sciences (FHSE) > Department of Behavioural and Cognitive Sciences (DBCS) > Cognitive Science and Assessment ; King’s College London, Department of Informatics, London, United Kingdom
Wang, Xinchao;  National University of Singapore, Department of Electrical and Computer Engineering, Singapore
Wang, Yanfeng ;  Shanghai Jiao Tong University, Cooperative Medianet Innovation Center, Shanghai, China ; Shanghai Artificial Intelligence Laboratory, Shanghai, China
External co-authors :
yes
Language :
English
Title :
Few-Shot Anomaly Detection via Category-Agnostic Registration Learning.
Publication date :
July 2025
Journal title :
IEEE Transactions on Neural Networks and Learning Systems
ISSN :
2162-237X
eISSN :
2162-2388
Publisher :
Institute of Electrical and Electronics Engineers Inc., United States
Volume :
36
Issue :
7
Pages :
13527 - 13541
Peer reviewed :
Peer Reviewed verified by ORBi
Funders :
National Key Research and Development Program of China
Science and Technology Commission of Shanghai Municipality
State Key Laboratory of UHD Video and Audio Production and Presentation
Ministry of Education Singapore, under its Academic Research Fund Tier 2, under Award
National Research Foundation, Singapore, under its AI Singapore Program (AISG), under Award
Funding text :
This work was supported in part by the National Key Research and Development Program of China under Grant 2022ZD0160702; in part by the STCSM under Grant 22DZ2229005; in part by the 111 Plan under Grant BP0719010; in part by the State Key Laboratory of UHD Video and Audio Production and Presentation; in part by the Ministry of Education Singapore, under its Academic Research Fund Tier 2, under Award MOE-T2EP20122-0006; and in part by the National Research Foundation, Singapore, under its AI Singapore Program (AISG), under Award AISG2-RP-2021023.Received 22 January 2024; revised 13 June 2024; accepted 10 September 2024. This work was supported in part by the National Key Research and Development Program of China under Grant 2022ZD0160702; in part by the STCSM under Grant 22DZ2229005; in part by the 111 Plan under Grant BP0719010; in part by the State Key Laboratory of UHD Video and Audio Production and Presentation; in part by the Ministry of Education Singapore, under its Academic Research Fund Tier 2, under Award MOE-T2EP20122-0006; and in part by the National Research Foundation, Singapore, under its AI Singapore Program (AISG), under Award AISG2-RP-2021023. (Chaoqin Huang and Haoyan Guan contributed equally to this work.) (Corresponding authors: Yanfeng Wang; Ya Zhang.) Chaoqin Huang was with the Cooperative Medianet Innovation Center, Shanghai Jiao Tong University, Shanghai 200240, China, and also with the Department of Electrical and Computer Engineering, National University of Singapore, Singapore 119077. He is now with the College of Information Communications, National University of Defense Technology, Wuhan 430030, China (e-mail: huangchaoqin@nudt.edu.cn).
Available on ORBilu :
since 22 August 2025

Statistics


Number of views
32 (0 by Unilu)
Number of downloads
0 (0 by Unilu)

Scopus citations®
 
5
Scopus citations®
without self-citations
5
OpenCitations
 
0
OpenAlex citations
 
8
WoS citations
 
7

Bibliography


Similar publications



Contact ORBilu