Paper published in a book (Scientific congresses, symposiums and conference proceedings)
FlakyCat: Predicting Flaky Tests Categories using Few-Shot Learning
AKLI, Amal; HABEN, Guillaume; Habchi, Sarra et al.
2023In FlakyCat: Predicting Flaky Tests Categories using Few-Shot Learning
Peer reviewed
 

Files


Full Text
FlakyCat.pdf
Author preprint (488.31 kB)
Download

All documents in ORBilu are protected by a user license.

Send to



Details



Keywords :
Software Testing; Flaky Tests; CodeBERT; Few-Shot learning; Siamese Networks
Abstract :
[en] Flaky tests are tests that yield different outcomes when run on the same version of a program. This non-deterministic behaviour plagues continuous integration with false signals, wasting developers’ time and reducing their trust in test suites. Studies highlighted the importance of keeping tests flakiness-free. Recently, the research community has been pushing towards the detection of flaky tests by suggesting many static and dynamic approaches. While promising, those approaches mainly focus on classifying tests as flaky or not and, even when high performances are reported, it remains challenging to understand the cause of flakiness. This part is crucial for researchers and developers that aim to fix it. To help with the comprehension of a given flaky test, we propose FlakyCat, the first approach to classify flaky tests based on their root cause category. FlakyCat relies on CodeBERT for code representation and leverages Siamese networks to train a multi-class classifier. We train and evaluate FlakyCat on a set of 451 flaky tests collected from open-source Java projects. Our evaluation shows that FlakyCat categorises flaky tests accurately, with an F1 score of 73%. Furthermore, we investigate the performance of our approach for each category, revealing that Async waits, Unordered collections and Time-related flaky tests are accurately classified, while Concurrency-related flaky tests are more challenging to predict. Finally, to facilitate the comprehension of FlakyCat’s predictions, we present a new technique for CodeBERT-based model interpretability that highlights code statements influencing the categorization.
Research center :
Interdisciplinary Centre for Security, Reliability and Trust (SnT) > SerVal - Security, Reasoning & Validation
Disciplines :
Computer science
Author, co-author :
AKLI, Amal ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SerVal
HABEN, Guillaume  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SerVal
Habchi, Sarra
PAPADAKIS, Mike ;  University of Luxembourg > Faculty of Science, Technology and Medicine (FSTM) > Department of Computer Science (DCS)
LE TRAON, Yves ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SerVal
External co-authors :
yes
Language :
English
Title :
FlakyCat: Predicting Flaky Tests Categories using Few-Shot Learning
Publication date :
May 2023
Event name :
4th International Conference on Automation of Software Test
Event date :
from 15-05-2023 to 16-05-2023
Main work title :
FlakyCat: Predicting Flaky Tests Categories using Few-Shot Learning
Peer reviewed :
Peer reviewed
Focus Area :
Security, Reliability and Trust
Available on ORBilu :
since 26 August 2023

Statistics


Number of views
68 (6 by Unilu)
Number of downloads
108 (5 by Unilu)

Scopus citations®
 
3
Scopus citations®
without self-citations
3

Bibliography


Similar publications



Contact ORBilu