Reference : Evaluating Search-Based Techniques With Statistical Tests |
Scientific congresses, symposiums and conference proceedings : Paper published in a book | |||
Engineering, computing & technology : Computer science | |||
http://hdl.handle.net/10993/35262 | |||
Evaluating Search-Based Techniques With Statistical Tests | |
English | |
Arcuri, Andrea ![]() | |
2018 | |
The Search-Based Software Testing (SBST) Workshop | |
No | |
The Search-Based Software Testing (SBST) Workshop | |
May | |
[en] This tutorial covers the basics of how to use statistical tests to
evaluate and compare search-algorithms, in particular when applied on software engineering problems. Search-algorithms like Hill Climbing and Genetic Algorithms are randomised. Running such randomised algorithms twice on the same problem can give different results. It is hence important to run such algorithms multiple times to collect average results, and avoid so publishing wrong conclusions that were based on just luck. However, there is the question of how often such runs should be repeated. Given a set of n repeated experiments, is such n large enough to draw sound conclusions? Or should had more experiments been run? Statistical tests like the Wilcoxon-Mann-Whitney U-test can be used to answer these important questions. | |
http://hdl.handle.net/10993/35262 | |
FnR ; FNR3949772 > Lionel Briand > VVLAB > Validation and Verification Laboratory > 01/01/2012 > 31/07/2018 > 2010 |
File(s) associated to this reference | ||||||||||||||
Fulltext file(s):
| ||||||||||||||
All documents in ORBilu are protected by a user license.