Reference : Experimental Evaluation of a Tool for Change Impact Prediction in Requirements Models...
Scientific congresses, symposiums and conference proceedings : Paper published in a book
Engineering, computing & technology : Computer science
Experimental Evaluation of a Tool for Change Impact Prediction in Requirements Models: Design, Results, and Lessons Learned
Göknil, Arda mailto [University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > >]
van Domburg, Roderick []
Kurtev, Ivan []
van den Berg, Klaas []
Wijnhoven, Fons []
The Fourth International Model-Driven Requirements Engineering (MoDRE) workshop
The Fourth International Model-Driven Requirements Engineering (MoDRE) workshop co-located with RE 2014
[en] Requirements management tools ; change impact analysis ; requirements models
[en] There are commercial tools like IBM Rational RequisitePro and DOORS that support semi-automatic change impact analysis for requirements. These tools capture the requirements relations and allow tracing the paths they form. In most of these tools, relation types do not say anything about the meaning of the relations except the direction. When a change is introduced to a requirement, the requirements engineer analyzes the impact of the change in related requirements. In case semantic information is missing to determine precisely how requirements are related to each other, the requirements engineer generally has to assume the worst case dependencies based on the available syntactic information only. We developed a tool that uses formal semantics of requirements relations to support change impact analysis and prediction in requirements models. The tool TRIC (Tool for Requirements Inferencing and Consistency checking) works on models that explicitly represent requirements and the relations among them with their formal semantics. In this paper we report on the evaluation of how TRIC improves the quality of change impact predictions. A quasi-experiment is systematically designed and executed to empirically validate the impact of TRIC. We conduct the quasi-experiment with 21 master’s degree students predicting change impact for five change scenarios in a real software requirements specification. The participants are assigned with Microsoft Excel, IBM RequisitePro or TRIC to perform change impact prediction for the change scenarios. It is hypothesized that using TRIC would positively impact the quality of change impact predictions. Two formal hypotheses are developed. As a result of the experiment, we are not able to reject the null hypotheses, and thus we are not able to show experimentally the effectiveness of our tool. In the paper we discuss reasons for the failure to reject the null hypotheses in the experiment.

File(s) associated to this reference

Fulltext file(s):

Open access
modre-preprint.pdfPublisher postprint335.26 kBView/Open

Bookmark and Share SFX Query

All documents in ORBilu are protected by a user license.