TechRxiv
Construct_validity.Sjoberg.TSE.submitted.pdf (2.15 MB)

Construct Validity in Software Engineering

Download (2.15 MB)
preprint
posted on 06.03.2021, 13:27 by Dag SjøbergDag Sjøberg, Gunnar BergersenGunnar Bergersen
Empirical research aims to establish generalizable claims from data. Such claims involve concepts that often must be measured indirectly by using indicators. Construct validity is concerned with whether one can justifiably make claims at the conceptual level that are supported by results at the operational level. We report a quantitative analysis of the awareness of construct validity in the software engineering literature between 2000 and 2019 and a qualitative review of 83 articles about human-centric experiments published in five high-quality journals between 2015 and 2019. Over the two decades, the appearance in the literature of the term construct validity increased sevenfold. Some of the reviewed articles we reviewed employed various ways to ensure that the indicators span the concept in an unbiased manner. We also found articles that reuse formerly validated constructs. However, the articles disagree about how to define construct validity. Several interpret construct validity excessively by including threats to internal, external, or statistical conclusion validity. A few articles also include fundamental challenges of a study, such as cheating and misunderstandings of experiment material. The diversity of topics discussed makes us recommend a minimalist approach to construct validity. We propose seven guidelines to establish a common ground for addressing construct validity in software engineering.

History

Email Address of Submitting Author

dagsj@ifi.uio.no

ORCID of Submitting Author

0000-0002-4941-7240

Submitting Author's Institution

University of Oslo

Submitting Author's Country

Norway

Usage metrics

Licence

Exports