In this study, the robustness of an online tool for objectively assessing singing ability was examined by: (1) determining the internal consistency and test-retest reliability of the tool; (2) comparing the task performance of web-based participants (n = 285) with a group (n = 52) completing the tool in a controlled laboratory setting, and then determining the convergent validity between settings, and (3) comparing participants’ task performance with previous research using similar singing tasks and populations. Results indicated that the online singing tool exhibited high internal consistency (Cronbach’s alpha = .92), and moderate-to-high test-retest reliabilities (.65–.80) across an average 4.5-year-span. Task performance for web- and laboratory-based participants (n = 82) matched on age, sex, and music training were not significantly different. Moderate-to-large correlations (|r| =.31–.59) were found between self-rated singing ability and the various singing tasks, supporting convergent validity. Finally, task performance of the web-based sample was not significantly different to previously reported findings. Overall the findings support the robustness of the online tool for objectively measuring singing pitch accuracy beyond a controlled laboratory environment and its potential application in large-scale investigations of singing and music ability.
Establishing the Reliability and Validity of Web-based Singing Research
- Views Icon Views
- Share Icon Share
- Search Site
Yi Ting Tan, Isabelle Peretz, Gary E. McPherson, Sarah J. Wilson; Establishing the Reliability and Validity of Web-based Singing Research. Music Perception 1 April 2021; 38 (4): 386–405. doi: https://doi.org/10.1525/mp.2021.38.4.386
Download citation file: