Abstract
CHI researchers typically use a significance testing approach to statistical analysis when testing hypotheses during usability evaluations. However, the appropriateness of this approach is under increasing criticism, with statisticians, economists, and psychologists arguing against the use of routine interpretation of results using "canned" p values. Three problems with current practice - the fallacy of the transposed conditional, a neglect of power, and the reluctance to interpret the size of effects - can lead us to build weak theories based on vaguely specified hypothesis, resulting in empirical studies which produce results that are of limited practical or scientific use. Using publicly available data presented at CHI 2010 [19] as an example we address each of the three concerns and promote consideration of the magnitude and actual importance of effects, as opposed to statistical significance, as the new criteria for evaluating CHI research.
Original language | English |
---|---|
Title of host publication | CHI '12 Proceedings of the 2012 ACM annual conference on Human Factors in Computing Systems |
Place of Publication | Austin, Texas |
Publisher | Association for Computing Machinery |
Pages | 1105-1114 |
Number of pages | 10 |
ISBN (Print) | 978-1-4503-1015-4 |
DOIs | |
Publication status | Published - 2012 |
Keywords
- Usability evaluation
- Bayesian Statistics
- Research Methods