Computational validity
Abstract/Contents
- Abstract
- As online learning platforms and computerized testing become more common, an increasing amount of data are collected about users. These data include, but are not limited to, response time, keystroke logs, and raw text. The desire to observe these features of the response process reflect an underly- ing interest in the cognitive processes and behaviors respondents engage in while taking tests and navigating learning platforms. Alongside this interest in response processes is an increased desire to use computational methods, artificial intelligence, and machine learning to combine this process data with traditional item response data. This dissertation attempts to advance a notion of computational validity, defined as arguments for the use and interpretation of test scores from evidence collected or analyzed using computational methods. Chapter one looks at item position effects as a threat to comparability across test forms. The presence of position effects is especially threatening to computer adaptive tests. Here a mixture item response theory model that decomposes item and person components of position effects is developed and applied to a large Brazilian college admissions exam to demonstrate its ability to both recover information about position effects in items and automatically produce comparable ability distributions across forms. Chapter two looks at data from a browser-based reading assessment and proposes drift diffusion models as a theoretically grounded way to combine response time and response accuracy data. The individual parameters estimated from the drift diffusion models are combined with scores on the browser-based assessment and used to predict scores on more expensive and time consuming measures of reading ability. This provides a line of evidence justifying the clinical and classroom use of the tool for diagnostic purposes. Chapter three looks at the issue of model selection. To make decisions from a model requires justification that the model is appropriate for the data. Current methods of model selection are well suited to make decisions on which model fits better, but not to communicate about how much better or if the degree of fit is appropriate. This paper uses a tool called the InterModel Vigorish that aims to provide measures of increases in predictive performance on an interpretable scale and extends it to polytomous item response theory models. Simulations are performed to explore what can be learned about data generating processes from this tool and then it is applied to a selection of empirical datasets. Combined, this work aims to highlight ways that modern computational methodologies can be used to improve not only assessments, but the decisions we make from them and our ability to communicate about them.
Description
Type of resource | text |
---|---|
Form | electronic resource; remote; computer; online resource |
Extent | 1 online resource. |
Place | California |
Place | [Stanford, California] |
Publisher | [Stanford University] |
Copyright date | 2023; ©2023 |
Publication date | 2023; 2023 |
Issuance | monographic |
Language | English |
Creators/Contributors
Author | Kanopka, Klint |
---|---|
Degree supervisor | Ruiz-Primo, Maria Araceli |
Thesis advisor | Ruiz-Primo, Maria Araceli |
Thesis advisor | Domingue, Ben |
Thesis advisor | Yeatman, Jason |
Degree committee member | Domingue, Ben |
Degree committee member | Yeatman, Jason |
Associated with | Stanford University, Graduate School of Education |
Subjects
Genre | Theses |
---|---|
Genre | Text |
Bibliographic information
Statement of responsibility | Klint Kanopka. |
---|---|
Note | Submitted to the Graduate School of Education. |
Thesis | Thesis Ph.D. Stanford University 2023. |
Location | https://purl.stanford.edu/tr545td7650 |
Access conditions
- Copyright
- © 2023 by Klint Kanopka
- License
- This work is licensed under a Creative Commons Attribution Non Commercial 3.0 Unported license (CC BY-NC).
Also listed in
Loading usage metrics...