Evaluating mode comparability in early elementary grades

With the widespread use of technology in the assessment field, many testing programs use both computer-based tests (CBTs) and paper-and-pencil tests (PPTs). Both the Standards for Educational and Psychological Testing (AERA, APA, & NCME, 2014) and the International Guidelines on Computer-Based a...

Full description

Bibliographic Details
Main Author: Lin, Ye
Other Authors: Welch, Catherine J.
Format: Others
Language:English
Published: University of Iowa 2018
Subjects:
Online Access:https://ir.uiowa.edu/etd/6604
https://ir.uiowa.edu/cgi/viewcontent.cgi?article=8103&context=etd
id ndltd-uiowa.edu-oai-ir.uiowa.edu-etd-8103
record_format oai_dc
spelling ndltd-uiowa.edu-oai-ir.uiowa.edu-etd-81032019-10-13T04:27:57Z Evaluating mode comparability in early elementary grades Lin, Ye With the widespread use of technology in the assessment field, many testing programs use both computer-based tests (CBTs) and paper-and-pencil tests (PPTs). Both the Standards for Educational and Psychological Testing (AERA, APA, & NCME, 2014) and the International Guidelines on Computer-Based and Internet Delivered Testing (International Test Commission, 2005) have called for studies on the equivalence of scores from different modes to support the uses and interpretations of scores across modes. Studies of administration mode effects, however, are quite limited and have found mixed results in the early childhood literature. In addition, little research has focused on both construct comparability and score comparability. The purpose of this study was to examine comparability in two stages. The first stage consisted of a series of analyses performed to investigate construct comparability through methods such as Confirmatory Factor Analysis (CFA), Multivariate Analysis of Variance (MANOVA), Classical Test Theory (CTT) and Item Response Theory (IRT). The second stage included summary analyses performed to investigate score comparability by evaluating the means, standard deviations, score distributions and reliabilities for the overall test scores. Correlations between the two modes and Test Characteristic Curves (TCCs) for the two modes were evaluated. Results indicated that, in general, the constructs and scores were comparable between PPTs and CBTs. The item and domain level analysis suggested that several items and domains seemed to be influenced slightly different by mode, while the scores at the total test level were not impacted by mode. This information could be useful for test developers when making decisions about what items to include on both modes. The current study sought to address gaps in the existing literature. First, this study examined how young test takers perform in a CBT testing environment. This work adds to previous literature in that young test takers have more access to technology than they did when many research studies were previously conducted. Second, this study discussed potential sources of mode effects such as test items and the characteristics of test takers in early elementary grades, another area in which comparability research is lacking. Third, the study evaluated comparability issues in two stages in a comprehensive manner. 2018-12-01T08:00:00Z dissertation application/pdf https://ir.uiowa.edu/etd/6604 https://ir.uiowa.edu/cgi/viewcontent.cgi?article=8103&context=etd Copyright © 2018 Ye Lin Theses and Dissertations eng University of IowaWelch, Catherine J. Dunbar, Stephen B. Educational Psychology
collection NDLTD
language English
format Others
sources NDLTD
topic Educational Psychology
spellingShingle Educational Psychology
Lin, Ye
Evaluating mode comparability in early elementary grades
description With the widespread use of technology in the assessment field, many testing programs use both computer-based tests (CBTs) and paper-and-pencil tests (PPTs). Both the Standards for Educational and Psychological Testing (AERA, APA, & NCME, 2014) and the International Guidelines on Computer-Based and Internet Delivered Testing (International Test Commission, 2005) have called for studies on the equivalence of scores from different modes to support the uses and interpretations of scores across modes. Studies of administration mode effects, however, are quite limited and have found mixed results in the early childhood literature. In addition, little research has focused on both construct comparability and score comparability. The purpose of this study was to examine comparability in two stages. The first stage consisted of a series of analyses performed to investigate construct comparability through methods such as Confirmatory Factor Analysis (CFA), Multivariate Analysis of Variance (MANOVA), Classical Test Theory (CTT) and Item Response Theory (IRT). The second stage included summary analyses performed to investigate score comparability by evaluating the means, standard deviations, score distributions and reliabilities for the overall test scores. Correlations between the two modes and Test Characteristic Curves (TCCs) for the two modes were evaluated. Results indicated that, in general, the constructs and scores were comparable between PPTs and CBTs. The item and domain level analysis suggested that several items and domains seemed to be influenced slightly different by mode, while the scores at the total test level were not impacted by mode. This information could be useful for test developers when making decisions about what items to include on both modes. The current study sought to address gaps in the existing literature. First, this study examined how young test takers perform in a CBT testing environment. This work adds to previous literature in that young test takers have more access to technology than they did when many research studies were previously conducted. Second, this study discussed potential sources of mode effects such as test items and the characteristics of test takers in early elementary grades, another area in which comparability research is lacking. Third, the study evaluated comparability issues in two stages in a comprehensive manner.
author2 Welch, Catherine J.
author_facet Welch, Catherine J.
Lin, Ye
author Lin, Ye
author_sort Lin, Ye
title Evaluating mode comparability in early elementary grades
title_short Evaluating mode comparability in early elementary grades
title_full Evaluating mode comparability in early elementary grades
title_fullStr Evaluating mode comparability in early elementary grades
title_full_unstemmed Evaluating mode comparability in early elementary grades
title_sort evaluating mode comparability in early elementary grades
publisher University of Iowa
publishDate 2018
url https://ir.uiowa.edu/etd/6604
https://ir.uiowa.edu/cgi/viewcontent.cgi?article=8103&context=etd
work_keys_str_mv AT linye evaluatingmodecomparabilityinearlyelementarygrades
_version_ 1719264572486975488