Impact of Ignoring Nested Data Structures on Ability Estimation

The literature is clear that intentional or unintentional clustering of data elements typically results in the inflation of the estimated standard error of fixed parameter estimates. This study is unique in that it examines the impact of multilevel data structures on subject ability which are rando...

Full description

Bibliographic Details
Main Author: Shropshire, Kevin O'Neil
Other Authors: Educational Leadership and Policy Studies
Format: Others
Published: Virginia Tech 2015
Subjects:
PSU
Online Access:http://hdl.handle.net/10919/64197
id ndltd-VTETD-oai-vtechworks.lib.vt.edu-10919-64197
record_format oai_dc
spelling ndltd-VTETD-oai-vtechworks.lib.vt.edu-10919-641972020-11-06T05:38:44Z Impact of Ignoring Nested Data Structures on Ability Estimation Shropshire, Kevin O'Neil Educational Leadership and Policy Studies Miyazaki, Yasuo House, Leanna L. Singh, Kusum Skaggs, Gary E. Savla, Jyoti S. Complex survey designs clustering PSU nested data multilevel data hierarchical data two-level HGLM three-level HGLM Rasch ability estimation The literature is clear that intentional or unintentional clustering of data elements typically results in the inflation of the estimated standard error of fixed parameter estimates. This study is unique in that it examines the impact of multilevel data structures on subject ability which are random effect predictions known as empirical Bayes estimates in the one-parameter IRT / Rasch model. The literature on the impact of complex survey design on latent trait models is mixed and there is no "best practice" established regarding how to handle this situation. A simulation study was conducted to address two questions related to ability estimation. First, what impacts does design based clustering have with respect to desirable statistical properties when estimating subject ability with the one-parameter IRT / Rasch model? Second, since empirical Bayes estimators have shrinkage properties, what impacts does clustering of first-stage sampling units have on measurement validity-does the first-stage sampling unit impact the ability estimate, and if so, is this desirable and equitable? Two models were fit to a factorial experimental design where the data were simulated over various conditions. The first model Rasch model formulated as a HGLM ignores the sample design (incorrect model) while the second incorporates a first-stage sampling unit (correct model). Study findings generally showed that the two models were comparable with respect to desirable statistical properties under a majority of the replicated conditions-more measurement error in ability estimation is found when the intra-class correlation is high and the item pool is small. In practice this is the exception rather than the norm. However, it was found that the empirical Bayes estimates were dependent upon the first-stage sampling unit raising the issue of equity and fairness in educational decision making. A real-world complex survey design with binary outcome data was also fit with both models. Analysis of the data supported the simulation design results which lead to the conclusion that modeling binary Rasch data may resort to a policy tradeoff between desirable statistical properties and measurement validity. Ph. D. 2015-11-26T07:01:30Z 2015-11-26T07:01:30Z 2014-06-03 Dissertation vt_gsexam:3063 http://hdl.handle.net/10919/64197 In Copyright http://rightsstatements.org/vocab/InC/1.0/ ETD application/pdf application/pdf Virginia Tech
collection NDLTD
format Others
sources NDLTD
topic Complex survey designs
clustering
PSU
nested data
multilevel data
hierarchical data
two-level HGLM
three-level HGLM
Rasch
ability estimation
spellingShingle Complex survey designs
clustering
PSU
nested data
multilevel data
hierarchical data
two-level HGLM
three-level HGLM
Rasch
ability estimation
Shropshire, Kevin O'Neil
Impact of Ignoring Nested Data Structures on Ability Estimation
description The literature is clear that intentional or unintentional clustering of data elements typically results in the inflation of the estimated standard error of fixed parameter estimates. This study is unique in that it examines the impact of multilevel data structures on subject ability which are random effect predictions known as empirical Bayes estimates in the one-parameter IRT / Rasch model. The literature on the impact of complex survey design on latent trait models is mixed and there is no "best practice" established regarding how to handle this situation. A simulation study was conducted to address two questions related to ability estimation. First, what impacts does design based clustering have with respect to desirable statistical properties when estimating subject ability with the one-parameter IRT / Rasch model? Second, since empirical Bayes estimators have shrinkage properties, what impacts does clustering of first-stage sampling units have on measurement validity-does the first-stage sampling unit impact the ability estimate, and if so, is this desirable and equitable? Two models were fit to a factorial experimental design where the data were simulated over various conditions. The first model Rasch model formulated as a HGLM ignores the sample design (incorrect model) while the second incorporates a first-stage sampling unit (correct model). Study findings generally showed that the two models were comparable with respect to desirable statistical properties under a majority of the replicated conditions-more measurement error in ability estimation is found when the intra-class correlation is high and the item pool is small. In practice this is the exception rather than the norm. However, it was found that the empirical Bayes estimates were dependent upon the first-stage sampling unit raising the issue of equity and fairness in educational decision making. A real-world complex survey design with binary outcome data was also fit with both models. Analysis of the data supported the simulation design results which lead to the conclusion that modeling binary Rasch data may resort to a policy tradeoff between desirable statistical properties and measurement validity. === Ph. D.
author2 Educational Leadership and Policy Studies
author_facet Educational Leadership and Policy Studies
Shropshire, Kevin O'Neil
author Shropshire, Kevin O'Neil
author_sort Shropshire, Kevin O'Neil
title Impact of Ignoring Nested Data Structures on Ability Estimation
title_short Impact of Ignoring Nested Data Structures on Ability Estimation
title_full Impact of Ignoring Nested Data Structures on Ability Estimation
title_fullStr Impact of Ignoring Nested Data Structures on Ability Estimation
title_full_unstemmed Impact of Ignoring Nested Data Structures on Ability Estimation
title_sort impact of ignoring nested data structures on ability estimation
publisher Virginia Tech
publishDate 2015
url http://hdl.handle.net/10919/64197
work_keys_str_mv AT shropshirekevinoneil impactofignoringnesteddatastructuresonabilityestimation
_version_ 1719355995348533248