Impact of Ignoring Nested Data Structures on Ability Estimation
The literature is clear that intentional or unintentional clustering of data elements typically results in the inflation of the estimated standard error of fixed parameter estimates. This study is unique in that it examines the impact of multilevel data structures on subject ability which are rando...
Main Author: | |
---|---|
Other Authors: | |
Format: | Others |
Published: |
Virginia Tech
2015
|
Subjects: | |
Online Access: | http://hdl.handle.net/10919/64197 |
id |
ndltd-VTETD-oai-vtechworks.lib.vt.edu-10919-64197 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-VTETD-oai-vtechworks.lib.vt.edu-10919-641972020-11-06T05:38:44Z Impact of Ignoring Nested Data Structures on Ability Estimation Shropshire, Kevin O'Neil Educational Leadership and Policy Studies Miyazaki, Yasuo House, Leanna L. Singh, Kusum Skaggs, Gary E. Savla, Jyoti S. Complex survey designs clustering PSU nested data multilevel data hierarchical data two-level HGLM three-level HGLM Rasch ability estimation The literature is clear that intentional or unintentional clustering of data elements typically results in the inflation of the estimated standard error of fixed parameter estimates. This study is unique in that it examines the impact of multilevel data structures on subject ability which are random effect predictions known as empirical Bayes estimates in the one-parameter IRT / Rasch model. The literature on the impact of complex survey design on latent trait models is mixed and there is no "best practice" established regarding how to handle this situation. A simulation study was conducted to address two questions related to ability estimation. First, what impacts does design based clustering have with respect to desirable statistical properties when estimating subject ability with the one-parameter IRT / Rasch model? Second, since empirical Bayes estimators have shrinkage properties, what impacts does clustering of first-stage sampling units have on measurement validity-does the first-stage sampling unit impact the ability estimate, and if so, is this desirable and equitable? Two models were fit to a factorial experimental design where the data were simulated over various conditions. The first model Rasch model formulated as a HGLM ignores the sample design (incorrect model) while the second incorporates a first-stage sampling unit (correct model). Study findings generally showed that the two models were comparable with respect to desirable statistical properties under a majority of the replicated conditions-more measurement error in ability estimation is found when the intra-class correlation is high and the item pool is small. In practice this is the exception rather than the norm. However, it was found that the empirical Bayes estimates were dependent upon the first-stage sampling unit raising the issue of equity and fairness in educational decision making. A real-world complex survey design with binary outcome data was also fit with both models. Analysis of the data supported the simulation design results which lead to the conclusion that modeling binary Rasch data may resort to a policy tradeoff between desirable statistical properties and measurement validity. Ph. D. 2015-11-26T07:01:30Z 2015-11-26T07:01:30Z 2014-06-03 Dissertation vt_gsexam:3063 http://hdl.handle.net/10919/64197 In Copyright http://rightsstatements.org/vocab/InC/1.0/ ETD application/pdf application/pdf Virginia Tech |
collection |
NDLTD |
format |
Others
|
sources |
NDLTD |
topic |
Complex survey designs clustering PSU nested data multilevel data hierarchical data two-level HGLM three-level HGLM Rasch ability estimation |
spellingShingle |
Complex survey designs clustering PSU nested data multilevel data hierarchical data two-level HGLM three-level HGLM Rasch ability estimation Shropshire, Kevin O'Neil Impact of Ignoring Nested Data Structures on Ability Estimation |
description |
The literature is clear that intentional or unintentional clustering of data elements typically results in the inflation of the estimated standard error of fixed parameter estimates. This study is unique in that it examines the impact of multilevel data structures on subject ability which are random effect predictions known as empirical Bayes estimates in the one-parameter IRT / Rasch model. The literature on the impact of complex survey design on latent trait models is mixed and there is no "best practice" established regarding how to handle this situation. A simulation study was conducted to address two questions related to ability estimation. First, what impacts does design based clustering have with respect to desirable statistical properties when estimating subject ability with the one-parameter IRT / Rasch model? Second, since empirical Bayes estimators have shrinkage properties, what impacts does clustering of first-stage sampling units have on measurement validity-does the first-stage sampling unit impact the ability estimate, and if so, is this desirable and equitable?
Two models were fit to a factorial experimental design where the data were simulated over various conditions. The first model Rasch model formulated as a HGLM ignores the sample design (incorrect model) while the second incorporates a first-stage sampling unit (correct model). Study findings generally showed that the two models were comparable with respect to desirable statistical properties under a majority of the replicated conditions-more measurement error in ability estimation is found when the intra-class correlation is high and the item pool is small. In practice this is the exception rather than the norm. However, it was found that the empirical Bayes estimates were dependent upon the first-stage sampling unit raising the issue of equity and fairness in educational decision making. A real-world complex survey design with binary outcome data was also fit with both models. Analysis of the data supported the simulation design results which lead to the conclusion that modeling binary Rasch data may resort to a policy tradeoff between desirable statistical properties and measurement validity. === Ph. D. |
author2 |
Educational Leadership and Policy Studies |
author_facet |
Educational Leadership and Policy Studies Shropshire, Kevin O'Neil |
author |
Shropshire, Kevin O'Neil |
author_sort |
Shropshire, Kevin O'Neil |
title |
Impact of Ignoring Nested Data Structures on Ability Estimation |
title_short |
Impact of Ignoring Nested Data Structures on Ability Estimation |
title_full |
Impact of Ignoring Nested Data Structures on Ability Estimation |
title_fullStr |
Impact of Ignoring Nested Data Structures on Ability Estimation |
title_full_unstemmed |
Impact of Ignoring Nested Data Structures on Ability Estimation |
title_sort |
impact of ignoring nested data structures on ability estimation |
publisher |
Virginia Tech |
publishDate |
2015 |
url |
http://hdl.handle.net/10919/64197 |
work_keys_str_mv |
AT shropshirekevinoneil impactofignoringnesteddatastructuresonabilityestimation |
_version_ |
1719355995348533248 |