Framework for evaluating statistical models in physics education research

Across the field of education research there has been an increased focus on the development, critique, and evaluation of statistical methods and data usage due to recently created, very large datasets and machine learning techniques. In physics education research (PER), this increased focus has rece...

Full description

Bibliographic Details
Main Authors: John M. Aiken, Riccardo De Bin, H. J. Lewandowski, Marcos D. Caballero
Format: Article
Language:English
Published: American Physical Society 2021-07-01
Series:Physical Review Physics Education Research
Online Access:http://doi.org/10.1103/PhysRevPhysEducRes.17.020104
id doaj-214d97928e824b449e4598b59d307920
record_format Article
spelling doaj-214d97928e824b449e4598b59d3079202021-07-28T14:21:18ZengAmerican Physical SocietyPhysical Review Physics Education Research2469-98962021-07-0117202010410.1103/PhysRevPhysEducRes.17.020104Framework for evaluating statistical models in physics education researchJohn M. AikenRiccardo De BinH. J. LewandowskiMarcos D. CaballeroAcross the field of education research there has been an increased focus on the development, critique, and evaluation of statistical methods and data usage due to recently created, very large datasets and machine learning techniques. In physics education research (PER), this increased focus has recently been shown through the 2019 Physical Review PER Focused Collection examining quantitative methods in PER. Quantitative PER has provided strong arguments for reforming courses by including interactive engagement, demonstrated that students often move away from scientistlike views due to science education, and has injected robust assessment into the physics classroom via concept inventories. The work presented here examines the impact that machine learning may have on physics education research, presents a framework for the entire process including data management, model evaluation, and results communication, and demonstrates the utility of this framework through the analysis of two types of survey data.http://doi.org/10.1103/PhysRevPhysEducRes.17.020104
collection DOAJ
language English
format Article
sources DOAJ
author John M. Aiken
Riccardo De Bin
H. J. Lewandowski
Marcos D. Caballero
spellingShingle John M. Aiken
Riccardo De Bin
H. J. Lewandowski
Marcos D. Caballero
Framework for evaluating statistical models in physics education research
Physical Review Physics Education Research
author_facet John M. Aiken
Riccardo De Bin
H. J. Lewandowski
Marcos D. Caballero
author_sort John M. Aiken
title Framework for evaluating statistical models in physics education research
title_short Framework for evaluating statistical models in physics education research
title_full Framework for evaluating statistical models in physics education research
title_fullStr Framework for evaluating statistical models in physics education research
title_full_unstemmed Framework for evaluating statistical models in physics education research
title_sort framework for evaluating statistical models in physics education research
publisher American Physical Society
series Physical Review Physics Education Research
issn 2469-9896
publishDate 2021-07-01
description Across the field of education research there has been an increased focus on the development, critique, and evaluation of statistical methods and data usage due to recently created, very large datasets and machine learning techniques. In physics education research (PER), this increased focus has recently been shown through the 2019 Physical Review PER Focused Collection examining quantitative methods in PER. Quantitative PER has provided strong arguments for reforming courses by including interactive engagement, demonstrated that students often move away from scientistlike views due to science education, and has injected robust assessment into the physics classroom via concept inventories. The work presented here examines the impact that machine learning may have on physics education research, presents a framework for the entire process including data management, model evaluation, and results communication, and demonstrates the utility of this framework through the analysis of two types of survey data.
url http://doi.org/10.1103/PhysRevPhysEducRes.17.020104
work_keys_str_mv AT johnmaiken frameworkforevaluatingstatisticalmodelsinphysicseducationresearch
AT riccardodebin frameworkforevaluatingstatisticalmodelsinphysicseducationresearch
AT hjlewandowski frameworkforevaluatingstatisticalmodelsinphysicseducationresearch
AT marcosdcaballero frameworkforevaluatingstatisticalmodelsinphysicseducationresearch
_version_ 1721268134569050112