Evaluating multiple-choice exams in large introductory physics courses
The reliability and validity of professionally written multiple-choice exams have been extensively studied for exams such as the SAT, graduate record examination, and the force concept inventory. Much of the success of these multiple-choice exams is attributed to the careful construction of each que...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
American Physical Society
2006-07-01
|
Series: | Physical Review Special Topics. Physics Education Research |
Online Access: | http://doi.org/10.1103/PhysRevSTPER.2.020102 |
id |
doaj-e47086ab92d44a4f8a2a3c4947d92b54 |
---|---|
record_format |
Article |
spelling |
doaj-e47086ab92d44a4f8a2a3c4947d92b542020-11-24T20:47:09ZengAmerican Physical SocietyPhysical Review Special Topics. Physics Education Research1554-91782006-07-012202010210.1103/PhysRevSTPER.2.020102Evaluating multiple-choice exams in large introductory physics coursesMichael ScottTim StelzerGary GladdingThe reliability and validity of professionally written multiple-choice exams have been extensively studied for exams such as the SAT, graduate record examination, and the force concept inventory. Much of the success of these multiple-choice exams is attributed to the careful construction of each question, as well as each response. In this study, the reliability and validity of scores from multiple-choice exams written for and administered in the large introductory physics courses at the University of Illinois, Urbana-Champaign were investigated. The reliability of exam scores over the course of a semester results in approximately a 3% uncertainty in students’ total semester exam score. This semester test score uncertainty yields an uncertainty in the students’ assigned letter grade that is less than 1/3 of a letter grade. To study the validity of exam scores, a subset of students were ranked independently based on their multiple-choice score, graded explanations, and student interviews. The ranking of these students based on their multiple-choice score was found to be consistent with the ranking assigned by physics instructors based on the students’ written explanations (r>0.94 at the 95% confidence level) and oral interviews (r=0.94_{−0.09}^{+0.06}).http://doi.org/10.1103/PhysRevSTPER.2.020102 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Michael Scott Tim Stelzer Gary Gladding |
spellingShingle |
Michael Scott Tim Stelzer Gary Gladding Evaluating multiple-choice exams in large introductory physics courses Physical Review Special Topics. Physics Education Research |
author_facet |
Michael Scott Tim Stelzer Gary Gladding |
author_sort |
Michael Scott |
title |
Evaluating multiple-choice exams in large introductory physics courses |
title_short |
Evaluating multiple-choice exams in large introductory physics courses |
title_full |
Evaluating multiple-choice exams in large introductory physics courses |
title_fullStr |
Evaluating multiple-choice exams in large introductory physics courses |
title_full_unstemmed |
Evaluating multiple-choice exams in large introductory physics courses |
title_sort |
evaluating multiple-choice exams in large introductory physics courses |
publisher |
American Physical Society |
series |
Physical Review Special Topics. Physics Education Research |
issn |
1554-9178 |
publishDate |
2006-07-01 |
description |
The reliability and validity of professionally written multiple-choice exams have been extensively studied for exams such as the SAT, graduate record examination, and the force concept inventory. Much of the success of these multiple-choice exams is attributed to the careful construction of each question, as well as each response. In this study, the reliability and validity of scores from multiple-choice exams written for and administered in the large introductory physics courses at the University of Illinois, Urbana-Champaign were investigated. The reliability of exam scores over the course of a semester results in approximately a 3% uncertainty in students’ total semester exam score. This semester test score uncertainty yields an uncertainty in the students’ assigned letter grade that is less than 1/3 of a letter grade. To study the validity of exam scores, a subset of students were ranked independently based on their multiple-choice score, graded explanations, and student interviews. The ranking of these students based on their multiple-choice score was found to be consistent with the ranking assigned by physics instructors based on the students’ written explanations (r>0.94 at the 95% confidence level) and oral interviews (r=0.94_{−0.09}^{+0.06}). |
url |
http://doi.org/10.1103/PhysRevSTPER.2.020102 |
work_keys_str_mv |
AT michaelscott evaluatingmultiplechoiceexamsinlargeintroductoryphysicscourses AT timstelzer evaluatingmultiplechoiceexamsinlargeintroductoryphysicscourses AT garygladding evaluatingmultiplechoiceexamsinlargeintroductoryphysicscourses |
_version_ |
1716810981079777280 |