Inter-rater agreement in the scoring of abstracts submitted to a primary care research conference

<p>Abstract</p> <p>Background</p> <p>Checklists for peer review aim to guide referees when assessing the quality of papers, but little evidence exists on the extent to which referees agree when evaluating the same paper. The aim of this study was to investigate agreemen...

Full description

Bibliographic Details
Main Authors: Evans Philip H, Graham Anna, Montgomery Alan A, Fahey Tom
Format: Article
Language:English
Published: BMC 2002-03-01
Series:BMC Health Services Research
Online Access:http://www.biomedcentral.com/1472-6963/2/8
id doaj-7a59647b04a34f08902f64ab29afbeb5
record_format Article
spelling doaj-7a59647b04a34f08902f64ab29afbeb52020-11-25T02:34:03ZengBMCBMC Health Services Research1472-69632002-03-0121810.1186/1472-6963-2-8Inter-rater agreement in the scoring of abstracts submitted to a primary care research conferenceEvans Philip HGraham AnnaMontgomery Alan AFahey Tom<p>Abstract</p> <p>Background</p> <p>Checklists for peer review aim to guide referees when assessing the quality of papers, but little evidence exists on the extent to which referees agree when evaluating the same paper. The aim of this study was to investigate agreement on dimensions of a checklist between two referees when evaluating abstracts submitted for a primary care conference.</p> <p>Methods</p> <p>Anonymised abstracts were scored using a structured assessment comprising seven categories. Between one (poor) and four (excellent) marks were awarded for each category, giving a maximum possible score of 28 marks. Every abstract was assessed independently by two referees and agreement measured using intraclass correlation coefficients. Mean total scores of abstracts accepted and rejected for the meeting were compared using an unpaired t test.</p> <p>Results</p> <p>Of 52 abstracts, agreement between reviewers was greater for three components relating to study design (adjusted intraclass correlation coefficients 0.40 to 0.45) compared to four components relating to more subjective elements such as the importance of the study and likelihood of provoking discussion (0.01 to 0.25). Mean score for accepted abstracts was significantly greater than those that were rejected (17.4 versus 14.6, 95% CI for difference 1.3 to 4.1, p = 0.0003).</p> <p>Conclusions</p> <p>The findings suggest that inclusion of subjective components in a review checklist may result in greater disagreement between reviewers. However in terms of overall quality scores, abstracts accepted for the meeting were rated significantly higher than those that were rejected.</p> http://www.biomedcentral.com/1472-6963/2/8
collection DOAJ
language English
format Article
sources DOAJ
author Evans Philip H
Graham Anna
Montgomery Alan A
Fahey Tom
spellingShingle Evans Philip H
Graham Anna
Montgomery Alan A
Fahey Tom
Inter-rater agreement in the scoring of abstracts submitted to a primary care research conference
BMC Health Services Research
author_facet Evans Philip H
Graham Anna
Montgomery Alan A
Fahey Tom
author_sort Evans Philip H
title Inter-rater agreement in the scoring of abstracts submitted to a primary care research conference
title_short Inter-rater agreement in the scoring of abstracts submitted to a primary care research conference
title_full Inter-rater agreement in the scoring of abstracts submitted to a primary care research conference
title_fullStr Inter-rater agreement in the scoring of abstracts submitted to a primary care research conference
title_full_unstemmed Inter-rater agreement in the scoring of abstracts submitted to a primary care research conference
title_sort inter-rater agreement in the scoring of abstracts submitted to a primary care research conference
publisher BMC
series BMC Health Services Research
issn 1472-6963
publishDate 2002-03-01
description <p>Abstract</p> <p>Background</p> <p>Checklists for peer review aim to guide referees when assessing the quality of papers, but little evidence exists on the extent to which referees agree when evaluating the same paper. The aim of this study was to investigate agreement on dimensions of a checklist between two referees when evaluating abstracts submitted for a primary care conference.</p> <p>Methods</p> <p>Anonymised abstracts were scored using a structured assessment comprising seven categories. Between one (poor) and four (excellent) marks were awarded for each category, giving a maximum possible score of 28 marks. Every abstract was assessed independently by two referees and agreement measured using intraclass correlation coefficients. Mean total scores of abstracts accepted and rejected for the meeting were compared using an unpaired t test.</p> <p>Results</p> <p>Of 52 abstracts, agreement between reviewers was greater for three components relating to study design (adjusted intraclass correlation coefficients 0.40 to 0.45) compared to four components relating to more subjective elements such as the importance of the study and likelihood of provoking discussion (0.01 to 0.25). Mean score for accepted abstracts was significantly greater than those that were rejected (17.4 versus 14.6, 95% CI for difference 1.3 to 4.1, p = 0.0003).</p> <p>Conclusions</p> <p>The findings suggest that inclusion of subjective components in a review checklist may result in greater disagreement between reviewers. However in terms of overall quality scores, abstracts accepted for the meeting were rated significantly higher than those that were rejected.</p>
url http://www.biomedcentral.com/1472-6963/2/8
work_keys_str_mv AT evansphiliph interrateragreementinthescoringofabstractssubmittedtoaprimarycareresearchconference
AT grahamanna interrateragreementinthescoringofabstractssubmittedtoaprimarycareresearchconference
AT montgomeryalana interrateragreementinthescoringofabstractssubmittedtoaprimarycareresearchconference
AT faheytom interrateragreementinthescoringofabstractssubmittedtoaprimarycareresearchconference
_version_ 1724810506737287168