Evaluating human versus machine learning performance in classifying research abstracts

Yeow Chong Goh, Xin Qing Cai, Walter Theseira, Giovanni Ko, Khiam Aik Khor*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

40 Citations (Scopus)

Abstract

We study whether humans or machine learning (ML) classification models are better at classifying scientific research abstracts according to a fixed set of discipline groups. We recruit both undergraduate and postgraduate assistants for this task in separate stages, and compare their performance against the support vectors machine ML algorithm at classifying European Research Council Starting Grant project abstracts to their actual evaluation panels, which are organised by discipline groups. On average, ML is more accurate than human classifiers, across a variety of training and test datasets, and across evaluation panels. ML classifiers trained on different training sets are also more reliable than human classifiers, meaning that different ML classifiers are more consistent in assigning the same classifications to any given abstract, compared to different human classifiers. While the top five percentile of human classifiers can outperform ML in limited cases, selection and training of such classifiers is likely costly and difficult compared to training ML models. Our results suggest ML models are a cost effective and highly accurate method for addressing problems in comparative bibliometric analysis, such as harmonising the discipline classifications of research from different funding agencies or countries.

Original languageEnglish
Pages (from-to)1197-1212
Number of pages16
JournalScientometrics
Volume125
Issue number2
DOIs
Publication statusPublished - Nov 2020
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2020, The Author(s).

ASJC Scopus Subject Areas

  • General Social Sciences
  • Computer Science Applications
  • Library and Information Sciences

Keywords

  • Discipline classification
  • Supervised classification
  • Text classification

Fingerprint

Dive into the research topics of 'Evaluating human versus machine learning performance in classifying research abstracts'. Together they form a unique fingerprint.

Cite this