Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Research evaluation with ChatGPT: is it age, country, length, or field biased?
8
Zitationen
2
Autoren
2025
Jahr
Abstract
Abstract Some research now suggests that ChatGPT can estimate the quality of journal articles from their titles and abstracts. This has created the possibility to use ChatGPT quality scores, perhaps alongside citation-based formulae, to support peer review for research evaluation. Nevertheless, ChatGPT’s internal processes are effectively opaque, despite it writing a report to support its scores, and its biases are unknown. This article investigates whether publication date and field are biasing factors. Based on submitting a monodisciplinary journal-balanced set of 117,650 articles from 26 fields published in the years 2003, 2008, 2013, 2018 and 2023 to ChatGPT 4o-mini, the results show that average scores increased over time, and this was not due to author nationality or title and abstract length changes. The results also varied substantially between fields, and first author countries. In addition, articles with longer abstracts tended to receive higher scores, mostly due to such articles tending to be better (e.g., more likely to be in higher impact journals) but also partly due to ChatGPT analysing more text. For the most accurate research quality evaluation results from ChatGPT, it is important to normalise ChatGPT scores for field and year and check for anomalies caused by sets of articles with short abstracts.
Ähnliche Arbeiten
An index to quantify an individual's scientific research output
2005 · 11.350 Zit.
How to conduct a bibliometric analysis: An overview and guidelines
2021 · 11.060 Zit.
Estimating the reproducibility of psychological science
2015 · 8.537 Zit.
Analyzing the Past to Prepare for the Future: Writing a Literature Review
2002 · 6.939 Zit.
Bibliometric Methods in Management and Organization
2014 · 6.204 Zit.