Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Assessing the Applicability of Artificial Intelligence (AI) in Fabrication and Plagiarism of Scientific Literature in the Field of Orthodontics
0
Zitationen
3
Autoren
2025
Jahr
Abstract
Objectives: The study aims to evaluate the applicability of Chat Generative Pre-trained Transformer (ChatGPT-4) to fabricate an orthodontic research paper and assess the ability of human reviewers (clinicians and academic reviewers) and artificial intelligence (AI) detection tools to identify such fabricated content. Methods: This study employed a descriptive exploratory research design. ChatGPT-4 was used to generate a research paper on the orthodontic topic. A search strategy was used to extract relevant research studies from three databases, then rewritten using ChatGPT-4. A panel of 10 orthodontic experts were given a mix of AI-rewritten and human-written abstracts to identify fake and real ones. Additionally, AI detection tools were tested for their efficiency in identifying AI-generated content. The results were analyzed using descriptive statistics and comparative percentages. Results: Orthodontic experts and researchers correctly identified only 68% of AIgenerated abstracts as fake, while 86% of human-written abstracts were recognized as real. Online AI detection tools demonstrated weak performance, with overall sensitivity and specificity of just 89% and accuracy of 44.5%. GPTZero had 100% specificity and 100% predictive values, along with the highest sensitivity and accuracy compared to others. However, writers AI demonstrates the poorest performance, indicating 0 predictive values in identifying AI-fabricated research works. Conclusions: Reviewers experienced difficulties in distinguishing between AI-generated abstracts and those written by humans due to the substantial resemblance to real papers. To safeguard scientific integrity, it is crucial to use measures such as improving detection tools, refining peer review, and creating a transparent culture around AI use.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.456 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.332 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.779 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.533 Zit.