Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Exploring students’ lived experiences with artificial intelligence in higher education: A case study of ChatGPT
1
Zitationen
2
Autoren
2025
Jahr
Abstract
This study adopted a phenomenological perspective to explore 84 university students’ experiences and perceptions of using Chat Generative Pre-trained Transformer (ChatGPT), an artificial intelligence (AI)-based software that generates contextualised responses. This study aimed to capture individual experiences and to understand how students interact with and perceive this AI tool; to fulfil this purpose, we addressed the following question: What are university students’ perceptions and experiences of using ChatGPT in their academic activities? The triangulation strategy was employed, which allowed us to analyse the results by combining information from different sources: narrative interviews, literature review, and direct unstructured observations. The results indicate that student responses reveal a positive perception towards ChatGPT; however, when examining lived experiences, concerns arise about the potential impact on developing critical and analytical skills. Such concerns underscore the ongoing need to investigate the role of AI in education, emphasising the importance of better understanding how students, from their subjective experience, interact with and relate to technology to inform and refine educational practices. We conclude that, despite the usefulness of ChatGPT, it is crucial to recognise its limitations and make informed decisions, especially in contexts that require a deep understanding of human intentions and emotions. Furthermore, it is important to consider the context in which ChatGPT is used and be aware of its potential biases or inaccuracies, as it may be influenced by biases present in its training data, which could result in misleading or incorrect responses.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.336 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.207 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.607 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.476 Zit.