Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Comparative Analysis of ChatGPT, DeepSeek, and Gemini for Automated Code Generation
3
Zitationen
2
Autoren
2025
Jahr
Abstract
Code generation using artificial intelligence (AI) has revolutionized software development, providing automated coding solutions. This study conducts a systematic comparative analysis of three leading large language models (LLMs) such as ChatGPT (O1), DeepSeek (R1) and Gemini (2.0 Flash thinking), for Python code generation, evaluating their performance in correctness, code quality, and computational efficiency. Using a curated dataset of Codeforces programming problems that span various difficulty levels <tex xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">$(\mathbf{8 0 0 - 2 0 0 0}$</tex> complexity), the research employs a rigorous evaluation framework that integrates online judge validation, static code analysis, and runtime profiling. The experimental results reveal that DeepSeek achieves comparatively higher correctness by consistently producing accepted solutions in fewer attempts, although its reasoning time increases with problem complexity. Gemini, on the other hand, is remarkably fast, delivering results in a fraction of the time, but its correctness deteriorates on more complex tasks. ChatGPT offers balanced performance with intermediate correctness and efficiency; however, it sometimes exhibits lower code quality. Overall, our findings underscore the inherent trade-offs between efficiency, accuracy, and quality in AI-generated code. The study provides actionable insights for developers, emphasizing the need to align model selection with project requirements, and contributes a replicable framework for future evaluations of AI code generation tools.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.245 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.102 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.468 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.429 Zit.