Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Challenges and Improvements in Mathematical Reasoning for Large Models
0
Zitationen
1
Autoren
2026
Jahr
Abstract
Mathematical reasoning is a fundamental criterion for AI, and this is also a problem to be solved before arriving at AGI. Large Language Models(LLMs) have become prominent in this domain.. LLM’s benchmark performance (GSM8K, MATH): However, a large number of investigation indicate that they only mimic what the training data are, so they do not properly understand logical problems, and that means their reasoning ability is extraordinarily flaky, meaning that even a small perturbation in how problems are phrased could completely alter the reasoning which is really the difference between being smart statistically and understanding. So, if you are going to try and reason around these sorts of limitations, to find a way to a more robust application of math intelligence, you need to consider all research directions being explored right now. Analysis is broken down into 3 directions:(1) The first direction involves methods that include tweaking methods from a processing intervention standpoint towards reasoning chains. (2). The second direction is placeholder self-verification and reflection, where the models are urged to think about and improve their own reasoning. (3). The third direction is scientific verification and a measure of efficiency, focusing on thorough benchmarking and ways to test like tool integration. This survey is meaningful, as it is on a topic that is ever-changing, and will provide a baseline to help a lay audience, in addition to showcasing what the state-of-the-art approaches are, while still critically examining whether it can overcome any of the inherent challenges associated with LLMs. I hope it can be useful in conducting research in the future for more logical/more trustworthy AIs.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.239 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.095 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.463 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.428 Zit.