TY - JOUR
T1 - Challenging ChatGPT with Different Types of Physics Education Questions
AU - López Simó, Víctor
AU - Rezende Junior, Mikael Frank
PY - 2024/4/1
Y1 - 2024/4/1
N2 - Will my job be replaced by ChatGPT? Can artificial intelligence (AI) engines do homework for students? How do I know if a delivered assignment was made by a robot? These and many other questions have been occupying the minds of professionals from different areas, including professors and researchers, especially since ChatGPT was launched in November 2022. This Generative Pretrained Transformer mechanism works through a chat interface that allows establishing conversations based on the targeted processing of a large volume of data, but the inner functioning of ChatGPT still acts as a "black box" for most of us. As physics educators, we are particularly interested in understanding the kind of information it can provide to students, how reliable this information can be, and where it may still fall short. This helps us better understand how we can use it. In the last months, different investigations have indicated the need for detailed studies to better understand both the potentialities and limitations of AI in physics teaching and learning scenarios. On the one hand, and presented different strategies to use Chat GTP in the physics classroom, presenting easy-to-implement examples of how ChatGPT can be used in physics classrooms to foster critical thinking skills at the secondary school level [1] and to generate bad examples to be addressed with students to critique and fix them [2]. On the other hand, some investigations have analyzed the level of performance of this IA tool to solve physics problems. According to [3], ChatGPT would narrowly pass a calculus-based physics course while exhibiting many of the preconceptions and errors of a beginning learner. In parallel, found that ChatGPT3.5 can match or exceed the median performance of a university student who has completed one semester of college physics, and found very impressive basic problem-solving capabilities of ChatGPT in interpreting simple physics problems, assuming relevant parameters, and writing correct codes. Despite those previous contributions focus either on identifying ChatGPT-based physics education good practices or testing ChatGPT physics' performance in comparison with real students, our particular interest lies in understanding how different typologies of physics education problems may influence both the correctness and the variability of the answers provided by the tool. It is well known in Physics Education Research that the typology of physics education questions strongly affects the ways of reasoning and obtaining the answer. For this reason, our research question is: How the correctness and the variability of the answers provided by ChatGPT are affected by the typology of physics education question?
AB - Will my job be replaced by ChatGPT? Can artificial intelligence (AI) engines do homework for students? How do I know if a delivered assignment was made by a robot? These and many other questions have been occupying the minds of professionals from different areas, including professors and researchers, especially since ChatGPT was launched in November 2022. This Generative Pretrained Transformer mechanism works through a chat interface that allows establishing conversations based on the targeted processing of a large volume of data, but the inner functioning of ChatGPT still acts as a "black box" for most of us. As physics educators, we are particularly interested in understanding the kind of information it can provide to students, how reliable this information can be, and where it may still fall short. This helps us better understand how we can use it. In the last months, different investigations have indicated the need for detailed studies to better understand both the potentialities and limitations of AI in physics teaching and learning scenarios. On the one hand, and presented different strategies to use Chat GTP in the physics classroom, presenting easy-to-implement examples of how ChatGPT can be used in physics classrooms to foster critical thinking skills at the secondary school level [1] and to generate bad examples to be addressed with students to critique and fix them [2]. On the other hand, some investigations have analyzed the level of performance of this IA tool to solve physics problems. According to [3], ChatGPT would narrowly pass a calculus-based physics course while exhibiting many of the preconceptions and errors of a beginning learner. In parallel, found that ChatGPT3.5 can match or exceed the median performance of a university student who has completed one semester of college physics, and found very impressive basic problem-solving capabilities of ChatGPT in interpreting simple physics problems, assuming relevant parameters, and writing correct codes. Despite those previous contributions focus either on identifying ChatGPT-based physics education good practices or testing ChatGPT physics' performance in comparison with real students, our particular interest lies in understanding how different typologies of physics education problems may influence both the correctness and the variability of the answers provided by the tool. It is well known in Physics Education Research that the typology of physics education questions strongly affects the ways of reasoning and obtaining the answer. For this reason, our research question is: How the correctness and the variability of the answers provided by ChatGPT are affected by the typology of physics education question?
KW - Chat GPT
KW - Physics Education
KW - Artificial intelligence
KW - Physics teaching and learning
KW - Secondary school level
UR - http://www.scopus.com/inward/record.url?scp=85189307933&partnerID=8YFLogxK
UR - https://www.mendeley.com/catalogue/1262bbf2-2b3d-306d-9235-e0d2fa9eee6e/
U2 - 10.1119/5.0160160
DO - 10.1119/5.0160160
M3 - Article
SN - 0031-921X
VL - 62
SP - 290
EP - 294
JO - Physics Teacher
JF - Physics Teacher
IS - 4
ER -