Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Student motivation and need satisfaction in GenAI-supported classrooms: A self-determination theory perspective
0
Zitationen
3
Autoren
2026
Jahr
Abstract
This study examines how generative AI (GenAI), specifically ChatGPT, relates to student motivation and basic psychological need satisfaction in lower-secondary education from a self-determination theory perspective. While generative AI is increasingly used in classrooms, it remains unclear how its use is associated with different motivational profiles across instructional contexts. Using a cross-sectional sample of 2,469 students (grades 7–8), we compared two instructional contexts: (a) rubric-based, self-regulated competency-based learning (CBL) implemented with or without ChatGPT integration, and (b) teacher-directed learning (TDL) organized in a traditional 45-minute lesson format with limited self-regulated phases. To capture both overall differences and profile-specific patterns, we combined person-centered latent profile analysis with variable-centered mean comparisons. The results show that ChatGPT use in CBL is associated with higher autonomy support and competence satisfaction, but also with a greater proportion of low-quality motivational profiles compared to non-AI CBL. At the same time, these profiles in the ChatGPT-supported context exhibit higher intrinsic and identified motivation than comparable profiles in both non-AI CBL and TDL. In contrast, relatedness appears less pronounced in the ChatGPT-supported context, which may help explain why the overall motivational distribution does not exceed the already high baseline of non-AI CBL. These findings point to systematic differences in motivational quality associated with ChatGPT use and highlight the importance of considering learner heterogeneity when integrating generative AI into autonomy-supportive instructional settings.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.393 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.259 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.688 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.502 Zit.