NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 3 results Save | Export
Peer reviewed Peer reviewed
Clayton Cohn; Surya Rayala; Caitlin Snyder; Joyce Horn Fonteles; Shruti Jain; Naveeduddin Mohammed; Umesh Timalsina; Sarah K. Burriss; Ashwin T. S.; Namrata Srivastava; Menton Deweese; Angela Eeds; Gautam Biswas – Grantee Submission, 2025
Collaborative dialogue offers rich insights into students' learning and critical thinking. This is essential for adapting pedagogical agents to students' learning and problem-solving skills in STEM+C settings. While large language models (LLMs) facilitate dynamic pedagogical interactions, potential hallucinations can undermine confidence, trust,…
Descriptors: STEM Education, Computer Science Education, Artificial Intelligence, Natural Language Processing
Peer reviewed Peer reviewed
Priti Oli; Rabin Banjade; Jeevan Chapagain; Vasile Rus – Grantee Submission, 2023
This paper systematically explores how Large Language Models (LLMs) generate explanations of code examples of the type used in intro-to-programming courses. As we show, the nature of code explanations generated by LLMs varies considerably based on the wording of the prompt, the target code examples being explained, the programming language, the…
Descriptors: Computational Linguistics, Programming, Computer Science Education, Programming Languages
Saira Anwar; Ahmed Ashraf Butt; Muhsin Menekse – Grantee Submission, 2023
This study explored the effectiveness of scaffolding in students' reflection writing process. We compared two sections of an introductory computer programming course (N=188). In Section 1, students did not receive any scaffolding while generating reflections, whereas in Section 2, students were scaffolded during the reflection writing process.…
Descriptors: Scaffolding (Teaching Technique), Writing Instruction, Writing Processes, Writing (Composition)