Abstract
Novice students make various mistakes in the process of learning computer programming. In courses with more than 100 students, it is difficult to provide accurate and detailed feedback regarding errors in the source code submitted for their assignments. Therefore, we created a source code analyzer and developed a tool to provide detailed feedback to each student. It performs unit tests with misspelled classes and method names. From the results, the tool generates comments, such as "Let us check the method name" or "Let us check the execution result.” The tool can generate an average of more than 8,000 Japanese characters per assignment in an actual programming lecture with more than 100 students. In this study, we report on the developed tool, its adaptation to an existing learning management system, and its evaluation.
Author Information
Tatsuyuki Takano, Kanto Gakuin University, Japan
Osamu Miyakawa, Tokyo Denki University, Japan
Takashi Kohama, Tokyo Denki University, Japan
Paper Information
Conference: ACEID2023
Stream: Design
This paper is part of the ACEID2023 Conference Proceedings (View)
Full Paper
View / Download the full paper in a new tab/window
To cite this article:
Takano T., Miyakawa O., & Kohama T. (2023) Development of a Tool to Analyze Source Code Submitted by Novice Programmers and Provide Learning Support Feedback With Comments ISSN: 2189-101X – The Asian Conference on Education & International Development 2023 Official Conference Proceedings (pp. 777-789) https://doi.org/10.22492/issn.2189-101X.2023.64
To link to this article: https://doi.org/10.22492/issn.2189-101X.2023.64
Comments
Powered by WP LinkPress