APPLYING AUTOMATIC CODE GRADING SYSTEM FOR TEACHING PROGRAMMING LANGUAGES IN UNIVERSITIES | Hoàn | TNU Journal of Science and Technology

APPLYING AUTOMATIC CODE GRADING SYSTEM FOR TEACHING PROGRAMMING LANGUAGES IN UNIVERSITIES

About this article

Received: 01/12/23                Revised: 27/12/23                Published: 27/12/23

Authors

1. Le Hoan Email to author, Electric Power University
2. Nguyen Quynh Anh, Electric Power University
3. Pham Nhat Linh, Electric Power University

Abstract


The rapid growth of online education platforms highlights the need for effective automatic grading systems, especially for programming courses in universities. This paper introduces an Automatic Code Grading System named AutoChecking, tailored for evaluating code submissions in various programming languages at the university level. AutoChecking combines static code analysis and dynamic execution to ensure thorough and equitable assessment of students' programming assignments. It features a sandbox environment for securely running student codes against predefined test cases, evaluating not just accuracy but also efficiency and coding style. The system is adaptable, with a flexible scoring mechanism to meet different course and instructor needs. AutoChecking integrates with existing online Learning Management Systems to streamline submission, grading, and feedback, making it easier for instructors to manage assignments and monitor student progress with detailed analytics. We evaluated the effectiveness of the AutoChecking through a semester-long deployment in introductory and advanced programming courses at three universities. The results showed that AutoChecking saves instructors' time, provides consistent feedback to students, enhancing their learning, and plays a crucial role in maintaining academic integrity by detecting code similarities.

Keywords


AutoChecking; Learning management systems; Automatic Code Grading System; Automated assessment tools; Sandbox environment

Full Text:

PDF

References


[1] J. C. Paiva, J. P. Leal, and Á. Figueira, “Automated Assessment in Computer Science Education: A State-of-the-Art Review,” ACM Trans. Comput. Educ., vol. 22, no. 3, pp. 34:1-34:40, Jun. 2022, doi: 10.1145/3513140.

[2] J. Schneider, R. Richner, and M. Riser, “Towards Trustworthy AutoGrading of Short, Multi-lingual, Multi-type Answers,” Int. J. Artif. Intell. Educ., vol. 33, no. 1, pp. 88–118, Mar. 2023, doi: 10.1007/s40593-022-00289-z.

[3] Z. Cakmak and I. Akgün, “A Theoretical Perspective on the Case Study Method,” J. Educ. Learn., vol. 7, p. 96, Oct. 2017, doi: 10.5539/jel.v7n1p96.

[4] V. Ljubovic and E. Pajic, “Plagiarism Detection in Computer Programming Using Feature Extraction From Ultra-Fine-Grained Repositories,” IEEE Access, vol. 8, pp. 96505–96514, 2020, doi: 10.1109/ACCESS.2020.2996146.

[5] M. Messer, N. C. C. Brown, M. Kölling, and M. Shi, “Machine Learning-Based Automated Grading and Feedback Tools for Programming: A Meta-Analysis,” in Proceedings of the 2023 Conference on Innovation and Technology in Computer Science Education V. 1, in ITiCSE 2023. New York, NY, USA: Association for Computing Machinery, Jun. 2023, pp. 491–497, doi: 10.1145/3587102.3588822.

[6] N. G. Fonseca, L. Macedo, M. J. Marcelino, and A. J. Mendes, “Augmenting the teacher’s perspective on programming student’s performance via permanent monitoring,” in 2018 IEEE Frontiers in Education Conference (FIE), Oct. 2018, pp. 1–9, doi: 10.1109/FIE.2018.8658924.

[7] S. Nayak, R. Agarwal, and S. K. Khatri, “Automated Assessment Tools for grading of programming Assignments: A review,” in 2022 International Conference on Computer Communication and Informatics (ICCCI), Jan. 2022, pp. 1–4, doi: 10.1109/ICCCI54379.2022.9740769.

[8] Á. Figueira, J. P. Leal, and J. C. Paiva, “Automated Assessment in Computer Science Education: A State-of-the-Art Review,” 2022. [Online]. Available: https://repositorio.inesctec.pt/handle/ 123456789/14301. [Accessed Nov. 29, 2023].

[9] S. Marwan, G. Gao, S. Fisk, T. W. Price, and T. Barnes, “Adaptive Immediate Feedback Can Improve Novice Programming Engagement and Intention to Persist in Computer Science,” in Proceedings of the 2020 ACM Conference on International Computing Education Research, New York, NY, USA: Association for Computing Machinery, Aug. 2020, pp. 194–203, doi: 10.1145/3372782.3406264.

[10] F. M. Van der Kleij, R. C. W. Feskens, and T. J. H. M. Eggen, “Effects of Feedback in a Computer-Based Learning Environment on Students’ Learning Outcomes: A Meta-Analysis,” Rev. Educ. Res., vol. 85, no. 4, pp. 475–511, Dec. 2015, doi: 10.3102/0034654314564881.

[11] L. Vu, “A Case Study of Peer Assessment in a Composition MOOC: Students’ Perceptions and Peer-Grading Scores vs. Instructor-Grading Scores,” in Learning and Performance Assessment: Concepts, Methodologies, Tools, and Applications, IGI Global, 2020, pp. 918–956, doi: 10.4018/978-1-7998-0420-8.ch043.

[12] B. M. Panjaitan, S. A. Rukmono, and R. S. Perdana, “Integration Model for Learning Management Systems, Source Control Management, and Autograders using Service-Oriented Architecture Principle,” in 2021 International Conference on Data and Software Engineering (ICoDSE), Nov. 2021, pp. 1–6, doi: 10.1109/ICoDSE53690.2021.9648450.

[13] “Home | Moodle.org.” [Online]. Available: https://moodle.org/ [Accessed Nov. 28, 2023].

[14] “Never Stop Learning - IT for all.” [Online]. Available: https://courses.it4all.vn/ [Accessed Dec. 04, 2023].




DOI: https://doi.org/10.34238/tnu-jst.9328

Refbacks

  • There are currently no refbacks.
TNU Journal of Science and Technology
Rooms 408, 409 - Administration Building - Thai Nguyen University
Tan Thinh Ward - Thai Nguyen City
Phone: (+84) 208 3840 288 - E-mail: jst@tnu.edu.vn
Based on Open Journal Systems
©2018 All Rights Reserved