Analysis of the ability of artificial intelligence tools to generate student-style code and detect such generated code

Keywords: tools with artificial intelligence, Copilot, ChatGPT, Gemini, programming, AI generated code

Abstract

It is known that artificial intelligence tools (AIT) can generate program code in different styles: the general style, the styles of other AITs, the teacher's style, the style of a professional developer, etc. But if a first-year student provides code for a task that was generated using an AIT in some such professional style then the teacher can easily determine the non-independency of performing such work. The article studies the possibility of AITs to generate code in the "first-year student" style and the ability of AITs to determine that this code was generated using an AIT, and was not actually written by a first-year student. The aim of the article is to investigate how easily a first-year student can mask the use of AITs when writing programs in C language for educational tasks in the disciplines of the first-year study. This study was conducted for the code generated in the student's style by the latest free versions of three AITs, which are currently among the most popular and most used by students: ChatGPT, Copilot, and Gemini. As a result of the study, it was found that the AITs ChatGPT, Copilot can mask their code as the student's code quite well, while the AIT ​​Gemini did not cope with this task. In any case, this applies to programs written in C language for tasks of initial complexity of the first year of study. The conclusions summarize the results of the study and make recommendations to teachers regarding the use of the studied AITs. The following are noted as areas for further research. Since the power of AI is developing explosively quickly, it seems appropriate to repeat the research of this article again in a year or two. Another direction can be proposed to conduct a similar study using other programming languages, as well as when performing tasks in other disciplines and/or of greater complexity

References

1. Dayu Yang, Tianyang Liu, Daoan Zhang, Antoine Simoulin, Xiaoyi Liu, Yuwei Cao, Zhaopu Teng, Xin Qian, Grey Yang, Jiebo Luo, and Julian McAuley, “ Code to Think, Think to Code: A Survey on Code-Enhanced Reasoning and Reasoning-Driven Code Intelligence in LLMs”, arXiv:2502.19411v1 [cs.CL] 26 Feb 2025.
2. Jiang, Fan Wang, Jiasi Shen, Sungju Kim, and Sunghun Kim, “A Survey on Large Language Models for Code Generation”, arXiv:2406.00515v2 [cs.CL] 10 Nov 2024. Available: https://arxiv.org/pdf/2406.00515. Accessed on: May 15, 2025.
3. Janek Bevendorff, Matti Wiegmann, Jussi Karlgren, Luise Dürlich, Evangelia Gogoulou, Aarne Talman, Efstathios Stamatatos, Martin Potthast, and Benno Stein, “Overview of the “Voight-Kampg” Generative AI Authorship Verification Task at PAN and ELOQUENT 2024”, Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2024), Grenoble, France, 9-12 September, 2024. pp. 2486-2506
4. Jijie Huang, Yang Chen, Man Luo, and Yonglan Li, “Generative AI Authorship Verification Of Tri-Sentence Analysis Base On The Bert Model”, Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2024), Grenoble, France, 9-12 September, 2024. pp. 2632-2637
5. Xuan Ji, “Detecting ai-generated code in C++: A comprehensive guide”, BytePlus, Apr 25, 2025.

Abstract views: 18
PDF Downloads: 7
Published
2025-06-16
How to Cite
Marchenko , O., & Marchenko, O. (2025). Analysis of the ability of artificial intelligence tools to generate student-style code and detect such generated code. COMPUTER-INTEGRATED TECHNOLOGIES: EDUCATION, SCIENCE, PRODUCTION, (59), 183-193. https://doi.org/10.36910/6775-2524-0560-2025-59-24
Section
Computer science and computer engineering