Our paper, “An Empirical Study of Code Smells in Transformer-based Code Generation Techniques”, got accepted for the 22nd IEEE International Working Conference on Source Code Analysis and Manipulation (SCAM 2022) in the research track.
We carry out a thorough empirical analysis of code smells in Python transformer-based code generation models’ training sets and look at how these bad patterns get up in the output. In order to carry out this study, we obtained three open-source datasets (CodeXGlue, APPS, and Code Clippy) that are frequently used to train Python code generation techniques and checked to see how much code smells were present in them. We also looked into the possibility of code smells in the code produced by transformer-based models. For this experiment, we calculated the code smells in the outputs produced by the open-source and closed-source code generating tools GPT-Code-Clippy and GitHub Copilot, respectively.
This paper makes three contributions:
Pre-print: SCAM 2022 Source Code: GitHub
Subscribe to this blog via RSS.
Paper 12
Research 12
Tool 2
Llm 9
Dataset 2
Survey 1
"SALLM: Security Assessment of Generated Code" accepted at ASYDE 2024 (ASE Workshop)
Posted on 07 Sep 2024Paper (12) Research (12) Tool (2) Llm (9) Dataset (2) Qualitative-analysis (1) Survey (1)