Treffer: Understanding Code Quality: A Qualitative Evaluation of LLM-Generated vs. Human-Written Code.
Weitere Informationen
As Large Language Models (LLMs) like GPT and Gemini become increasingly integrated into software development, understanding their capabilities and limitations is essential. This project evaluates the effectiveness of these models in code generation by comparing AIgenerated code to human-written code in C++ and Python. Key software quality metrics—including cyclomatic complexity, lines of code, and space and time complexity—are used to assess the performance, efficiency, and readability of the generated code. The study also examines how prompt complexity, analyzed at two distinct levels, influences the quality of code produced by the models. By highlighting the strengths and weaknesses of LLMs in handling programming tasks of varying difficulty, this research provides valuable insights for developers, researchers, and industry professionals. The findings aim to inform best practices for integrating AI assistance into development workflows, ensuring a balance between automation and human oversight. Ultimately, this work contributes to more efficient and maintainable coding practices in an AI-augmented development landscape. [ABSTRACT FROM AUTHOR]