Similar Items: Robustness evaluation and enhancement of LLMs in code generation: an empirical study
- Do code LLMs do static analysis?
- How challenging it is to identify real code authors: an empirical study
- Prompt engineering in LLMs for automated unit test generation: A large-scale study
- ArkTS code generation: A comprehensive evaluation with large language models
- An evaluation study of large language models for addressing code quality issues
- Performance analysis of AI-generated code: A case study of Copilot, Copilot Chat, CodeLlaMa, and DeepSeek-Coder models