earticle

논문검색

Human-Machine Interaction Technology (HIT)

Utilizing Generative AI for Test Case Generation : Comparative Analysis and Guidelines

초록

영어

The advancement of generative AI technologies has significantly impacted various domains in software engineering, particularly in automating test case generation. As software systems become increasingly complex, manual test case creation faces limitations in terms of efficiency and coverage. This study analyzes the capabilities and limitations of major generative AI models—ChatGPT, Copilot, and Gemini—in generating software test cases. We focus on evaluating their performance in boundary value analysis, exception handling, and property-based testing. Using the ArrayUtils.indexOf() function from the Apache Commons Lang library as the test subject, we conducted experiments to compare the quality and effectiveness of the test cases generated by each model. Our findings indicate that while generative AI can efficiently produce a substantial number of high-quality test cases, there are instances of incorrect test cases and test codes. To address these issues, we propose guidelines for developers to enhance the reliability and consistency of test case generation using generative AI. Future research will explore the application of these models to more complex software systems and further methods to improve their test generation capabilities.

목차

Abstract
1. Introduction
2. Related Works
3. Generative AI's Test Case Generation Capabilities
3.1 Test Case Generation According to Boundary Value Analysis
3.2 Test Case Generation with Exception Objects
3.3 Property-Based Testing Technique
4. Comparative Analysis
4.1 Code Coverage Measurement
4.2 Mutation Testing
4.3 Analysis Results and Guidelines
5. Conclusion
Acknowledgement
References

저자정보

  • Woochang Shin Professor, Dept. of Computer Science, Seokyeong University, Korea

참고문헌

자료제공 : 네이버학술정보

    함께 이용한 논문

      ※ 원문제공기관과의 협약기간이 종료되어 열람이 제한될 수 있습니다.

      0개의 논문이 장바구니에 담겼습니다.