Please use this identifier to cite or link to this item:
http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/15215
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kumar, Dhruv | - |
dc.date.accessioned | 2024-08-12T11:23:15Z | - |
dc.date.available | 2024-08-12T11:23:15Z | - |
dc.date.issued | 2023-12 | - |
dc.identifier.uri | https://arxiv.org/abs/2312.10622 | - |
dc.identifier.uri | http://dspace.bits-pilani.ac.in:8080/jspui/xmlui/handle/123456789/15215 | - |
dc.description.abstract | Generating unit tests is a crucial task in software development, demanding substantial time and effort from programmers. The advent of Large Language Models (LLMs) introduces a novel avenue for unit test script generation. This research aims to experimentally investigate the effectiveness of LLMs, specifically exemplified by ChatGPT, for generating unit test scripts for Python programs, and how the generated test cases compare with those generated by an existing unit test generator (Pynguin). For experiments, we consider three types of code units: 1) Procedural scripts, 2) Function-based modular code, and 3) Class-based code. The generated test cases are evaluated based on criteria such as coverage, correctness, and readability. Our results show that ChatGPT's performance is comparable with Pynguin in terms of coverage, though for some cases its performance is superior to Pynguin. We also find that about a third of assertions generated by ChatGPT for some categories were incorrect. Our results also show that there is minimal overlap in missed statements between ChatGPT and Pynguin, thus, suggesting that a combination of both tools may enhance unit test generation performance. Finally, in our experiments, prompt engineering improved ChatGPT's performance, achieving a much higher coverage. | en_US |
dc.language.iso | en | en_US |
dc.subject | Computer Science | en_US |
dc.subject | Autogeneration Tools | en_US |
dc.subject | Large Language Models (LLMs) | en_US |
dc.subject | ChatGPT(3.5) | en_US |
dc.title | Unit Test Generation using Generative AI : A Comparative Performance Analysis of Autogeneration Tools | en_US |
dc.type | Preprint | en_US |
Appears in Collections: | Department of Computer Science and Information Systems |
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.