
Unit tests serve a critical role in software development in ensuring quality and pre-dictability for components of code. While unit tests are important, they are a resourceheavy part of the development life cycle. Hence, the automation of unit tests is an inten-sive research area. This paper looks at the use of Large Language Models (LLMs) to achievethis task by looking at both fine-tuning and in-context learning of various sizes in order toautomatically generate unit tests given a function, and seeing how they perform in terms ofa similarity score. In order to achieve this, this thesis proposes an implementation pipelinewhich consist of creating a dataset that maps functions to unit tests, training an LLM, Wiz-ardCoder, of various sizes on a GPU-cluster, running inference, and evaluating the qualityof the generated tests. These results that the fine-tuned models consistently outperformthe models using in-context learning, independent of size. Moreover, within the scope offine-tuned models, the models using more parameters performed slightly better than theirsmaller counterparts.
Machine Learning, Unit Tests, LLM, Telekommunikation, Large Language Models, Test generation, Telecommunications, C++, Similarity
Machine Learning, Unit Tests, LLM, Telekommunikation, Large Language Models, Test generation, Telecommunications, C++, Similarity
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
