Please use this identifier to cite or link to this item: http://dr.iiserpune.ac.in:8080/xmlui/handle/123456789/9958
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorKumar, Sudhir-
dc.contributor.authorDESHPANDE, BHUSHAN-
dc.date.accessioned2025-05-19T04:31:56Z-
dc.date.available2025-05-19T04:31:56Z-
dc.date.issued2025-05-
dc.identifier.citation112en_US
dc.identifier.urihttp://dr.iiserpune.ac.in:8080/xmlui/handle/123456789/9958-
dc.description.abstractLarge Language Models (LLMs) have been shown to capture the syntax, semantics, and structure of programming languages, enabling the generation of accurate code for similar test cases through Few Shot Learning (FSL) and prompt engineering. Although LLMs perform exceptionally well with small context-length inputs, they struggle to produce accurate results with large context-length inputs and out-of-distribution-dataset, thereby limiting their applicability for large-scale code generation tasks. Our work focuses on large code generation based on the custom dataset using LLMs. This research explored a number of small-sized, open-source, state-of-the-art LLMs and various configurations of LLMs, with temperature and other hyper-parameter settings. With pre-trained LLMs, code generation accuracy was less than 20% without tuning any hyper-parameters. By implementing Retrieval-Augmented Generation (RAG) to retrieve contextually relevant examples, the initial accuracy of the generated code was improved, reaching 65% to 70% correctness based on expert evaluations. A framework for reviewing the generated code called ‘LLM Judge’, was developed to identify correctness, issues, and places of improvement. By iteratively generating and refining code based on feedback from the ‘LLM Judge’, the accuracy of the generated code improved to 75%–80% at the end of the second iteration. These results highlight the potential of LLM to automate the test code generation. This work reduces the time required to write custom code to automate test cases from, on average, two days to a few hours, thereby simplifying the development process for engineers.en_US
dc.language.isoenen_US
dc.subjectRetrieval Augmented Generation (RAG)en_US
dc.subjectCode Generationen_US
dc.subjectCode Evaluationen_US
dc.subjectLarge Language Models (LLMs)en_US
dc.titleRetrieval-Augmented Large Code Generation and Evaluation using Large Language Modelsen_US
dc.typeThesisen_US
dc.description.embargoNo Embargoen_US
dc.type.degreeBS-MSen_US
dc.contributor.departmentDept. of Data Scienceen_US
dc.contributor.registration20201224en_US
Appears in Collections:MS THESES

Files in This Item:
File Description SizeFormat 
20201224_Bhushan_Deshpande_MS_Thesis.pdfMS Thesis3.08 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.