Rajat Singh (rajat.singh@cse.iitd.ac.in), Srikanta Bedathur (srikanta@cse.iitd.ac.in)
We use Python 3.9.18 to run each file. We use Pytorch '2.2.1' with cuda '12.1.105'.
For the dataset, please fill out this Google Form: https://forms.gle/cJiRNeGbvcTDYTyt6, and we will share the dataset link with you afterwards.
The 'result' folder contains three folders, each containing code for each dataset. The code in these folders is arranged according to the prompting pipeline given in the paper. To get the result, you must run the '.ipynb' files in this folder; before running these .ipynb files, please make the following changes:
- Update the API keys for Gemini-Pro and GPT in the code (.ipynb files).
- Download the relevant models (specifically CodeLLama-7B and CodeLLama-13B) from Hugging Face (https://huggingface.co).
- Update all the paths in the config file.
The 'finetune' folder contains all the files required for fine-tuning. The dataset we used to finetune is in the 'finetune/dataset' folder, and the script used to fine-tune the model is in the 'finetune\script' folder. Before running the finetuning code ('finetune.ipynb'), please make the following changes:
- Download the relevant models (specifically CodeLLama-13B) from Hugging Face (https://huggingface.co).
- Update all the paths in 'finetune.ipynb' file.