This repository contains the dataset employed for the experimental evaluation of the nine DL models, including:
-
3 Fine-tuned encoder-decoder models
- CodeBERT
- CodeT5+
- PLBart
-
3 Fine-tuned decoder-only models
- CodeGen
- CodeGPT
- CodeParrot
-
3 Zero-shot prompted, instruction-tuned LLMs
- DeepSeek-Coder-6.7b
- Qwen2.5-Coder-7b
- StableCode-3b
The repository also contains the script designed to run inference using the 3 instruction-tuned LLMs. We adopted the 4-bit quantized version of these model, prompting them via the llama-cli
interface.
- To run inference, you need
llama-cli
installed and properly configured. - Ensure you have the required model files, e.g, deepseek-coder-6.7b-instruct.Q4_K_M.gguf.
- Adjust the script configuration:
- Uncomment the desired model configuration block in the script.
- Update the
MODEL_PATH
,NGL
, andMAX_TOKENS
variables as needed.
Run the script from the command line:
python run_inference.py