Skip to content

Code and results for the paper "Evaluating Explanations for Software Patches Generated by Large Language Models".

Notifications You must be signed in to change notification settings

SOLAR-group/ExplanationsForSoftwarePatches

Repository files navigation

Evaluating Explanations for Software Patches Generated by Large Language Models

In the following, we explain the scripts and datasets used for the experiments of the SSBSE 2023 challenge paper "Evaluating Explanations for Software Patches Generated by Large Language Models":

ARJAe/

This folder contains the ARJAe-generated patches. Patches taken from: https://github.com/yyxhdy/arja/tree/arja-e/

Human/

This folder contains the human-written patches. Patches taken from: https://github.com/rjust/defects4j

responses/

This folder contains the requests and the LLM's responses for all considered patches and runs.

results/CompleteEval.csv

A CSV file containing the results for all patches and runs.

getARJAEDiffs.py and getHumanDiffs.py

Scripts to build long diffs (the complete file). E.g., ARJAe/Chart_5/long_diff.patch.

get_explanations.py

The script needed to run the LLM requests via API.

You need an additional file named open_api_key.py with the following contents to run this script:

api_key = 'Your OpenAI API key'

helper_functions.py

Contains helper functions needed by the get_explanations.py script.

evaluation.py

Generates the data for the results table.

About

Code and results for the paper "Evaluating Explanations for Software Patches Generated by Large Language Models".

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages