Skip to content

The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".

License

Notifications You must be signed in to change notification settings

thunlp/DebugBench

Repository files navigation

Repo of DebugBench

icon

Overview

Implementation for paper DebugBench: Evaluating Debugging Capabilities of Large Language Models with datasets, prompts, model outputs.

Benchmark

Please refer to the Hugging Face Dataset for the data source and evaluation script if you want to use the benchmark.

DebugBench is a Large Language Model (LLM) debugging benchmark introduced in the paper "DebugBench: Evaluating Debugging Capability of Large Language Models" [url]. We collect code snippets from the LeetCode community and implant bugs into source data with GPT-4.

  • It consists of 4,253 instances.
  • It covers four major bug categories and 18 minor types.
  • It includes C++, Java, and Python instances.
  • It contains three difficulty levels: easy, medium, and hard.
  • All the instances were released after June 2022.
  • Please refer to the article [url] for more details.

Repo Content

This repository contains the implementation for benchmark construction and evaluation.

  • benchmark directory contains the 51 JSON shards of different languages and bug types of the benchmark.

  • dataset_construction directory contains the implementation for bug implantation to solution code via LLMs.

  • evaluation directory contains the implementation for evaluating the debugging capabilities of LLMs with API.

  • evalution_result directory contains the model output of gpt-4-0613 , gpt-3.5-turbo-0613 and CodeLlama-34b-instruct under different scenarios.

More elements will be added to the repository soon.

Citations

Please cite the paper and star the repo if you use DebugBench and find it helpful.

Feel free to contact [email protected] or open an issue if you have any questions.

@misc{tian2024debugbench,
      title={DebugBench: Evaluating Debugging Capability of Large Language Models}, 
      author={Runchu Tian and Yining Ye and Yujia Qin and Xin Cong and Yankai Lin and Zhiyuan Liu and Maosong Sun},
      year={2024},
      eprint={2401.04621},
      archivePrefix={arXiv},
      primaryClass={cs.SE}
}

About

The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published