Datasets:

Modalities:
Text
Formats:
parquet
Libraries:
Datasets
pandas
License:
FEA-Bench / README.md
CharonBony's picture
Upload README.md with huggingface_hub
55ee2f7 verified
metadata
license: other
license_name: other
license_link: LICENSE
dataset_info:
  features:
    - name: instance_id
      dtype: string
    - name: pull_number
      dtype: int64
    - name: repo
      dtype: string
    - name: version
      dtype: string
    - name: base_commit
      dtype: string
    - name: environment_setup_commit
      dtype: string
    - name: created_at
      dtype: string
    - name: FAIL_TO_PASS
      sequence: string
    - name: PASS_TO_PASS
      sequence: string
  splits:
    - name: test
      num_bytes: 9959442
      num_examples: 1401
  download_size: 1227364
  dataset_size: 9959442
configs:
  - config_name: default
    data_files:
      - split: test
        path: data/test-*

Dataset Card for FEA-Bench

A Benchmark for Evaluating Repository-Level Code Generation for Feature Implementation.

Dataset Details

Dataset Description

The FEA-Bench is a benchmark with a test set that contains 1,401 task instances from 83 Github repositories. This benchmark aims to evaluate the capabilities of repository-level incremental code development. The task instances are collected from Github pull requests, which have the purpose of new feature implementation. Each task instance includes the repo and the base commit sha256, and the PR number and the status of unit test.

  • Curated by: the authors of the FEA-Bench paper: Wei Li, Xin Zhang, Zhongxin Guo, Shaoguang Mao and their collaborators.
  • Language(s) (NLP): English
  • License: Others; We list all licenses of involved github repositories in the last part.

Uses

This dataset is designed to evaluate performances of LLMs on repository-level code development, which is a complicated software engineering task.

  • Repository-level incremental code development: The FEA-Bench can be used to evaluate a model for the the capabilities of repository-level incremental code development. Success on this task is typically measured by achieving a high/low resolved ratio. The leaderboard will soon be published as a website.

Direct Use

Use scripts from FEA-Bench repo to get info for task instances and organize them into prompt, which can be used to LLMs' inference. Also, you can get info or use agents to directly solve the PRs with code changes.

Out-of-Scope Use

This dataset is not aimed at training for LLMs. You should not take the FEA-Bench as the training dataset to avoid contamination.

Dataset Structure

An example:

{
    "instance_id": "huggingface__accelerate-270",
    "pull_number": 270,
    "repo": "huggingface/accelerate",
    "version": null,
    "base_commit": "515fcca9ed2b36c274c595dbdff75f1c2da635de",
    "environment_setup_commit": "08101b9dde2b1a9658c2e363e3e9f5663ba06073",
    "FAIL_TO_PASS": [
        "tests/test_state_checkpointing.py::CheckpointTest::test_can_resume_training",
        "tests/test_state_checkpointing.py::CheckpointTest::test_invalid_registration",
        "tests/test_state_checkpointing.py::CheckpointTest::test_with_scheduler"
    ],
    "PASS_TO_PASS": []
}

Dataset Creation

Curation Rationale

Implementing new features in repository-level codebases is a crucial application of code generation models. However, current benchmarks lack a dedicated evaluation framework for this capability. To fill this gap, we introduce FEA-Bench, a benchmark designed to assess the ability of large language models (LLMs) to perform incremental development within code repositories.

Source Data

Data Collection and Processing

We collect pull requests from 83 GitHub repositories and use rule-based and intent-based filtering to construct task instances focused on new feature development. Each task instance containing code changes is paired with relevant unit test files to ensure that the solution can be verified.

Who are the source data producers?

Authors of 83 Github repositories list in the last part.

Personal and Sensitive Information

The dataset does not include any personal or sensitive information.

Bias, Risks, and Limitations

  • The quantity of high-quality data suitable for repository-level incremental development is limited. High-quality and usable pull requests for new feature development are relatively scarce. Many repository-level code developments for implementing new functionalities were committed during the early stages of repositories, without going through the rigorous code review process typical of the open-source community, resulting in lower data quality that cannot be utilized.
  • Furthermore, the software's early-stage developments might not even have been conducted using the GitHub platform, posing a challenge for data collection and utilization.
  • The repository-level incremental code development may not just include new feature implementation tasks.
  • Only Python repositories are involved in FEA-Bench.
  • The inference results of the task instances from the benchmark may contain code that is harmful to computer systems.

Recommendations

Evaluation by docker is recommended, just like SWE-bench. We will also publish a patch for SWE-bench to make it compatible for our tasks' evaluation.

BibTeX:

To be appeared after publishing the ArXiv paper.

APA:

To be appeared after publishing the ArXiv paper.

Dataset Card Contact

For further information or questions, please contact Xin Zhang (xinzhang3@microsoft.com).

All involved Github repositories in the FEA-Bench

Repo Name License Topic
astropy/astropy BSD-3-Clause Scientific/Engineering::Astronomy
django/django BSD-3-Clause Internet::WWW/HTTP
matplotlib/matplotlib Other Scientific/Engineering::Visualization
mwaskom/seaborn BSD-3-Clause Scientific/Engineering::Visualization
pallets/flask BSD-3-Clause Internet::WWW/HTTP
pvlib/pvlib-python BSD-3-Clause Scientific/Engineering::Physics
pydata/xarray Apache-2.0 Scientific/Engineering::Information Analysis
pydicom/pydicom Others Scientific/Engineering::Medical Science Apps.
pylint-dev/astroid LGPL-2.1 Software Development::Libraries
pylint-dev/pylint GPL-2.0 Software Development::Quality Assurance
pyvista/pyvista MIT Scientific/Engineering::Information Analysis
scikit-learn/scikit-learn BSD-3-Clause Scientific/Engineering::Artificial Intelligence
sphinx-doc/sphinx BSD-2-Clause Text Processing::Markup
sqlfluff/sqlfluff MIT Software Development::Quality Assurance
sympy/sympy Others Scientific/Engineering::Mathematics
Aider-AI/aider Apache-2.0 Software Development::Code Generators
Cog-Creators/Red-DiscordBot GPL-3.0 Communications::Chat
DLR-RM/stable-baselines3 MIT Scientific/Engineering::Artificial Intelligence
EleutherAI/lm-evaluation-harness MIT Scientific/Engineering::Artificial Intelligence
Project-MONAI/MONAI Apache-2.0 Scientific/Engineering::Medical Science Apps.
PyThaiNLP/pythainlp Apache-2.0 Text Processing::Linguistic
RDFLib/rdflib BSD-3-Clause Software Development::Libraries
Textualize/rich MIT Software Development::Libraries
Textualize/textual MIT Software Development::User Interfaces
TileDB-Inc/TileDB-Py MIT Software Development::Libraries
astronomer/astronomer-cosmos Apache-2.0 Software Development::Build Tools
atlassian-api/atlassian-python-api Apache-2.0 Internet::WWW/HTTP
aws-cloudformation/cfn-lint MIT-0 Software Development::Quality Assurance
aws-powertools/powertools-lambda-python MIT-0 Software Development::Libraries
aws/sagemaker-python-sdk Apache-2.0 Scientific/Engineering::Artificial Intelligence
biopragmatics/bioregistry MIT Scientific/Engineering::Bio-Informatics
boto/boto3 Apache-2.0 Software Development::Libraries
boto/botocore Apache-2.0 Software Development::Libraries
cocotb/cocotb BSD-3-Clause Scientific/Engineering::Electronic Design Automation (EDA)
conan-io/conan MIT Software Development::Build Tools
deepset-ai/haystack Apache-2.0 Scientific/Engineering::Artificial Intelligence
docker/docker-py Apache-2.0 Software Development::Libraries
dpkp/kafka-python Apache-2.0 Software Development::Libraries
embeddings-benchmark/mteb Apache-2.0 Scientific/Engineering::Artificial Intelligence
facebookresearch/hydra MIT Software Development::Libraries
fairlearn/fairlearn MIT Scientific/Engineering::Artificial Intelligence
falconry/falcon Apache-2.0 Internet::WWW/HTTP
google-deepmind/optax Apache-2.0 Scientific/Engineering::Artificial Intelligence
googleapis/python-aiplatform Apache-2.0 Scientific/Engineering::Artificial Intelligence
googleapis/python-bigquery Apache-2.0 Internet::WWW/HTTP
gradio-app/gradio Apache-2.0 Scientific/Engineering::Human Machine Interfaces
graphql-python/graphene MIT Software Development::Libraries
huggingface/accelerate Apache-2.0 Scientific/Engineering::Artificial Intelligence
huggingface/datasets Apache-2.0 Scientific/Engineering::Artificial Intelligence
huggingface/huggingface_hub Apache-2.0 Scientific/Engineering::Artificial Intelligence
huggingface/pytorch-image-models Apache-2.0 Software Development::Libraries
huggingface/trl Apache-2.0 Scientific/Engineering::Artificial Intelligence
joblib/joblib BSD-3-Clause Software Development::Libraries
joke2k/faker MIT Software Development::Testing
lark-parser/lark MIT Text Processing::Linguistic
minio/minio-py Apache-2.0 Software Development::Libraries
open-mmlab/mmengine Apache-2.0 Utilities
openvinotoolkit/datumaro MIT Scientific/Engineering::Image Processing
pgmpy/pgmpy MIT Scientific/Engineering::Artificial Intelligence
pre-commit/pre-commit MIT Software Development::Quality Assurance
prometheus/client_python Apache-2.0 System::Monitoring
prompt-toolkit/python-prompt-toolkit BSD-3-Clause Software Development::User Interfaces
pygments/pygments BSD-2-Clause Software Development::Documentation
pyocd/pyOCD Apache-2.0 Software Development::Debuggers
pypa/hatch MIT Software Development::Build Tools
pyro-ppl/pyro Apache-2.0 Scientific/Engineering::Artificial Intelligence
python-hyper/h2 MIT Internet::WWW/HTTP
roboflow/supervision MIT Scientific/Engineering::Image Processing
rytilahti/python-miio GPL-3.0 Home Automation
saleweaver/python-amazon-sp-api MIT Internet::WWW/HTTP
scrapy/scrapy BSD-3-Clause Software Development::Libraries
scverse/scanpy BSD-3-Clause Scientific/Engineering::Bio-Informatics
slackapi/bolt-python MIT Communications::Chat
slackapi/python-slack-sdk MIT Communications::Chat
snowflakedb/snowflake-connector-python Apache-2.0 Software Development::Libraries
softlayer/softlayer-python MIT Software Development::Libraries
spec-first/connexion Apache-2.0 Internet::WWW/HTTP
statsmodels/statsmodels BSD-3-Clause Scientific/Engineering::Information Analysis
tfranzel/drf-spectacular BSD-3-Clause Software Development::Documentation
tobymao/sqlglot MIT Database::Database Engines/Servers
tornadoweb/tornado Apache-2.0 Internet::WWW/HTTP
tortoise/tortoise-orm Apache-2.0 Database::Front-Ends
wagtail/wagtail BSD-3-Clause Internet::WWW/HTTP