Datasets:

Modalities:
Text
Formats:
parquet
ArXiv:
Tags:
code
License:
Dataset Viewer
Auto-converted to Parquet Duplicate
repo
stringlengths
9
49
instance_id
stringlengths
16
55
base_commit
stringlengths
40
40
patch
stringlengths
485
4.69k
test_patch
stringlengths
519
22.5k
problem_statement
stringlengths
123
10.8k
hints_text
stringlengths
0
38.3k
created_at
stringdate
2025-01-03 00:43:09
2025-01-31 23:26:11
version
stringlengths
3
20
meta
dict
install_config
dict
FAIL_TO_PASS
listlengths
1
12
PASS_TO_PASS
listlengths
0
2.23k
environment_setup_commit
stringlengths
40
40
docker_image
stringlengths
55
94
image_name
stringlengths
55
94
interface
stringclasses
1 value
tobymao/sqlglot
tobymao__sqlglot-4661
bae0489044a1368556f03f637c171a1873b6f05c
diff --git a/sqlglot/dialects/presto.py b/sqlglot/dialects/presto.py index de9e0a2c..9e42c4c5 100644 --- a/sqlglot/dialects/presto.py +++ b/sqlglot/dialects/presto.py @@ -392,7 +392,6 @@ class Presto(Dialect): exp.Encode: lambda self, e: encode_decode_sql(self, e, "TO_UTF8"), exp.FileFormatPro...
diff --git a/tests/dialects/test_presto.py b/tests/dialects/test_presto.py index c84f0003..9443b4e6 100644 --- a/tests/dialects/test_presto.py +++ b/tests/dialects/test_presto.py @@ -1083,6 +1083,9 @@ class TestPresto(Validator): "snowflake": "CURRENT_USER()", }, ) + self.v...
Trino first_value and last_value being incorrectly converted on sql generation - Read dialect = Trino - Write dialect = Trino - Still exists on main FIRST_VALUE and LAST_VALUE window functions are being converted to ARBITRARY. I suspect because of this [function](https://github.com/tobymao/sqlglot/blob/v26.3.8/sqlglot...
2025-01-27 08:58:17
26.3
{ "commit_name": "merge_commit", "failed_lite_validators": [ "has_many_modified_files" ], "has_test_patch": true, "is_lite": false, "num_modified_files": 2 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e .[dev]", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": "pytest", "pip_packages": [ "pytest" ], "pre_install": null, "python": "3.9", "reqs_path": null, "test_cmd": "pytest --no-header -rA --tb=line --col...
[ "tests/dialects/test_presto.py::TestPresto::test_presto", "tests/dialects/test_trino.py::TestTrino::test_ddl" ]
[ "tests/dialects/test_presto.py::TestPresto::test_analyze", "tests/dialects/test_presto.py::TestPresto::test_cast", "tests/dialects/test_presto.py::TestPresto::test_ddl", "tests/dialects/test_presto.py::TestPresto::test_encode_decode", "tests/dialects/test_presto.py::TestPresto::test_hex_unhex", "tests/dia...
1904b7605a7308608ac64e5cfb3c8424d3e55c17
swerebench/sweb.eval.x86_64.tobymao_1776_sqlglot-4661:latest
swerebench/sweb.eval.x86_64.tobymao_1776_sqlglot-4661:latest
Blaizzy/mlx-vlm
Blaizzy__mlx-vlm-179
1e4579ad0c99b41c71f7309a4cb59d0c4e49fca1
diff --git a/mlx_vlm/models/qwen2_vl/vision.py b/mlx_vlm/models/qwen2_vl/vision.py index bd699a4..bd32514 100644 --- a/mlx_vlm/models/qwen2_vl/vision.py +++ b/mlx_vlm/models/qwen2_vl/vision.py @@ -88,7 +88,7 @@ class VisionRotaryEmbedding(nn.Module): inv_freq = 1.0 / ( self.theta ** (mx.arange(0, ...
diff --git a/mlx_vlm/tests/test_trainer.py b/mlx_vlm/tests/test_trainer.py index 97339e7..70dd370 100644 --- a/mlx_vlm/tests/test_trainer.py +++ b/mlx_vlm/tests/test_trainer.py @@ -47,15 +47,15 @@ class TestDataset(unittest.TestCase): mock_get_prompt.return_value = "Mocked prompt" - mock_prepare_inp...
LoRa fine-tuning is not working with Llama 3.2 vision model. After I pull the current main branch, lora fine-tuning of the Llama-3.2-11B-Vision-Instruct stopped working with the following error. ``` python3 -m mlx_vlm.lora --dataset /myspace/datasets --model-path /myspace/Llama-3.2-11B-Vision-Instruct --epochs 5 ...
2025-01-11 21:48:06
0.1
{ "commit_name": "merge_commit", "failed_lite_validators": [ "has_many_modified_files" ], "has_test_patch": true, "is_lite": false, "num_modified_files": 2 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e .", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": "requirements.txt", "pip_packages": [ "pytest" ], "pre_install": [ "apt-get update", "apt-get install -y gcc" ], "python": "3.10", "reqs_path": [...
[ "mlx_vlm/tests/test_trainer.py::TestDataset::test_dataset_getitem" ]
[ "mlx_vlm/tests/test_trainer.py::TestDataset::test_dataset_initialization", "mlx_vlm/tests/test_trainer.py::TestTrainer::test_loss_fn", "mlx_vlm/tests/test_trainer.py::TestTrainer::test_train_step", "mlx_vlm/tests/test_trainer.py::TestTrainer::test_trainer_initialization" ]
7dc05f95c13f8912323648f7d8e5979a3140620a
swerebench/sweb.eval.x86_64.blaizzy_1776_mlx-vlm-179:latest
swerebench/sweb.eval.x86_64.blaizzy_1776_mlx-vlm-179:latest
pydata/xarray
pydata__xarray-9974
609412d8544217247ddf2f72f988da1b38ef01bc
diff --git a/doc/whats-new.rst b/doc/whats-new.rst index 17af655c..9b40a323 100644 --- a/doc/whats-new.rst +++ b/doc/whats-new.rst @@ -71,6 +71,8 @@ Bug fixes By `Kai Mühlbauer <https://github.com/kmuehlbauer>`_. - Use zarr-fixture to prevent thread leakage errors (:pull:`9967`). By `Kai Mühlbauer <https://githu...
diff --git a/xarray/tests/test_dataset.py b/xarray/tests/test_dataset.py index 8a90a05a..f3867bd6 100644 --- a/xarray/tests/test_dataset.py +++ b/xarray/tests/test_dataset.py @@ -6685,11 +6685,15 @@ class TestDataset: assert len(out.data_vars) == 0 def test_polyfit_weighted(self) -> None: - # Mak...
Weighted polyfit is broken for arrays with more than two dimensions ### What happened? I get an error when trying to use the keyword argument w with arrays that have more than two dimensions. This worked in older versions of xarray. The issue was introduced in xarray v2024.11.0. ### What did you expect to happen? _N...
max-sixty: I'll merge, since the failures look unrelated
2025-01-22 15:16:55
2025.01
{ "commit_name": "merge_commit", "failed_lite_validators": [ "has_hyperlinks", "has_many_modified_files" ], "has_test_patch": true, "is_lite": false, "num_modified_files": 2 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e .[complete]", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": "requirements.txt", "pip_packages": [ "pytest", "pytest-cov", "pytest-xdist", "pytest-env", "pytest-timeout" ], "pre_install": null, ...
[ "xarray/tests/test_dataset.py::TestDataset::test_polyfit_weighted" ]
[ "xarray/tests/test_dataset.py::TestDataset::test_repr", "xarray/tests/test_dataset.py::TestDataset::test_repr_multiindex", "xarray/tests/test_dataset.py::TestDataset::test_repr_period_index", "xarray/tests/test_dataset.py::TestDataset::test_unicode_data", "xarray/tests/test_dataset.py::TestDataset::test_rep...
bd927827836cb331015fd54040337bc6aaa2310f
swerebench/sweb.eval.x86_64.pydata_1776_xarray-9974:latest
swerebench/sweb.eval.x86_64.pydata_1776_xarray-9974:latest
probabl-ai/skore
probabl-ai__skore-1133
e6b5de827c7de2ef717c2824382c88b526bbd728
diff --git a/skore/src/skore/sklearn/find_ml_task.py b/skore/src/skore/sklearn/find_ml_task.py index 20d652a..0af490d 100644 --- a/skore/src/skore/sklearn/find_ml_task.py +++ b/skore/src/skore/sklearn/find_ml_task.py @@ -8,9 +8,19 @@ from skore.externals._sklearn_compat import is_clusterer from skore.sklearn.types imp...
diff --git a/skore/tests/unit/sklearn/test_utils.py b/skore/tests/unit/sklearn/test_utils.py index 9df684e..631d645 100644 --- a/skore/tests/unit/sklearn/test_utils.py +++ b/skore/tests/unit/sklearn/test_utils.py @@ -1,3 +1,4 @@ +import numpy import pytest from sklearn.cluster import KMeans from sklearn.datasets imp...
feat(train_test_split): Refine detection of ML task ### Is your feature request related to a problem? Please describe. @koaning got the following error when running `skore.train_test_split`: > /workspaces/joblib-modal/.venv/lib/python3.12/site-packages/skore/sklearn/train_test_split/train_test_split.py:165: HighClass...
github-actions[bot]: <!-- Pytest Coverage Comment: display-backend-coverage --> <a href="https://github.com/probabl-ai/skore/blob/7d115ad74e9990a285b40b7c1ba1034e5400542f/README.md"><img alt="Coverage" src="https://img.shields.io/badge/Coverage-93%25-brightgreen.svg" /></a><details><summary>Coverage Report for backend ...
2025-01-16 10:36:41
0.5
{ "commit_name": "merge_commit", "failed_lite_validators": [ "has_many_hunks" ], "has_test_patch": true, "is_lite": false, "num_modified_files": 1 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e './skore[test,sphinx]'", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": "pytest", "pip_packages": [ "pytest", "pytest-cov", "pytest-xdist", "pytest-mock", "pytest-asyncio" ], "pre_install": null...
[ "skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target8-regression]", "skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target7-regression]", "skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target5-regression]" ]
[ "skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_with_estimator[True-X0-y0-estimator0-binary-classification]", "skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target0-binary-classification]", "skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[targ...
4bd8f2fba221dad77e37a2bbffa6c8201ace9f51
swerebench/sweb.eval.x86_64.probabl-ai_1776_skore-1133:latest
swerebench/sweb.eval.x86_64.probabl-ai_1776_skore-1133:latest
brightway-lca/brightway2-io
brightway-lca__brightway2-io-294
774b44635cbfcc833dd0a40e803ef66e0b204630
diff --git a/bw2io/strategies/ecospold2.py b/bw2io/strategies/ecospold2.py index 4a4375d..f506849 100644 --- a/bw2io/strategies/ecospold2.py +++ b/bw2io/strategies/ecospold2.py @@ -1177,6 +1177,8 @@ def add_cpc_classification_from_single_reference_product(db): "classifications": [], } + T...
diff --git a/tests/strategies/ecospold2.py b/tests/strategies/ecospold2.py index 22d05c4..7e27bc5 100644 --- a/tests/strategies/ecospold2.py +++ b/tests/strategies/ecospold2.py @@ -1,4 +1,4 @@ -from stats_arrays import * +from stats_arrays import LognormalUncertainty, UndefinedUncertainty from bw2io.strategies.ecosp...
Extraction of CPC & HS2017 classifications fail The fix for #282 does not seem to work. See the example below: ### setup - OS: Ubuntu 24.04.1 LTS - Python: 3.12.8 - bw2data version: (4, 3) - bw2io version: 0.9.3 ### description of the issue - CPC classification is truncated and only the first digit seems ex...
2025-01-08 18:01:30
0.9
{ "commit_name": "head_commit", "failed_lite_validators": [], "has_test_patch": true, "is_lite": true, "num_modified_files": 1 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e .[dev]", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": null, "pip_packages": [ "pytest" ], "pre_install": null, "python": "3.9", "reqs_path": null, "test_cmd": "pytest --no-header -rA --tb=line --color=n...
[ "tests/strategies/ecospold2.py::test_add_cpc_classification_from_single_reference_product_single_classif" ]
[ "tests/strategies/ecospold2.py::test_drop_temporary_outdated_biosphere_flows", "tests/strategies/ecospold2.py::test_reparametrize_lognormal_to_agree_with_static_amount", "tests/strategies/ecospold2.py::test_fix_unreasonably_high_lognormal_uncertainties", "tests/strategies/ecospold2.py::test_remove_uncertainty...
839c10b85851eefe8bfe077ce38f1c1ec4d3c937
swerebench/sweb.eval.x86_64.brightway-lca_1776_brightway2-io-294:latest
swerebench/sweb.eval.x86_64.brightway-lca_1776_brightway2-io-294:latest
sphinx-doc/sphinx
sphinx-doc__sphinx-13261
019a6661f4b4ae6bffcfe015387e90d8b0190c8b
diff --git a/CHANGES.rst b/CHANGES.rst index 139ce579a..6271505bc 100644 --- a/CHANGES.rst +++ b/CHANGES.rst @@ -92,6 +92,8 @@ Bugs fixed methods and attributes. Patch by Bénédikt Tran. * #12975: Avoid rendering a trailing comma in C and C++ multi-line signatures. +* #13178: autodoc: Fix resolution for ``pathlib...
diff --git a/tests/test_extensions/test_ext_autodoc_configs.py b/tests/test_extensions/test_ext_autodoc_configs.py index 348e84b2b..02ef26b30 100644 --- a/tests/test_extensions/test_ext_autodoc_configs.py +++ b/tests/test_extensions/test_ext_autodoc_configs.py @@ -696,11 +696,6 @@ def test_mocked_module_imports(app): ...
References to `pathlib` objects are incorrectly picked up by autodoc ### Describe the bug Type annotations using `pathlib.Path` of callable that are handed by autodoc generate the following warning on Python 3.13: ``` {filename}.py:docstring of {callable}:1: WARNING: py:class reference target not found: pathlib....
2025-01-23 02:57:43
8.1
{ "commit_name": "merge_commit", "failed_lite_validators": [ "has_hyperlinks", "has_many_modified_files" ], "has_test_patch": true, "is_lite": false, "num_modified_files": 2 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e .[docs,test]", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": "requirements.txt", "pip_packages": [ "pytest", "pytest-cov" ], "pre_install": [ "apt-get update", "apt-get install -y graphviz" ], ...
[ "tests/test_util/test_util_typing.py::test_is_invalid_builtin_class" ]
[ "tests/test_extensions/test_ext_autodoc_configs.py::test_autoclass_content_class", "tests/test_extensions/test_ext_autodoc_configs.py::test_autoclass_content_init", "tests/test_extensions/test_ext_autodoc_configs.py::test_autodoc_class_signature_mixed", "tests/test_extensions/test_ext_autodoc_configs.py::test...
b678d17e002aec62c54acc176123d9a2d4be8409
swerebench/sweb.eval.x86_64.sphinx-doc_1776_sphinx-13261:latest
swerebench/sweb.eval.x86_64.sphinx-doc_1776_sphinx-13261:latest
vyperlang/vyper
vyperlang__vyper-4462
9e396fb8de9e0c8c72b9500365ade44435ef4bef
diff --git a/vyper/codegen/core.py b/vyper/codegen/core.py index aaf6f350..789cc775 100644 --- a/vyper/codegen/core.py +++ b/vyper/codegen/core.py @@ -560,7 +560,7 @@ def _get_element_ptr_array(parent, key, array_bounds_check): return IRnode.from_list("~empty", subtype) if parent.value == "multi": - ...
diff --git a/tests/functional/codegen/features/iteration/test_for_in_list.py b/tests/functional/codegen/features/iteration/test_for_in_list.py index 036e7c06..4b9eea3c 100644 --- a/tests/functional/codegen/features/iteration/test_for_in_list.py +++ b/tests/functional/codegen/features/iteration/test_for_in_list.py @@ -2...
panic when iterating over static arrays ### Version Information * vyper Version (output of `vyper --version`): 762eec682450a6631342fad06898471ea2a6aa90 ### What's your issue about? following contract panics when compiling: ```vyper X: constant(uint24[4]) = [1,2,3,4] @external def foo(): x: uint24 = 0 for s: ...
2025-01-22 00:27:39
0.4
{ "commit_name": "merge_commit", "failed_lite_validators": [ "has_git_commit_hash", "has_many_modified_files" ], "has_test_patch": true, "is_lite": false, "num_modified_files": 2 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e .[dev]", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": "pytest", "pip_packages": [ "pytest" ], "pre_install": null, "python": "3.10", "reqs_path": null, "test_cmd": "pytest --no-header -rA --tb=line --co...
[ "tests/functional/codegen/features/iteration/test_for_in_list.py::test_constant_list_iter" ]
[ "tests/functional/codegen/features/iteration/test_for_in_list.py::test_bad_code[17", "tests/functional/codegen/features/iteration/test_for_in_list.py::test_iterator_modification_module_attribute", "tests/functional/codegen/features/iteration/test_for_in_list.py::test_bad_code[19", "tests/functional/codegen/fe...
9e396fb8de9e0c8c72b9500365ade44435ef4bef
swerebench/sweb.eval.x86_64.vyperlang_1776_vyper-4462:latest
swerebench/sweb.eval.x86_64.vyperlang_1776_vyper-4462:latest
CrossGL/crosstl
CrossGL__crosstl-257
bf894dbc90466d1ed9ae942f44396fbd04f18298
diff --git a/crosstl/translator/lexer.py b/crosstl/translator/lexer.py index 84cad4f..c260361 100644 --- a/crosstl/translator/lexer.py +++ b/crosstl/translator/lexer.py @@ -21,6 +21,7 @@ TOKENS = OrderedDict( ("FRAGMENT", r"\bfragment\b"), ("FLOAT_NUMBER", r"\d*\.\d+|\d+\.\d*"), ("FLOAT", r"\...
diff --git a/tests/test_translator/test_lexer.py b/tests/test_translator/test_lexer.py index f4793b0..5164c59 100644 --- a/tests/test_translator/test_lexer.py +++ b/tests/test_translator/test_lexer.py @@ -152,6 +152,7 @@ def test_data_types_tokenization(): float c; double d; bool e; + unsigned int f; ...
ToDO for translator frontend - this task is to add support for `translator` frontend add token in `lexer` so it can also identify these tokens. for this task the one file `crosstl/src/translator/lexer.py` List of task - [x] #42 - [x] #43 - [x] #44 - [x] #45 - [x] #46 - [x] #47 - [x] #48 - [x] #4...
github-actions[bot]: This pull request has been marked as stale because it has been inactive for more than 14 days. Please update this pull request or it will be automatically closed in 7 days.
2025-01-06 15:20:49
0.0
{ "commit_name": "merge_commit", "failed_lite_validators": [ "has_hyperlinks", "has_issue_reference" ], "has_test_patch": true, "is_lite": false, "num_modified_files": 1 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e .", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": "requirements.txt", "pip_packages": [ "pytest" ], "pre_install": [ "apt-get update", "apt-get install -y gcc" ], "python": "3.9", "reqs_path": [ ...
[ "tests/test_translator/test_lexer.py::test_data_types_tokenization" ]
[ "tests/test_translator/test_lexer.py::test_struct_tokenization", "tests/test_translator/test_lexer.py::test_if_statement_tokenization", "tests/test_translator/test_lexer.py::test_for_statement_tokenization", "tests/test_translator/test_lexer.py::test_else_statement_tokenization", "tests/test_translator/test...
36bed5871a8d102f73cfebf82c8d8495aaa89e87
swerebench/sweb.eval.x86_64.crossgl_1776_crosstl-257:latest
swerebench/sweb.eval.x86_64.crossgl_1776_crosstl-257:latest
sympy/sympy
sympy__sympy-27512
7f0b42c29807be66eb3bc181c818a431f211ffd4
diff --git a/sympy/assumptions/handlers/ntheory.py b/sympy/assumptions/handlers/ntheory.py index 38cedecd49..cfe63ba646 100644 --- a/sympy/assumptions/handlers/ntheory.py +++ b/sympy/assumptions/handlers/ntheory.py @@ -63,7 +63,14 @@ def _(expr, assumptions): return _PrimePredicate_number(expr, assumptions) ...
diff --git a/sympy/assumptions/tests/test_query.py b/sympy/assumptions/tests/test_query.py index 4bf385d857..f6da573b13 100644 --- a/sympy/assumptions/tests/test_query.py +++ b/sympy/assumptions/tests/test_query.py @@ -1790,7 +1790,21 @@ def test_prime(): assert ask(Q.prime(x**2), Q.integer(x)) is False ass...
ask(Q.prime(x ** y)) wrongly returns False when x and y are positive integers ```python >>>from sympy import Symbol, ask, Q >>>x = Symbol('x', positive=True, integer=True) >>>y = Symbol('y', positive=True, integer=True) >>>ask(Q.prime(x ** y)) False ``` ~~1 is a positive integer, so 1^1 is a counterexample.~~ 2^1 = 2 ...
sympy-bot: Hi, I am the [SymPy bot](https://github.com/sympy/sympy-bot). I'm here to help you write a release notes entry. Please read the [guide on how to write release notes](https://github.com/sympy/sympy/wiki/Writing-Release-Notes). :x: There was an issue with the release notes. **Please do not close this pull r...
2025-01-26 08:51:38
1.13
{ "commit_name": "head_commit", "failed_lite_validators": [ "has_issue_reference", "has_many_modified_files", "has_many_hunks" ], "has_test_patch": true, "is_lite": false, "num_modified_files": 3 }
{ "env_vars": null, "env_yml_path": null, "install": "pip install -e .[dev]", "log_parser": "parse_log_pytest", "no_use_env": null, "packages": "requirements.txt", "pip_packages": [ "pytest" ], "pre_install": [ "apt-get update", "apt-get install -y gcc" ], "python": "3.9", "reqs_path...
[ "sympy/assumptions/tests/test_query.py::test_prime", "sympy/stats/tests/test_discrete_rv.py::test_negative_binomial", "sympy/stats/tests/test_discrete_rv.py::test_moment_generating_functions" ]
[ "sympy/assumptions/tests/test_query.py::test_int_1", "sympy/assumptions/tests/test_query.py::test_int_11", "sympy/assumptions/tests/test_query.py::test_int_12", "sympy/assumptions/tests/test_query.py::test_float_1", "sympy/assumptions/tests/test_query.py::test_zero_0", "sympy/assumptions/tests/test_query....
490c6d0f7df7b02a70e04c0ed4ead899cd430e70
swerebench/sweb.eval.x86_64.sympy_1776_sympy-27512:latest
swerebench/sweb.eval.x86_64.sympy_1776_sympy-27512:latest
probabl-ai/skore
probabl-ai__skore-1229
c976c4ba555a599a17dcb48b3f3a44d9e6366583
"diff --git a/skore/src/skore/sklearn/find_ml_task.py b/skore/src/skore/sklearn/find_ml_task.py\nind(...TRUNCATED)
"diff --git a/skore/tests/unit/sklearn/test_utils.py b/skore/tests/unit/sklearn/test_utils.py\nindex(...TRUNCATED)
"refactor: Change output of `_find_ml_task` from \"unsupported\" to \"unknown\"\nAs discussed with @(...TRUNCATED)
"github-actions[bot]: <!-- Pytest Coverage Comment: display-backend-coverage -->\n<a href=\"https://(...TRUNCATED)
2025-01-27 10:46:39
0.6
{"commit_name":"merge_commit","failed_lite_validators":["has_many_modified_files","has_many_hunks"],(...TRUNCATED)
{"env_vars":null,"env_yml_path":null,"install":"pip install -e './skore[test,sphinx]'","log_parser":(...TRUNCATED)
["skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target4-unknown]","sko(...TRUNCATED)
["skore/tests/unit/utils/test_accessors.py::test_register_accessor","skore/tests/unit/utils/test_acc(...TRUNCATED)
035e4f65d914d492ac75c387c7b227fb1ed9e1c2
swerebench/sweb.eval.x86_64.probabl-ai_1776_skore-1229:latest
swerebench/sweb.eval.x86_64.probabl-ai_1776_skore-1229:latest
End of preview. Expand in Data Studio

Dataset Summary

SWE-rebench-leaderboard is a continuously updated, curated subset of the full SWE-rebench corpus, tailored for benchmarking software engineering agents on real-world tasks. These tasks are used in the SWE-rebench leaderboard. For more details on the benchmark methodology and data collection process, please refer to our paper SWE-rebench: An Automated Pipeline for Task Collection and Decontaminated Evaluation of Software Engineering Agents.

All Docker images required to run the tasks are pre-built and publicly available on Docker Hub. You do not need to build them yourself. The specific image for each task is listed in the docker_image column.

To get the exact subset of tasks used for a specific month's SWE-rebench-leaderboard, you can filter the dataset by the created_at field.

News

[2025/09/19] Added a split for each month.
[2025/09/01] Added 52 August tasks, each with a corresponding Docker image.
[2025/08/04] Added 34 July tasks, each with a corresponding Docker image.

How to Use

from datasets import load_dataset
ds = load_dataset('nebius/SWE-rebench-leaderboard')
ds_june_2025 = ds['test'].filter(lambda x: x['created_at'].startswith('2025-06'))

Dataset Structure

The SWE-rebench dataset schema extends the original SWE-bench schema with additional fields to support richer analysis. The complete schema is detailed in the table below. For more information about this data and methodology behind collecting it, please refer to our paper.

Field name Type Description
instance_id str A formatted instance identifier, usually as repo_owner__repo_name-PR-number.
patch str The gold patch, the patch generated by the PR (minus test-related code), that resolved the issue.
repo str The repository owner/name identifier from GitHub.
base_commit str The commit hash of the repository representing the HEAD of the repository before the solution PR is applied.
hints_text str Comments made on the issue prior to the creation of the solution PR’s first commit creation date.
created_at str The creation date of the pull request.
test_patch str A test-file patch that was contributed by the solution PR.
problem_statement str The issue title and body.
version str Installation version to use for running evaluation.
environment_setup_commit str Commit hash to use for environment setup and installation.
FAIL_TO_PASS str A JSON list of strings that represent the set of tests resolved by the PR and tied to the issue resolution.
PASS_TO_PASS str A JSON list of strings that represent tests that should pass before and after the PR application.
meta str A JSON dictionary indicating whether the instance is lite, along with a list of failed lite validators if it is not.
license_name str The type of license of the repository.
install_config str Installation configuration for setting up the repository.
docker_image str Docker image name for the instance.

To execute tasks from SWE-rebench (i.e., set up their environments, apply patches, and run tests), we provide a fork of the original SWE-bench execution framework, adapted for our dataset's structure and features. The primary modification introduces functionality to source environment installation constants directly from the install_config field present in each task instance within SWE-rebench. This allows for more flexible and task-specific environment setups.

You can find the details of this modification in the following commit

To build the necessary Docker images and run agents on SWE-rebench tasks, you have two main options:

  1. Use our SWE-bench fork directly: Clone the fork and utilize its scripts for building images and executing tasks. The framework will automatically use the install_config from each task.
  2. Integrate similar functionality into your existing codebase: If you have your own execution framework based on SWE-bench or a different system, you can adapt it by implementing a similar mechanism to parse and utilize the install_config field from the SWE-rebench task instances. The aforementioned commit can serve as a reference for this integration.

License

The dataset is licensed under the Creative Commons Attribution 4.0 license. However, please respect the license of each specific repository on which a particular instance is based. To facilitate this, the license of each repository at the time of the commit is provided for every instance.

Citation

@misc{badertdinov2025swerebenchautomatedpipelinetask,
      title={SWE-rebench: An Automated Pipeline for Task Collection and Decontaminated Evaluation of Software Engineering Agents}, 
      author={Ibragim Badertdinov and Alexander Golubev and Maksim Nekrashevich and Anton Shevtsov and Simon Karasik and Andrei Andriushchenko and Maria Trofimova and Daria Litvintseva and Boris Yangel},
      year={2025},
      eprint={2505.20411},
      archivePrefix={arXiv},
      primaryClass={cs.SE},
      url={https://arxiv.org/abs/2505.20411}
}
Downloads last month
5,313

Paper for nebius/SWE-rebench-leaderboard