repo stringlengths 9 49 | instance_id stringlengths 16 55 | base_commit stringlengths 40 40 | patch stringlengths 485 4.69k | test_patch stringlengths 519 22.5k | problem_statement stringlengths 123 10.8k | hints_text stringlengths 0 38.3k | created_at stringdate 2025-01-03 00:43:09 2025-01-31 23:26:11 | version stringlengths 3 20 | meta dict | install_config dict | FAIL_TO_PASS listlengths 1 12 | PASS_TO_PASS listlengths 0 2.23k | environment_setup_commit stringlengths 40 40 | docker_image stringlengths 55 94 | image_name stringlengths 55 94 | interface stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
tobymao/sqlglot | tobymao__sqlglot-4661 | bae0489044a1368556f03f637c171a1873b6f05c | diff --git a/sqlglot/dialects/presto.py b/sqlglot/dialects/presto.py
index de9e0a2c..9e42c4c5 100644
--- a/sqlglot/dialects/presto.py
+++ b/sqlglot/dialects/presto.py
@@ -392,7 +392,6 @@ class Presto(Dialect):
exp.Encode: lambda self, e: encode_decode_sql(self, e, "TO_UTF8"),
exp.FileFormatPro... | diff --git a/tests/dialects/test_presto.py b/tests/dialects/test_presto.py
index c84f0003..9443b4e6 100644
--- a/tests/dialects/test_presto.py
+++ b/tests/dialects/test_presto.py
@@ -1083,6 +1083,9 @@ class TestPresto(Validator):
"snowflake": "CURRENT_USER()",
},
)
+ self.v... | Trino first_value and last_value being incorrectly converted on sql generation
- Read dialect = Trino
- Write dialect = Trino
- Still exists on main
FIRST_VALUE and LAST_VALUE window functions are being converted to ARBITRARY. I suspect because of this [function](https://github.com/tobymao/sqlglot/blob/v26.3.8/sqlglot... | 2025-01-27 08:58:17 | 26.3 | {
"commit_name": "merge_commit",
"failed_lite_validators": [
"has_many_modified_files"
],
"has_test_patch": true,
"is_lite": false,
"num_modified_files": 2
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e .[dev]",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": "pytest",
"pip_packages": [
"pytest"
],
"pre_install": null,
"python": "3.9",
"reqs_path": null,
"test_cmd": "pytest --no-header -rA --tb=line --col... | [
"tests/dialects/test_presto.py::TestPresto::test_presto",
"tests/dialects/test_trino.py::TestTrino::test_ddl"
] | [
"tests/dialects/test_presto.py::TestPresto::test_analyze",
"tests/dialects/test_presto.py::TestPresto::test_cast",
"tests/dialects/test_presto.py::TestPresto::test_ddl",
"tests/dialects/test_presto.py::TestPresto::test_encode_decode",
"tests/dialects/test_presto.py::TestPresto::test_hex_unhex",
"tests/dia... | 1904b7605a7308608ac64e5cfb3c8424d3e55c17 | swerebench/sweb.eval.x86_64.tobymao_1776_sqlglot-4661:latest | swerebench/sweb.eval.x86_64.tobymao_1776_sqlglot-4661:latest | ||
Blaizzy/mlx-vlm | Blaizzy__mlx-vlm-179 | 1e4579ad0c99b41c71f7309a4cb59d0c4e49fca1 | diff --git a/mlx_vlm/models/qwen2_vl/vision.py b/mlx_vlm/models/qwen2_vl/vision.py
index bd699a4..bd32514 100644
--- a/mlx_vlm/models/qwen2_vl/vision.py
+++ b/mlx_vlm/models/qwen2_vl/vision.py
@@ -88,7 +88,7 @@ class VisionRotaryEmbedding(nn.Module):
inv_freq = 1.0 / (
self.theta ** (mx.arange(0, ... | diff --git a/mlx_vlm/tests/test_trainer.py b/mlx_vlm/tests/test_trainer.py
index 97339e7..70dd370 100644
--- a/mlx_vlm/tests/test_trainer.py
+++ b/mlx_vlm/tests/test_trainer.py
@@ -47,15 +47,15 @@ class TestDataset(unittest.TestCase):
mock_get_prompt.return_value = "Mocked prompt"
- mock_prepare_inp... | LoRa fine-tuning is not working with Llama 3.2 vision model.
After I pull the current main branch, lora fine-tuning of the Llama-3.2-11B-Vision-Instruct stopped working with the following error.
```
python3 -m mlx_vlm.lora --dataset /myspace/datasets --model-path /myspace/Llama-3.2-11B-Vision-Instruct --epochs 5
... | 2025-01-11 21:48:06 | 0.1 | {
"commit_name": "merge_commit",
"failed_lite_validators": [
"has_many_modified_files"
],
"has_test_patch": true,
"is_lite": false,
"num_modified_files": 2
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e .",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": "requirements.txt",
"pip_packages": [
"pytest"
],
"pre_install": [
"apt-get update",
"apt-get install -y gcc"
],
"python": "3.10",
"reqs_path": [... | [
"mlx_vlm/tests/test_trainer.py::TestDataset::test_dataset_getitem"
] | [
"mlx_vlm/tests/test_trainer.py::TestDataset::test_dataset_initialization",
"mlx_vlm/tests/test_trainer.py::TestTrainer::test_loss_fn",
"mlx_vlm/tests/test_trainer.py::TestTrainer::test_train_step",
"mlx_vlm/tests/test_trainer.py::TestTrainer::test_trainer_initialization"
] | 7dc05f95c13f8912323648f7d8e5979a3140620a | swerebench/sweb.eval.x86_64.blaizzy_1776_mlx-vlm-179:latest | swerebench/sweb.eval.x86_64.blaizzy_1776_mlx-vlm-179:latest | ||
pydata/xarray | pydata__xarray-9974 | 609412d8544217247ddf2f72f988da1b38ef01bc | diff --git a/doc/whats-new.rst b/doc/whats-new.rst
index 17af655c..9b40a323 100644
--- a/doc/whats-new.rst
+++ b/doc/whats-new.rst
@@ -71,6 +71,8 @@ Bug fixes
By `Kai Mühlbauer <https://github.com/kmuehlbauer>`_.
- Use zarr-fixture to prevent thread leakage errors (:pull:`9967`).
By `Kai Mühlbauer <https://githu... | diff --git a/xarray/tests/test_dataset.py b/xarray/tests/test_dataset.py
index 8a90a05a..f3867bd6 100644
--- a/xarray/tests/test_dataset.py
+++ b/xarray/tests/test_dataset.py
@@ -6685,11 +6685,15 @@ class TestDataset:
assert len(out.data_vars) == 0
def test_polyfit_weighted(self) -> None:
- # Mak... | Weighted polyfit is broken for arrays with more than two dimensions
### What happened?
I get an error when trying to use the keyword argument w with arrays that have more than two dimensions. This worked in older versions of xarray. The issue was introduced in xarray v2024.11.0.
### What did you expect to happen?
_N... | max-sixty: I'll merge, since the failures look unrelated | 2025-01-22 15:16:55 | 2025.01 | {
"commit_name": "merge_commit",
"failed_lite_validators": [
"has_hyperlinks",
"has_many_modified_files"
],
"has_test_patch": true,
"is_lite": false,
"num_modified_files": 2
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e .[complete]",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": "requirements.txt",
"pip_packages": [
"pytest",
"pytest-cov",
"pytest-xdist",
"pytest-env",
"pytest-timeout"
],
"pre_install": null,
... | [
"xarray/tests/test_dataset.py::TestDataset::test_polyfit_weighted"
] | [
"xarray/tests/test_dataset.py::TestDataset::test_repr",
"xarray/tests/test_dataset.py::TestDataset::test_repr_multiindex",
"xarray/tests/test_dataset.py::TestDataset::test_repr_period_index",
"xarray/tests/test_dataset.py::TestDataset::test_unicode_data",
"xarray/tests/test_dataset.py::TestDataset::test_rep... | bd927827836cb331015fd54040337bc6aaa2310f | swerebench/sweb.eval.x86_64.pydata_1776_xarray-9974:latest | swerebench/sweb.eval.x86_64.pydata_1776_xarray-9974:latest | |
probabl-ai/skore | probabl-ai__skore-1133 | e6b5de827c7de2ef717c2824382c88b526bbd728 | diff --git a/skore/src/skore/sklearn/find_ml_task.py b/skore/src/skore/sklearn/find_ml_task.py
index 20d652a..0af490d 100644
--- a/skore/src/skore/sklearn/find_ml_task.py
+++ b/skore/src/skore/sklearn/find_ml_task.py
@@ -8,9 +8,19 @@ from skore.externals._sklearn_compat import is_clusterer
from skore.sklearn.types imp... | diff --git a/skore/tests/unit/sklearn/test_utils.py b/skore/tests/unit/sklearn/test_utils.py
index 9df684e..631d645 100644
--- a/skore/tests/unit/sklearn/test_utils.py
+++ b/skore/tests/unit/sklearn/test_utils.py
@@ -1,3 +1,4 @@
+import numpy
import pytest
from sklearn.cluster import KMeans
from sklearn.datasets imp... | feat(train_test_split): Refine detection of ML task
### Is your feature request related to a problem? Please describe.
@koaning got the following error when running `skore.train_test_split`:
> /workspaces/joblib-modal/.venv/lib/python3.12/site-packages/skore/sklearn/train_test_split/train_test_split.py:165: HighClass... | github-actions[bot]: <!-- Pytest Coverage Comment: display-backend-coverage -->
<a href="https://github.com/probabl-ai/skore/blob/7d115ad74e9990a285b40b7c1ba1034e5400542f/README.md"><img alt="Coverage" src="https://img.shields.io/badge/Coverage-93%25-brightgreen.svg" /></a><details><summary>Coverage Report for backend ... | 2025-01-16 10:36:41 | 0.5 | {
"commit_name": "merge_commit",
"failed_lite_validators": [
"has_many_hunks"
],
"has_test_patch": true,
"is_lite": false,
"num_modified_files": 1
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e './skore[test,sphinx]'",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": "pytest",
"pip_packages": [
"pytest",
"pytest-cov",
"pytest-xdist",
"pytest-mock",
"pytest-asyncio"
],
"pre_install": null... | [
"skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target8-regression]",
"skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target7-regression]",
"skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target5-regression]"
] | [
"skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_with_estimator[True-X0-y0-estimator0-binary-classification]",
"skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target0-binary-classification]",
"skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[targ... | 4bd8f2fba221dad77e37a2bbffa6c8201ace9f51 | swerebench/sweb.eval.x86_64.probabl-ai_1776_skore-1133:latest | swerebench/sweb.eval.x86_64.probabl-ai_1776_skore-1133:latest | |
brightway-lca/brightway2-io | brightway-lca__brightway2-io-294 | 774b44635cbfcc833dd0a40e803ef66e0b204630 | diff --git a/bw2io/strategies/ecospold2.py b/bw2io/strategies/ecospold2.py
index 4a4375d..f506849 100644
--- a/bw2io/strategies/ecospold2.py
+++ b/bw2io/strategies/ecospold2.py
@@ -1177,6 +1177,8 @@ def add_cpc_classification_from_single_reference_product(db):
"classifications": [],
}
+ T... | diff --git a/tests/strategies/ecospold2.py b/tests/strategies/ecospold2.py
index 22d05c4..7e27bc5 100644
--- a/tests/strategies/ecospold2.py
+++ b/tests/strategies/ecospold2.py
@@ -1,4 +1,4 @@
-from stats_arrays import *
+from stats_arrays import LognormalUncertainty, UndefinedUncertainty
from bw2io.strategies.ecosp... | Extraction of CPC & HS2017 classifications fail
The fix for #282 does not seem to work. See the example below:
### setup
- OS: Ubuntu 24.04.1 LTS
- Python: 3.12.8
- bw2data version: (4, 3)
- bw2io version: 0.9.3
### description of the issue
- CPC classification is truncated and only the first digit seems ex... | 2025-01-08 18:01:30 | 0.9 | {
"commit_name": "head_commit",
"failed_lite_validators": [],
"has_test_patch": true,
"is_lite": true,
"num_modified_files": 1
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e .[dev]",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": null,
"pip_packages": [
"pytest"
],
"pre_install": null,
"python": "3.9",
"reqs_path": null,
"test_cmd": "pytest --no-header -rA --tb=line --color=n... | [
"tests/strategies/ecospold2.py::test_add_cpc_classification_from_single_reference_product_single_classif"
] | [
"tests/strategies/ecospold2.py::test_drop_temporary_outdated_biosphere_flows",
"tests/strategies/ecospold2.py::test_reparametrize_lognormal_to_agree_with_static_amount",
"tests/strategies/ecospold2.py::test_fix_unreasonably_high_lognormal_uncertainties",
"tests/strategies/ecospold2.py::test_remove_uncertainty... | 839c10b85851eefe8bfe077ce38f1c1ec4d3c937 | swerebench/sweb.eval.x86_64.brightway-lca_1776_brightway2-io-294:latest | swerebench/sweb.eval.x86_64.brightway-lca_1776_brightway2-io-294:latest | ||
sphinx-doc/sphinx | sphinx-doc__sphinx-13261 | 019a6661f4b4ae6bffcfe015387e90d8b0190c8b | diff --git a/CHANGES.rst b/CHANGES.rst
index 139ce579a..6271505bc 100644
--- a/CHANGES.rst
+++ b/CHANGES.rst
@@ -92,6 +92,8 @@ Bugs fixed
methods and attributes.
Patch by Bénédikt Tran.
* #12975: Avoid rendering a trailing comma in C and C++ multi-line signatures.
+* #13178: autodoc: Fix resolution for ``pathlib... | diff --git a/tests/test_extensions/test_ext_autodoc_configs.py b/tests/test_extensions/test_ext_autodoc_configs.py
index 348e84b2b..02ef26b30 100644
--- a/tests/test_extensions/test_ext_autodoc_configs.py
+++ b/tests/test_extensions/test_ext_autodoc_configs.py
@@ -696,11 +696,6 @@ def test_mocked_module_imports(app):
... | References to `pathlib` objects are incorrectly picked up by autodoc
### Describe the bug
Type annotations using `pathlib.Path` of callable that are handed by autodoc generate the following warning on Python 3.13:
```
{filename}.py:docstring of {callable}:1: WARNING: py:class reference target not found: pathlib.... | 2025-01-23 02:57:43 | 8.1 | {
"commit_name": "merge_commit",
"failed_lite_validators": [
"has_hyperlinks",
"has_many_modified_files"
],
"has_test_patch": true,
"is_lite": false,
"num_modified_files": 2
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e .[docs,test]",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": "requirements.txt",
"pip_packages": [
"pytest",
"pytest-cov"
],
"pre_install": [
"apt-get update",
"apt-get install -y graphviz"
],
... | [
"tests/test_util/test_util_typing.py::test_is_invalid_builtin_class"
] | [
"tests/test_extensions/test_ext_autodoc_configs.py::test_autoclass_content_class",
"tests/test_extensions/test_ext_autodoc_configs.py::test_autoclass_content_init",
"tests/test_extensions/test_ext_autodoc_configs.py::test_autodoc_class_signature_mixed",
"tests/test_extensions/test_ext_autodoc_configs.py::test... | b678d17e002aec62c54acc176123d9a2d4be8409 | swerebench/sweb.eval.x86_64.sphinx-doc_1776_sphinx-13261:latest | swerebench/sweb.eval.x86_64.sphinx-doc_1776_sphinx-13261:latest | ||
vyperlang/vyper | vyperlang__vyper-4462 | 9e396fb8de9e0c8c72b9500365ade44435ef4bef | diff --git a/vyper/codegen/core.py b/vyper/codegen/core.py
index aaf6f350..789cc775 100644
--- a/vyper/codegen/core.py
+++ b/vyper/codegen/core.py
@@ -560,7 +560,7 @@ def _get_element_ptr_array(parent, key, array_bounds_check):
return IRnode.from_list("~empty", subtype)
if parent.value == "multi":
- ... | diff --git a/tests/functional/codegen/features/iteration/test_for_in_list.py b/tests/functional/codegen/features/iteration/test_for_in_list.py
index 036e7c06..4b9eea3c 100644
--- a/tests/functional/codegen/features/iteration/test_for_in_list.py
+++ b/tests/functional/codegen/features/iteration/test_for_in_list.py
@@ -2... | panic when iterating over static arrays
### Version Information
* vyper Version (output of `vyper --version`): 762eec682450a6631342fad06898471ea2a6aa90
### What's your issue about?
following contract panics when compiling:
```vyper
X: constant(uint24[4]) = [1,2,3,4]
@external
def foo():
x: uint24 = 0
for s: ... | 2025-01-22 00:27:39 | 0.4 | {
"commit_name": "merge_commit",
"failed_lite_validators": [
"has_git_commit_hash",
"has_many_modified_files"
],
"has_test_patch": true,
"is_lite": false,
"num_modified_files": 2
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e .[dev]",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": "pytest",
"pip_packages": [
"pytest"
],
"pre_install": null,
"python": "3.10",
"reqs_path": null,
"test_cmd": "pytest --no-header -rA --tb=line --co... | [
"tests/functional/codegen/features/iteration/test_for_in_list.py::test_constant_list_iter"
] | [
"tests/functional/codegen/features/iteration/test_for_in_list.py::test_bad_code[17",
"tests/functional/codegen/features/iteration/test_for_in_list.py::test_iterator_modification_module_attribute",
"tests/functional/codegen/features/iteration/test_for_in_list.py::test_bad_code[19",
"tests/functional/codegen/fe... | 9e396fb8de9e0c8c72b9500365ade44435ef4bef | swerebench/sweb.eval.x86_64.vyperlang_1776_vyper-4462:latest | swerebench/sweb.eval.x86_64.vyperlang_1776_vyper-4462:latest | ||
CrossGL/crosstl | CrossGL__crosstl-257 | bf894dbc90466d1ed9ae942f44396fbd04f18298 | diff --git a/crosstl/translator/lexer.py b/crosstl/translator/lexer.py
index 84cad4f..c260361 100644
--- a/crosstl/translator/lexer.py
+++ b/crosstl/translator/lexer.py
@@ -21,6 +21,7 @@ TOKENS = OrderedDict(
("FRAGMENT", r"\bfragment\b"),
("FLOAT_NUMBER", r"\d*\.\d+|\d+\.\d*"),
("FLOAT", r"\... | diff --git a/tests/test_translator/test_lexer.py b/tests/test_translator/test_lexer.py
index f4793b0..5164c59 100644
--- a/tests/test_translator/test_lexer.py
+++ b/tests/test_translator/test_lexer.py
@@ -152,6 +152,7 @@ def test_data_types_tokenization():
float c;
double d;
bool e;
+ unsigned int f;
... | ToDO for translator frontend
- this task is to add support for `translator` frontend add token in `lexer` so it can also identify these tokens. for this task the one file `crosstl/src/translator/lexer.py`
List of task
- [x] #42
- [x] #43
- [x] #44
- [x] #45
- [x] #46
- [x] #47
- [x] #48
- [x] #4... | github-actions[bot]: This pull request has been marked as stale because it has been inactive for more than 14 days. Please update this pull request or it will be automatically closed in 7 days. | 2025-01-06 15:20:49 | 0.0 | {
"commit_name": "merge_commit",
"failed_lite_validators": [
"has_hyperlinks",
"has_issue_reference"
],
"has_test_patch": true,
"is_lite": false,
"num_modified_files": 1
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e .",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": "requirements.txt",
"pip_packages": [
"pytest"
],
"pre_install": [
"apt-get update",
"apt-get install -y gcc"
],
"python": "3.9",
"reqs_path": [
... | [
"tests/test_translator/test_lexer.py::test_data_types_tokenization"
] | [
"tests/test_translator/test_lexer.py::test_struct_tokenization",
"tests/test_translator/test_lexer.py::test_if_statement_tokenization",
"tests/test_translator/test_lexer.py::test_for_statement_tokenization",
"tests/test_translator/test_lexer.py::test_else_statement_tokenization",
"tests/test_translator/test... | 36bed5871a8d102f73cfebf82c8d8495aaa89e87 | swerebench/sweb.eval.x86_64.crossgl_1776_crosstl-257:latest | swerebench/sweb.eval.x86_64.crossgl_1776_crosstl-257:latest | |
sympy/sympy | sympy__sympy-27512 | 7f0b42c29807be66eb3bc181c818a431f211ffd4 | diff --git a/sympy/assumptions/handlers/ntheory.py b/sympy/assumptions/handlers/ntheory.py
index 38cedecd49..cfe63ba646 100644
--- a/sympy/assumptions/handlers/ntheory.py
+++ b/sympy/assumptions/handlers/ntheory.py
@@ -63,7 +63,14 @@ def _(expr, assumptions):
return _PrimePredicate_number(expr, assumptions)
... | diff --git a/sympy/assumptions/tests/test_query.py b/sympy/assumptions/tests/test_query.py
index 4bf385d857..f6da573b13 100644
--- a/sympy/assumptions/tests/test_query.py
+++ b/sympy/assumptions/tests/test_query.py
@@ -1790,7 +1790,21 @@ def test_prime():
assert ask(Q.prime(x**2), Q.integer(x)) is False
ass... | ask(Q.prime(x ** y)) wrongly returns False when x and y are positive integers
```python
>>>from sympy import Symbol, ask, Q
>>>x = Symbol('x', positive=True, integer=True)
>>>y = Symbol('y', positive=True, integer=True)
>>>ask(Q.prime(x ** y))
False
```
~~1 is a positive integer, so 1^1 is a counterexample.~~
2^1 = 2 ... | sympy-bot:
Hi, I am the [SymPy bot](https://github.com/sympy/sympy-bot). I'm here to help you write a release notes entry. Please read the [guide on how to write release notes](https://github.com/sympy/sympy/wiki/Writing-Release-Notes).
:x: There was an issue with the release notes. **Please do not close this pull r... | 2025-01-26 08:51:38 | 1.13 | {
"commit_name": "head_commit",
"failed_lite_validators": [
"has_issue_reference",
"has_many_modified_files",
"has_many_hunks"
],
"has_test_patch": true,
"is_lite": false,
"num_modified_files": 3
} | {
"env_vars": null,
"env_yml_path": null,
"install": "pip install -e .[dev]",
"log_parser": "parse_log_pytest",
"no_use_env": null,
"packages": "requirements.txt",
"pip_packages": [
"pytest"
],
"pre_install": [
"apt-get update",
"apt-get install -y gcc"
],
"python": "3.9",
"reqs_path... | [
"sympy/assumptions/tests/test_query.py::test_prime",
"sympy/stats/tests/test_discrete_rv.py::test_negative_binomial",
"sympy/stats/tests/test_discrete_rv.py::test_moment_generating_functions"
] | [
"sympy/assumptions/tests/test_query.py::test_int_1",
"sympy/assumptions/tests/test_query.py::test_int_11",
"sympy/assumptions/tests/test_query.py::test_int_12",
"sympy/assumptions/tests/test_query.py::test_float_1",
"sympy/assumptions/tests/test_query.py::test_zero_0",
"sympy/assumptions/tests/test_query.... | 490c6d0f7df7b02a70e04c0ed4ead899cd430e70 | swerebench/sweb.eval.x86_64.sympy_1776_sympy-27512:latest | swerebench/sweb.eval.x86_64.sympy_1776_sympy-27512:latest | |
probabl-ai/skore | probabl-ai__skore-1229 | c976c4ba555a599a17dcb48b3f3a44d9e6366583 | "diff --git a/skore/src/skore/sklearn/find_ml_task.py b/skore/src/skore/sklearn/find_ml_task.py\nind(...TRUNCATED) | "diff --git a/skore/tests/unit/sklearn/test_utils.py b/skore/tests/unit/sklearn/test_utils.py\nindex(...TRUNCATED) | "refactor: Change output of `_find_ml_task` from \"unsupported\" to \"unknown\"\nAs discussed with @(...TRUNCATED) | "github-actions[bot]: <!-- Pytest Coverage Comment: display-backend-coverage -->\n<a href=\"https://(...TRUNCATED) | 2025-01-27 10:46:39 | 0.6 | {"commit_name":"merge_commit","failed_lite_validators":["has_many_modified_files","has_many_hunks"],(...TRUNCATED) | {"env_vars":null,"env_yml_path":null,"install":"pip install -e './skore[test,sphinx]'","log_parser":(...TRUNCATED) | ["skore/tests/unit/sklearn/test_utils.py::test_find_ml_task_without_estimator[target4-unknown]","sko(...TRUNCATED) | ["skore/tests/unit/utils/test_accessors.py::test_register_accessor","skore/tests/unit/utils/test_acc(...TRUNCATED) | 035e4f65d914d492ac75c387c7b227fb1ed9e1c2 | swerebench/sweb.eval.x86_64.probabl-ai_1776_skore-1229:latest | swerebench/sweb.eval.x86_64.probabl-ai_1776_skore-1229:latest |
Dataset Summary
SWE-rebench-leaderboard is a continuously updated, curated subset of the full SWE-rebench corpus, tailored for benchmarking software engineering agents on real-world tasks. These tasks are used in the SWE-rebench leaderboard. For more details on the benchmark methodology and data collection process, please refer to our paper SWE-rebench: An Automated Pipeline for Task Collection and Decontaminated Evaluation of Software Engineering Agents.
All Docker images required to run the tasks are pre-built and publicly available on Docker Hub. You do not need to build them yourself. The specific image for each task is listed in the docker_image column.
To get the exact subset of tasks used for a specific month's SWE-rebench-leaderboard, you can filter the dataset by the created_at field.
News
[2025/09/19] Added a split for each month.
[2025/09/01] Added 52 August tasks, each with a corresponding Docker image.
[2025/08/04] Added 34 July tasks, each with a corresponding Docker image.
How to Use
from datasets import load_dataset
ds = load_dataset('nebius/SWE-rebench-leaderboard')
ds_june_2025 = ds['test'].filter(lambda x: x['created_at'].startswith('2025-06'))
Dataset Structure
The SWE-rebench dataset schema extends the original SWE-bench schema with additional fields to support richer analysis. The complete schema is detailed in the table below. For more information about this data and methodology behind collecting it, please refer to our paper.
| Field name | Type | Description |
|---|---|---|
instance_id |
str | A formatted instance identifier, usually as repo_owner__repo_name-PR-number. |
patch |
str | The gold patch, the patch generated by the PR (minus test-related code), that resolved the issue. |
repo |
str | The repository owner/name identifier from GitHub. |
base_commit |
str | The commit hash of the repository representing the HEAD of the repository before the solution PR is applied. |
hints_text |
str | Comments made on the issue prior to the creation of the solution PR’s first commit creation date. |
created_at |
str | The creation date of the pull request. |
test_patch |
str | A test-file patch that was contributed by the solution PR. |
problem_statement |
str | The issue title and body. |
version |
str | Installation version to use for running evaluation. |
environment_setup_commit |
str | Commit hash to use for environment setup and installation. |
FAIL_TO_PASS |
str | A JSON list of strings that represent the set of tests resolved by the PR and tied to the issue resolution. |
PASS_TO_PASS |
str | A JSON list of strings that represent tests that should pass before and after the PR application. |
meta |
str | A JSON dictionary indicating whether the instance is lite, along with a list of failed lite validators if it is not. |
license_name |
str | The type of license of the repository. |
install_config |
str | Installation configuration for setting up the repository. |
docker_image |
str | Docker image name for the instance. |
To execute tasks from SWE-rebench (i.e., set up their environments, apply patches, and run tests), we provide a fork of the original SWE-bench execution framework, adapted for our dataset's structure and features.
The primary modification introduces functionality to source environment installation constants directly from the install_config field present in each task instance within SWE-rebench. This allows for more flexible and task-specific environment setups.
You can find the details of this modification in the following commit
To build the necessary Docker images and run agents on SWE-rebench tasks, you have two main options:
- Use our SWE-bench fork directly: Clone the fork and utilize its scripts for building images and executing tasks. The framework will automatically use the
install_configfrom each task. - Integrate similar functionality into your existing codebase: If you have your own execution framework based on SWE-bench or a different system, you can adapt it by implementing a similar mechanism to parse and utilize the
install_configfield from the SWE-rebench task instances. The aforementioned commit can serve as a reference for this integration.
License
The dataset is licensed under the Creative Commons Attribution 4.0 license. However, please respect the license of each specific repository on which a particular instance is based. To facilitate this, the license of each repository at the time of the commit is provided for every instance.
Citation
@misc{badertdinov2025swerebenchautomatedpipelinetask,
title={SWE-rebench: An Automated Pipeline for Task Collection and Decontaminated Evaluation of Software Engineering Agents},
author={Ibragim Badertdinov and Alexander Golubev and Maksim Nekrashevich and Anton Shevtsov and Simon Karasik and Andrei Andriushchenko and Maria Trofimova and Daria Litvintseva and Boris Yangel},
year={2025},
eprint={2505.20411},
archivePrefix={arXiv},
primaryClass={cs.SE},
url={https://arxiv.org/abs/2505.20411}
}
- Downloads last month
- 5,313