Update BirdSet.py
Browse files- BirdSet.py +18 -13
BirdSet.py
CHANGED
|
@@ -17,6 +17,7 @@
|
|
| 17 |
import os
|
| 18 |
import datasets
|
| 19 |
import pandas as pd
|
|
|
|
| 20 |
|
| 21 |
from .classes import BIRD_NAMES_NIPS4BPLUS, BIRD_NAMES_AMAZON_BASIN, BIRD_NAMES_HAWAII, \
|
| 22 |
BIRD_NAMES_HIGH_SIERRAS, BIRD_NAMES_SIERRA_NEVADA, BIRD_NAMES_POWDERMILL_NATURE, BIRD_NAMES_SAPSUCKER, \
|
|
@@ -48,6 +49,7 @@ _BIRDSET_DESCRIPTION = """\
|
|
| 48 |
|
| 49 |
base_url = "https://huggingface.co/datasets/DBD-research-group/BirdSet/resolve/data"
|
| 50 |
|
|
|
|
| 51 |
class BirdSetConfig(datasets.BuilderConfig):
|
| 52 |
def __init__(
|
| 53 |
self,
|
|
@@ -58,7 +60,7 @@ class BirdSetConfig(datasets.BuilderConfig):
|
|
| 58 |
species_group_list,
|
| 59 |
order_list,
|
| 60 |
**kwargs):
|
| 61 |
-
super().__init__(version=datasets.Version("0.0.
|
| 62 |
|
| 63 |
features = datasets.Features({
|
| 64 |
"audio": datasets.Audio(sampling_rate=32_000, mono=True, decode=False),
|
|
@@ -437,7 +439,9 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 437 |
"meta_test_5s": os.path.join(self.config.data_dir, f"{ds_name}_metadata_test_5s.parquet"),
|
| 438 |
})
|
| 439 |
|
| 440 |
-
|
|
|
|
|
|
|
| 441 |
|
| 442 |
if self.config.name.startswith("XC") or self.config.name.endswith("_xc"):
|
| 443 |
return [
|
|
@@ -445,7 +449,7 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 445 |
name=datasets.Split.TRAIN,
|
| 446 |
gen_kwargs={
|
| 447 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["train"]],
|
| 448 |
-
"local_audio_archives_paths": local_audio_archives_paths["train"] if local_audio_archives_paths else None,
|
| 449 |
"metapath": dl_dir["metadata"],
|
| 450 |
"split": datasets.Split.TRAIN,
|
| 451 |
},
|
|
@@ -454,7 +458,7 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 454 |
name="valid",
|
| 455 |
gen_kwargs={
|
| 456 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["valid"]],
|
| 457 |
-
"local_audio_archives_paths": local_audio_archives_paths["valid"] if local_audio_archives_paths else None,
|
| 458 |
"metapath": dl_dir["meta_test_5s"],
|
| 459 |
"split": "valid",
|
| 460 |
},
|
|
@@ -467,7 +471,7 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 467 |
name=datasets.Split.TEST,
|
| 468 |
gen_kwargs={
|
| 469 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["test"]],
|
| 470 |
-
"local_audio_archives_paths": local_audio_archives_paths["test"] if local_audio_archives_paths else None,
|
| 471 |
"metapath": dl_dir["metadata"],
|
| 472 |
"split": datasets.Split.TEST,
|
| 473 |
},
|
|
@@ -476,7 +480,7 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 476 |
name="test_5s",
|
| 477 |
gen_kwargs={
|
| 478 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["test_5s"]],
|
| 479 |
-
"local_audio_archives_paths": local_audio_archives_paths["test_5s"] if local_audio_archives_paths else None,
|
| 480 |
"metapath": dl_dir["metadata_5s"],
|
| 481 |
"split": "test_multilabel"
|
| 482 |
},
|
|
@@ -488,7 +492,7 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 488 |
name=datasets.Split.TRAIN,
|
| 489 |
gen_kwargs={
|
| 490 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["train"]],
|
| 491 |
-
"local_audio_archives_paths": local_audio_archives_paths["train"] if local_audio_archives_paths else None,
|
| 492 |
"metapath": dl_dir["meta_train"],
|
| 493 |
"split": datasets.Split.TRAIN,
|
| 494 |
},
|
|
@@ -497,7 +501,7 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 497 |
name=datasets.Split.TEST,
|
| 498 |
gen_kwargs={
|
| 499 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["test"]],
|
| 500 |
-
"local_audio_archives_paths": local_audio_archives_paths["test"] if local_audio_archives_paths else None,
|
| 501 |
"metapath": dl_dir["meta_test"],
|
| 502 |
"split": datasets.Split.TEST,
|
| 503 |
},
|
|
@@ -506,27 +510,26 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 506 |
name="test_5s",
|
| 507 |
gen_kwargs={
|
| 508 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["test_5s"]],
|
| 509 |
-
"local_audio_archives_paths": local_audio_archives_paths["test_5s"] if local_audio_archives_paths else None,
|
| 510 |
"metapath": dl_dir["meta_test_5s"],
|
| 511 |
"split": "test_multilabel"
|
| 512 |
},
|
| 513 |
),
|
| 514 |
]
|
| 515 |
|
| 516 |
-
def _generate_examples(self, audio_archive_iterators,
|
| 517 |
metadata = pd.read_parquet(metapath)
|
| 518 |
idx = 0
|
| 519 |
for i, audio_archive_iterator in enumerate(audio_archive_iterators):
|
| 520 |
for audio_path_in_archive, audio_file in audio_archive_iterator:
|
| 521 |
id = os.path.split(audio_path_in_archive)[-1]
|
| 522 |
rows = metadata[metadata.index == (int(id[2:].split(".")[0]) if split == "train" else id)]
|
| 523 |
-
audio_path =
|
| 524 |
|
| 525 |
-
audio = audio_path if local_audio_archives_paths else audio_file.read()
|
| 526 |
for _, row in rows.iterrows():
|
| 527 |
idx += 1
|
| 528 |
yield id if split == "train" else idx, {
|
| 529 |
-
"audio":
|
| 530 |
"filepath": audio_path,
|
| 531 |
"start_time": row["start_time"],
|
| 532 |
"end_time": row["end_time"],
|
|
@@ -556,3 +559,5 @@ class BirdSet(datasets.GeneratorBasedBuilder):
|
|
| 556 |
"species_group_multilabel": row.get("species_group_multilabel", None),
|
| 557 |
"order_multilabel": row.get("order_multilabel", None),
|
| 558 |
}
|
|
|
|
|
|
|
|
|
| 17 |
import os
|
| 18 |
import datasets
|
| 19 |
import pandas as pd
|
| 20 |
+
import numpy as np
|
| 21 |
|
| 22 |
from .classes import BIRD_NAMES_NIPS4BPLUS, BIRD_NAMES_AMAZON_BASIN, BIRD_NAMES_HAWAII, \
|
| 23 |
BIRD_NAMES_HIGH_SIERRAS, BIRD_NAMES_SIERRA_NEVADA, BIRD_NAMES_POWDERMILL_NATURE, BIRD_NAMES_SAPSUCKER, \
|
|
|
|
| 49 |
|
| 50 |
base_url = "https://huggingface.co/datasets/DBD-research-group/BirdSet/resolve/data"
|
| 51 |
|
| 52 |
+
|
| 53 |
class BirdSetConfig(datasets.BuilderConfig):
|
| 54 |
def __init__(
|
| 55 |
self,
|
|
|
|
| 60 |
species_group_list,
|
| 61 |
order_list,
|
| 62 |
**kwargs):
|
| 63 |
+
super().__init__(version=datasets.Version("0.0.3"), name=name, **kwargs)
|
| 64 |
|
| 65 |
features = datasets.Features({
|
| 66 |
"audio": datasets.Audio(sampling_rate=32_000, mono=True, decode=False),
|
|
|
|
| 439 |
"meta_test_5s": os.path.join(self.config.data_dir, f"{ds_name}_metadata_test_5s.parquet"),
|
| 440 |
})
|
| 441 |
|
| 442 |
+
# TODO no more extraction
|
| 443 |
+
#local_audio_archives_paths = dl_manager.extract(dl_dir) if not dl_manager.is_streaming else None
|
| 444 |
+
#local_audio_archives_paths = dl_manager.iter_archive(dl_dir)
|
| 445 |
|
| 446 |
if self.config.name.startswith("XC") or self.config.name.endswith("_xc"):
|
| 447 |
return [
|
|
|
|
| 449 |
name=datasets.Split.TRAIN,
|
| 450 |
gen_kwargs={
|
| 451 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["train"]],
|
| 452 |
+
#"local_audio_archives_paths": local_audio_archives_paths["train"] if local_audio_archives_paths else None,
|
| 453 |
"metapath": dl_dir["metadata"],
|
| 454 |
"split": datasets.Split.TRAIN,
|
| 455 |
},
|
|
|
|
| 458 |
name="valid",
|
| 459 |
gen_kwargs={
|
| 460 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["valid"]],
|
| 461 |
+
#"local_audio_archives_paths": local_audio_archives_paths["valid"] if local_audio_archives_paths else None,
|
| 462 |
"metapath": dl_dir["meta_test_5s"],
|
| 463 |
"split": "valid",
|
| 464 |
},
|
|
|
|
| 471 |
name=datasets.Split.TEST,
|
| 472 |
gen_kwargs={
|
| 473 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["test"]],
|
| 474 |
+
#"local_audio_archives_paths": local_audio_archives_paths["test"] if local_audio_archives_paths else None,
|
| 475 |
"metapath": dl_dir["metadata"],
|
| 476 |
"split": datasets.Split.TEST,
|
| 477 |
},
|
|
|
|
| 480 |
name="test_5s",
|
| 481 |
gen_kwargs={
|
| 482 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["test_5s"]],
|
| 483 |
+
#"local_audio_archives_paths": local_audio_archives_paths["test_5s"] if local_audio_archives_paths else None,
|
| 484 |
"metapath": dl_dir["metadata_5s"],
|
| 485 |
"split": "test_multilabel"
|
| 486 |
},
|
|
|
|
| 492 |
name=datasets.Split.TRAIN,
|
| 493 |
gen_kwargs={
|
| 494 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["train"]],
|
| 495 |
+
#"local_audio_archives_paths": local_audio_archives_paths["train"] if local_audio_archives_paths else None,
|
| 496 |
"metapath": dl_dir["meta_train"],
|
| 497 |
"split": datasets.Split.TRAIN,
|
| 498 |
},
|
|
|
|
| 501 |
name=datasets.Split.TEST,
|
| 502 |
gen_kwargs={
|
| 503 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["test"]],
|
| 504 |
+
#"local_audio_archives_paths": local_audio_archives_paths["test"] if local_audio_archives_paths else None,
|
| 505 |
"metapath": dl_dir["meta_test"],
|
| 506 |
"split": datasets.Split.TEST,
|
| 507 |
},
|
|
|
|
| 510 |
name="test_5s",
|
| 511 |
gen_kwargs={
|
| 512 |
"audio_archive_iterators": [dl_manager.iter_archive(archive_path) for archive_path in dl_dir["test_5s"]],
|
| 513 |
+
#"local_audio_archives_paths": local_audio_archives_paths["test_5s"] if local_audio_archives_paths else None,
|
| 514 |
"metapath": dl_dir["meta_test_5s"],
|
| 515 |
"split": "test_multilabel"
|
| 516 |
},
|
| 517 |
),
|
| 518 |
]
|
| 519 |
|
| 520 |
+
def _generate_examples(self, audio_archive_iterators, metapath, split):
|
| 521 |
metadata = pd.read_parquet(metapath)
|
| 522 |
idx = 0
|
| 523 |
for i, audio_archive_iterator in enumerate(audio_archive_iterators):
|
| 524 |
for audio_path_in_archive, audio_file in audio_archive_iterator:
|
| 525 |
id = os.path.split(audio_path_in_archive)[-1]
|
| 526 |
rows = metadata[metadata.index == (int(id[2:].split(".")[0]) if split == "train" else id)]
|
| 527 |
+
audio_path = audio_path_in_archive
|
| 528 |
|
|
|
|
| 529 |
for _, row in rows.iterrows():
|
| 530 |
idx += 1
|
| 531 |
yield id if split == "train" else idx, {
|
| 532 |
+
"audio": audio_file.read(),
|
| 533 |
"filepath": audio_path,
|
| 534 |
"start_time": row["start_time"],
|
| 535 |
"end_time": row["end_time"],
|
|
|
|
| 559 |
"species_group_multilabel": row.get("species_group_multilabel", None),
|
| 560 |
"order_multilabel": row.get("order_multilabel", None),
|
| 561 |
}
|
| 562 |
+
|
| 563 |
+
os.remove(audio_archive_iterator.args[0])
|