Datasets:
Simplify and remove BuilderConfigs
Browse files
rixvox.py
CHANGED
|
@@ -50,18 +50,6 @@ class Rixvox(datasets.GeneratorBasedBuilder):
|
|
| 50 |
VERSION = datasets.Version("1.0.0")
|
| 51 |
DEFAULT_CONFIG_NAME = "all"
|
| 52 |
|
| 53 |
-
BUILDER_CONFIGS = [
|
| 54 |
-
datasets.BuilderConfig(
|
| 55 |
-
name="train", version=VERSION, description="Training set of the RixVox dataset. 5383 hours of speech."
|
| 56 |
-
),
|
| 57 |
-
datasets.BuilderConfig(
|
| 58 |
-
name="dev", version=VERSION, description="Development set of the RixVox dataset. 52 hours of speech."
|
| 59 |
-
),
|
| 60 |
-
datasets.BuilderConfig(
|
| 61 |
-
name="test", version=VERSION, description="Test set of the RixVox dataset. 59 hours of speech."
|
| 62 |
-
),
|
| 63 |
-
]
|
| 64 |
-
|
| 65 |
def _info(self):
|
| 66 |
|
| 67 |
features = datasets.Features(
|
|
@@ -104,20 +92,10 @@ class Rixvox(datasets.GeneratorBasedBuilder):
|
|
| 104 |
splits = ["train", "dev", "test"]
|
| 105 |
meta_urls = {split: [_META_URL.format(split=split)] for split in splits}
|
| 106 |
|
| 107 |
-
|
| 108 |
-
|
| 109 |
-
|
| 110 |
-
|
| 111 |
-
}
|
| 112 |
-
else:
|
| 113 |
-
archive_urls = {
|
| 114 |
-
self.config.name: [
|
| 115 |
-
_DATA_URL.format(split=self.config.name, shard_idx=idx)
|
| 116 |
-
for idx in range(0, _N_SHARDS[self.config.name])
|
| 117 |
-
]
|
| 118 |
-
}
|
| 119 |
-
# Choose single split
|
| 120 |
-
meta_urls = {self.config.name: meta_urls[self.config.name]}
|
| 121 |
|
| 122 |
archive_paths = dl_manager.download(archive_urls)
|
| 123 |
local_extracted_archives = dl_manager.extract(archive_paths) if not dl_manager.is_streaming else {}
|
|
@@ -130,28 +108,14 @@ class Rixvox(datasets.GeneratorBasedBuilder):
|
|
| 130 |
"test": datasets.Split.TEST,
|
| 131 |
}
|
| 132 |
|
| 133 |
-
|
| 134 |
-
for split in splits:
|
| 135 |
-
split_generators.append(
|
| 136 |
-
datasets.SplitGenerator(
|
| 137 |
-
name=split_names.get(split),
|
| 138 |
-
gen_kwargs={
|
| 139 |
-
"local_extracted_archive_paths": local_extracted_archives.get(split),
|
| 140 |
-
"archive_iters": [dl_manager.iter_archive(path) for path in archive_paths.get(split)],
|
| 141 |
-
"meta_paths": meta_paths[split],
|
| 142 |
-
},
|
| 143 |
-
),
|
| 144 |
-
)
|
| 145 |
-
else:
|
| 146 |
split_generators.append(
|
| 147 |
datasets.SplitGenerator(
|
| 148 |
-
name=split_names.get(
|
| 149 |
gen_kwargs={
|
| 150 |
-
"local_extracted_archive_paths": local_extracted_archives.get(
|
| 151 |
-
"archive_iters": [
|
| 152 |
-
|
| 153 |
-
],
|
| 154 |
-
"meta_paths": meta_paths, # meta_paths is a dict with one key, the split name
|
| 155 |
},
|
| 156 |
),
|
| 157 |
)
|
|
@@ -167,14 +131,11 @@ class Rixvox(datasets.GeneratorBasedBuilder):
|
|
| 167 |
):
|
| 168 |
key = 0
|
| 169 |
|
| 170 |
-
|
| 171 |
-
|
| 172 |
-
|
| 173 |
-
data.append(pd.read_parquet(meta_path))
|
| 174 |
|
| 175 |
-
|
| 176 |
-
else:
|
| 177 |
-
df_meta = pd.read_parquet(meta_paths[self.config.name])
|
| 178 |
|
| 179 |
df_meta = df_meta.set_index("filename", drop=False)
|
| 180 |
# Column contains NAType, so we convert to object type column and NAType to None values.
|
|
|
|
| 50 |
VERSION = datasets.Version("1.0.0")
|
| 51 |
DEFAULT_CONFIG_NAME = "all"
|
| 52 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
def _info(self):
|
| 54 |
|
| 55 |
features = datasets.Features(
|
|
|
|
| 92 |
splits = ["train", "dev", "test"]
|
| 93 |
meta_urls = {split: [_META_URL.format(split=split)] for split in splits}
|
| 94 |
|
| 95 |
+
archive_urls = {
|
| 96 |
+
split: [_DATA_URL.format(split=split, shard_idx=idx) for idx in range(0, _N_SHARDS[split])]
|
| 97 |
+
for split in splits
|
| 98 |
+
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 99 |
|
| 100 |
archive_paths = dl_manager.download(archive_urls)
|
| 101 |
local_extracted_archives = dl_manager.extract(archive_paths) if not dl_manager.is_streaming else {}
|
|
|
|
| 108 |
"test": datasets.Split.TEST,
|
| 109 |
}
|
| 110 |
|
| 111 |
+
for split in splits:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 112 |
split_generators.append(
|
| 113 |
datasets.SplitGenerator(
|
| 114 |
+
name=split_names.get(split),
|
| 115 |
gen_kwargs={
|
| 116 |
+
"local_extracted_archive_paths": local_extracted_archives.get(split),
|
| 117 |
+
"archive_iters": [dl_manager.iter_archive(path) for path in archive_paths.get(split)],
|
| 118 |
+
"meta_paths": meta_paths[split],
|
|
|
|
|
|
|
| 119 |
},
|
| 120 |
),
|
| 121 |
)
|
|
|
|
| 131 |
):
|
| 132 |
key = 0
|
| 133 |
|
| 134 |
+
data = []
|
| 135 |
+
for meta_path in meta_paths:
|
| 136 |
+
data.append(pd.read_parquet(meta_path))
|
|
|
|
| 137 |
|
| 138 |
+
df_meta = pd.concat(data)
|
|
|
|
|
|
|
| 139 |
|
| 140 |
df_meta = df_meta.set_index("filename", drop=False)
|
| 141 |
# Column contains NAType, so we convert to object type column and NAType to None values.
|