add hour splits
Browse files- libriheavy.py +40 -2
libriheavy.py
CHANGED
@@ -169,11 +169,41 @@ class Libriheavy(datasets.GeneratorBasedBuilder):
|
|
169 |
name="odd",
|
170 |
gen_kwargs={"speaker_chunks": odd_speaker_chunks, "split": "odd"}
|
171 |
),
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
172 |
]
|
173 |
|
174 |
-
def _generate_examples(self, speaker_chunks, split):
|
175 |
"""Yields examples."""
|
|
|
|
|
|
|
|
|
176 |
for chunk in speaker_chunks:
|
|
|
|
|
177 |
npz = dict(np.load(chunk["audio"], allow_pickle=True))
|
178 |
utterances = npz.keys()
|
179 |
with gzip.open(chunk["text"], "rt") as f:
|
@@ -211,7 +241,11 @@ class Libriheavy(datasets.GeneratorBasedBuilder):
|
|
211 |
"srmr": npz_item["overall_srmr"],
|
212 |
},
|
213 |
}
|
|
|
214 |
yield chunk["speaker_id"] + "_" + utterance_id, result
|
|
|
|
|
|
|
215 |
else:
|
216 |
# only use the last utterance
|
217 |
utterance_id = sorted(list(text.keys()))[-1]
|
@@ -244,4 +278,8 @@ class Libriheavy(datasets.GeneratorBasedBuilder):
|
|
244 |
"srmr": npz_item["overall_srmr"],
|
245 |
},
|
246 |
}
|
247 |
-
|
|
|
|
|
|
|
|
|
|
169 |
name="odd",
|
170 |
gen_kwargs={"speaker_chunks": odd_speaker_chunks, "split": "odd"}
|
171 |
),
|
172 |
+
datasets.SplitGenerator(
|
173 |
+
name="even-100",
|
174 |
+
gen_kwargs={"speaker_chunks": even_speaker_chunks, "split": "even", "hours": 100}
|
175 |
+
),
|
176 |
+
datasets.SplitGenerator(
|
177 |
+
name="odd-100",
|
178 |
+
gen_kwargs={"speaker_chunks": odd_speaker_chunks, "split": "odd", "hours": 100}
|
179 |
+
),
|
180 |
+
datasets.SplitGenerator(
|
181 |
+
name="even-500",
|
182 |
+
gen_kwargs={"speaker_chunks": even_speaker_chunks, "split": "even", "hours": 500}
|
183 |
+
),
|
184 |
+
datasets.SplitGenerator(
|
185 |
+
name="odd-500",
|
186 |
+
gen_kwargs={"speaker_chunks": odd_speaker_chunks, "split": "odd", "hours": 500}
|
187 |
+
),
|
188 |
+
datasets.SplitGenerator(
|
189 |
+
name="even-1000",
|
190 |
+
gen_kwargs={"speaker_chunks": even_speaker_chunks, "split": "even", "hours": 1000}
|
191 |
+
),
|
192 |
+
datasets.SplitGenerator(
|
193 |
+
name="odd-1000",
|
194 |
+
gen_kwargs={"speaker_chunks": odd_speaker_chunks, "split": "odd", "hours": 1000}
|
195 |
+
),
|
196 |
]
|
197 |
|
198 |
+
def _generate_examples(self, speaker_chunks, split, hours=None):
|
199 |
"""Yields examples."""
|
200 |
+
hours_streamed = 0
|
201 |
+
finish_stream = False
|
202 |
+
if hours is None:
|
203 |
+
hours = float("inf")
|
204 |
for chunk in speaker_chunks:
|
205 |
+
if finish_stream:
|
206 |
+
break
|
207 |
npz = dict(np.load(chunk["audio"], allow_pickle=True))
|
208 |
utterances = npz.keys()
|
209 |
with gzip.open(chunk["text"], "rt") as f:
|
|
|
241 |
"srmr": npz_item["overall_srmr"],
|
242 |
},
|
243 |
}
|
244 |
+
hours_streamed += (utterance["word_segments"][-1][1] - utterance["word_segments"][0][0]) / 3600
|
245 |
yield chunk["speaker_id"] + "_" + utterance_id, result
|
246 |
+
if hours_streamed >= hours:
|
247 |
+
finish_stream = True
|
248 |
+
break
|
249 |
else:
|
250 |
# only use the last utterance
|
251 |
utterance_id = sorted(list(text.keys()))[-1]
|
|
|
278 |
"srmr": npz_item["overall_srmr"],
|
279 |
},
|
280 |
}
|
281 |
+
hours_streamed += (utterance["word_segments"][-1][1] - utterance["word_segments"][0][0]) / 3600
|
282 |
+
yield chunk["speaker_id"] + "_" + utterance_id, result
|
283 |
+
if hours_streamed >= hours:
|
284 |
+
finish_stream = True
|
285 |
+
break
|