asahi417 commited on
Commit
1af401f
·
1 Parent(s): 6c57588
Files changed (2) hide show
  1. main_s2s.sh +22 -10
  2. tokenize_dataset_s2s.py +5 -1
main_s2s.sh CHANGED
@@ -84,7 +84,7 @@ do
84
  python fetch_dataset_s2s.py
85
  done
86
  # tokenize
87
- for i in $(seq 81 100);
88
  do
89
  export N_POOL=15
90
  export DATASET_ID=${i}
@@ -95,7 +95,7 @@ do
95
  python tokenize_dataset_s2s.py
96
  done
97
  # speaker embedding
98
- for i in $(seq 81 100);
99
  do
100
  export DATASET_ID=${i}
101
  export DIRECTION="enA-zhA"
@@ -132,7 +132,7 @@ do
132
  python fetch_dataset_s2s.py
133
  done
134
  # tokenize
135
- for i in $(seq 1 30);
136
  do
137
  export N_POOL=15
138
  export DATASET_ID=${i}
@@ -142,7 +142,7 @@ do
142
  echo ${LINE_NO_START}
143
  python tokenize_dataset_s2s.py
144
  done
145
- for i in $(seq 300 330);
146
  do
147
  export N_POOL=15
148
  export DATASET_ID=${i}
@@ -153,14 +153,14 @@ do
153
  python tokenize_dataset_s2s.py
154
  done
155
  # speaker embedding
156
- for i in $(seq 1 30);
157
  do
158
  export DATASET_ID=${i}
159
  export DIRECTION="enA-frA"
160
  python attach_speaker_embedding_s2s.py
161
  done
162
  # speaker embedding
163
- for i in $(seq 300 330);
164
  do
165
  export DATASET_ID=${i}
166
  export DIRECTION="enA-frA"
@@ -197,7 +197,19 @@ do
197
  python fetch_dataset_s2s.py
198
  done
199
  # tokenize
200
- for i in $(seq 1 20);
 
 
 
 
 
 
 
 
 
 
 
 
201
  do
202
  export N_POOL=15
203
  export DATASET_ID=${i}
@@ -207,7 +219,7 @@ do
207
  echo ${LINE_NO_START}
208
  python tokenize_dataset_s2s.py
209
  done
210
- for i in $(seq 200 240);
211
  do
212
  export N_POOL=15
213
  export DATASET_ID=${i}
@@ -218,14 +230,14 @@ do
218
  python tokenize_dataset_s2s.py
219
  done
220
  # speaker embedding
221
- for i in $(seq 1 20);
222
  do
223
  export DATASET_ID=${i}
224
  export DIRECTION="deA-enA"
225
  python attach_speaker_embedding_s2s.py
226
  done
227
  # speaker embedding
228
- for i in $(seq 200 240);
229
  do
230
  export DATASET_ID=${i}
231
  export DIRECTION="deA-enA"
 
84
  python fetch_dataset_s2s.py
85
  done
86
  # tokenize
87
+ for i in $(seq 101 150);
88
  do
89
  export N_POOL=15
90
  export DATASET_ID=${i}
 
95
  python tokenize_dataset_s2s.py
96
  done
97
  # speaker embedding
98
+ for i in $(seq 101 150);
99
  do
100
  export DATASET_ID=${i}
101
  export DIRECTION="enA-zhA"
 
132
  python fetch_dataset_s2s.py
133
  done
134
  # tokenize
135
+ for i in $(seq 31 60);
136
  do
137
  export N_POOL=15
138
  export DATASET_ID=${i}
 
142
  echo ${LINE_NO_START}
143
  python tokenize_dataset_s2s.py
144
  done
145
+ for i in $(seq 331 360);
146
  do
147
  export N_POOL=15
148
  export DATASET_ID=${i}
 
153
  python tokenize_dataset_s2s.py
154
  done
155
  # speaker embedding
156
+ for i in $(seq 31 60);
157
  do
158
  export DATASET_ID=${i}
159
  export DIRECTION="enA-frA"
160
  python attach_speaker_embedding_s2s.py
161
  done
162
  # speaker embedding
163
+ for i in $(seq 331 360);
164
  do
165
  export DATASET_ID=${i}
166
  export DIRECTION="enA-frA"
 
197
  python fetch_dataset_s2s.py
198
  done
199
  # tokenize
200
+ #for i in 215 216 232;
201
+ #do
202
+ # export N_POOL=15
203
+ # export DATASET_ID=${i}
204
+ # export DIRECTION="deA-enA"
205
+ # export LINE_NO_START=$(((DATASET_ID-1) * 2500))
206
+ # export LINE_NO_END=$((DATASET_ID * 2500))
207
+ # echo ${LINE_NO_START}
208
+ # python tokenize_dataset_s2s.py
209
+ #
210
+ #done
211
+
212
+ for i in $(seq 21 50);
213
  do
214
  export N_POOL=15
215
  export DATASET_ID=${i}
 
219
  echo ${LINE_NO_START}
220
  python tokenize_dataset_s2s.py
221
  done
222
+ for i in $(seq 241 290);
223
  do
224
  export N_POOL=15
225
  export DATASET_ID=${i}
 
230
  python tokenize_dataset_s2s.py
231
  done
232
  # speaker embedding
233
+ for i in $(seq 21 50);
234
  do
235
  export DATASET_ID=${i}
236
  export DIRECTION="deA-enA"
237
  python attach_speaker_embedding_s2s.py
238
  done
239
  # speaker embedding
240
+ for i in $(seq 241 290);
241
  do
242
  export DATASET_ID=${i}
243
  export DIRECTION="deA-enA"
tokenize_dataset_s2s.py CHANGED
@@ -23,7 +23,9 @@ audio_loader = Audio()
23
  def error_file(example):
24
  for side in sides:
25
  try:
26
- audio_loader.decode_example(example[f"{side}.audio"])
 
 
27
  except LibsndfileError:
28
  return False
29
  return True
@@ -41,6 +43,8 @@ print(f"Num examples (after filtering): {len(dataset)}")
41
  def tokenize(example):
42
  for side in sides:
43
  wav = torch.as_tensor(example[f"{side}.audio"]["array"].reshape(1, 1, -1), dtype=torch.float32)
 
 
44
  example[f"{side}.audio.tokens"] = tokenizer.wav_to_tokens(
45
  wav=wav, sample_rate=example[f"{side}.audio"]["sampling_rate"]
46
  ).numpy().tolist()[0]
 
23
  def error_file(example):
24
  for side in sides:
25
  try:
26
+ wav = audio_loader.decode_example(example[f"{side}.audio"])
27
+ if len(wav["array"]) == 0:
28
+ return False
29
  except LibsndfileError:
30
  return False
31
  return True
 
43
  def tokenize(example):
44
  for side in sides:
45
  wav = torch.as_tensor(example[f"{side}.audio"]["array"].reshape(1, 1, -1), dtype=torch.float32)
46
+ if len(wav) == 0:
47
+ return None
48
  example[f"{side}.audio.tokens"] = tokenizer.wav_to_tokens(
49
  wav=wav, sample_rate=example[f"{side}.audio"]["sampling_rate"]
50
  ).numpy().tolist()[0]