ru / tokenizer /g2p /tokenizer.py
TeraSpace's picture
Upload 13 files
dfc143a
raw
history blame
1.46 kB
import re
from .g2p import *
import json
import os
class Tokenizer():
def __init__(self, data_path: str) -> None:
self.dic = {}
for line in open(os.path.join(data_path, "dictionary.txt")):
items = line.split()
self.dic[items[0]] = " ".join(items[1:])
self.config = json.load(open(os.path.join(data_path, "config.json")))
def g2p(self, text):
text = re.sub("—", "-", text)
text = re.sub("([!'(),-.:;?])", r' \1 ', text)
phonemes = []
for word in text.split():
if re.match("[!'(),-.:;?]", word):
phonemes.append(word)
continue
word = word.lower()
if len(phonemes) > 0: phonemes.append(' ')
if word in self.dic:
phonemes.extend(self.dic[word].split())
else:
phonemes.extend(convert(word).split())
phoneme_id_map = self.config["phoneme_id_map"]
phoneme_ids = []
phoneme_ids.extend(phoneme_id_map["^"])
phoneme_ids.extend(phoneme_id_map["_"])
for p in phonemes:
if p in phoneme_id_map:
phoneme_ids.extend(phoneme_id_map[p])
phoneme_ids.extend(phoneme_id_map["_"])
phoneme_ids.extend(phoneme_id_map["$"])
return phoneme_ids, phonemes
def _get_seq(self, text: str) -> list[int]:
seq = self.g2p(text)[0]
return seq