| | import pickle |
| | import os |
| | import re |
| | from g2p_en import G2p |
| |
|
| | from . import symbols |
| |
|
| | from .english_utils.abbreviations import expand_abbreviations |
| | from .english_utils.time_norm import expand_time_english |
| | from .english_utils.number_norm import normalize_numbers |
| | from .japanese import distribute_phone |
| |
|
| | from transformers import AutoTokenizer |
| |
|
| | current_file_path = os.path.dirname(__file__) |
| | CMU_DICT_PATH = os.path.join(current_file_path, "cmudict.rep") |
| | CACHE_PATH = os.path.join(current_file_path, "cmudict_cache.pickle") |
| | _g2p = G2p() |
| |
|
| | arpa = { |
| | "AH0", |
| | "S", |
| | "AH1", |
| | "EY2", |
| | "AE2", |
| | "EH0", |
| | "OW2", |
| | "UH0", |
| | "NG", |
| | "B", |
| | "G", |
| | "AY0", |
| | "M", |
| | "AA0", |
| | "F", |
| | "AO0", |
| | "ER2", |
| | "UH1", |
| | "IY1", |
| | "AH2", |
| | "DH", |
| | "IY0", |
| | "EY1", |
| | "IH0", |
| | "K", |
| | "N", |
| | "W", |
| | "IY2", |
| | "T", |
| | "AA1", |
| | "ER1", |
| | "EH2", |
| | "OY0", |
| | "UH2", |
| | "UW1", |
| | "Z", |
| | "AW2", |
| | "AW1", |
| | "V", |
| | "UW2", |
| | "AA2", |
| | "ER", |
| | "AW0", |
| | "UW0", |
| | "R", |
| | "OW1", |
| | "EH1", |
| | "ZH", |
| | "AE0", |
| | "IH2", |
| | "IH", |
| | "Y", |
| | "JH", |
| | "P", |
| | "AY1", |
| | "EY0", |
| | "OY2", |
| | "TH", |
| | "HH", |
| | "D", |
| | "ER0", |
| | "CH", |
| | "AO1", |
| | "AE1", |
| | "AO2", |
| | "OY1", |
| | "AY2", |
| | "IH1", |
| | "OW0", |
| | "L", |
| | "SH", |
| | } |
| |
|
| |
|
| | def post_replace_ph(ph): |
| | rep_map = { |
| | ":": ",", |
| | ";": ",", |
| | ",": ",", |
| | "。": ".", |
| | "!": "!", |
| | "?": "?", |
| | "\n": ".", |
| | "·": ",", |
| | "、": ",", |
| | "...": "…", |
| | "v": "V", |
| | } |
| | if ph in rep_map.keys(): |
| | ph = rep_map[ph] |
| | if ph in symbols: |
| | return ph |
| | if ph not in symbols: |
| | ph = "UNK" |
| | return ph |
| |
|
| |
|
| | def read_dict(): |
| | g2p_dict = {} |
| | start_line = 49 |
| | with open(CMU_DICT_PATH) as f: |
| | line = f.readline() |
| | line_index = 1 |
| | while line: |
| | if line_index >= start_line: |
| | line = line.strip() |
| | word_split = line.split(" ") |
| | word = word_split[0] |
| |
|
| | syllable_split = word_split[1].split(" - ") |
| | g2p_dict[word] = [] |
| | for syllable in syllable_split: |
| | phone_split = syllable.split(" ") |
| | g2p_dict[word].append(phone_split) |
| |
|
| | line_index = line_index + 1 |
| | line = f.readline() |
| |
|
| | return g2p_dict |
| |
|
| |
|
| | def cache_dict(g2p_dict, file_path): |
| | with open(file_path, "wb") as pickle_file: |
| | pickle.dump(g2p_dict, pickle_file) |
| |
|
| |
|
| | def get_dict(): |
| | if os.path.exists(CACHE_PATH): |
| | with open(CACHE_PATH, "rb") as pickle_file: |
| | g2p_dict = pickle.load(pickle_file) |
| | else: |
| | g2p_dict = read_dict() |
| | cache_dict(g2p_dict, CACHE_PATH) |
| |
|
| | return g2p_dict |
| |
|
| |
|
| | eng_dict = get_dict() |
| |
|
| |
|
| | def refine_ph(phn): |
| | tone = 0 |
| | if re.search(r"\d$", phn): |
| | tone = int(phn[-1]) + 1 |
| | phn = phn[:-1] |
| | return phn.lower(), tone |
| |
|
| |
|
| | def refine_syllables(syllables): |
| | tones = [] |
| | phonemes = [] |
| | for phn_list in syllables: |
| | for i in range(len(phn_list)): |
| | phn = phn_list[i] |
| | phn, tone = refine_ph(phn) |
| | phonemes.append(phn) |
| | tones.append(tone) |
| | return phonemes, tones |
| |
|
| |
|
| | def text_normalize(text): |
| | text = text.lower() |
| | text = expand_time_english(text) |
| | text = normalize_numbers(text) |
| | text = expand_abbreviations(text) |
| | return text |
| |
|
| | model_id = 'bert-base-uncased' |
| | tokenizer = AutoTokenizer.from_pretrained(model_id) |
| | def g2p_old(text): |
| | tokenized = tokenizer.tokenize(text) |
| | |
| | phones = [] |
| | tones = [] |
| | words = re.split(r"([,;.\-\?\!\s+])", text) |
| | for w in words: |
| | if w.upper() in eng_dict: |
| | phns, tns = refine_syllables(eng_dict[w.upper()]) |
| | phones += phns |
| | tones += tns |
| | else: |
| | phone_list = list(filter(lambda p: p != " ", _g2p(w))) |
| | for ph in phone_list: |
| | if ph in arpa: |
| | ph, tn = refine_ph(ph) |
| | phones.append(ph) |
| | tones.append(tn) |
| | else: |
| | phones.append(ph) |
| | tones.append(0) |
| | |
| | word2ph = [1 for i in phones] |
| |
|
| | phones = [post_replace_ph(i) for i in phones] |
| | return phones, tones, word2ph |
| |
|
| | def g2p(text, pad_start_end=True, tokenized=None): |
| | if tokenized is None: |
| | tokenized = tokenizer.tokenize(text) |
| | |
| | phs = [] |
| | ph_groups = [] |
| | for t in tokenized: |
| | if not t.startswith("#"): |
| | ph_groups.append([t]) |
| | else: |
| | ph_groups[-1].append(t.replace("#", "")) |
| | |
| | phones = [] |
| | tones = [] |
| | word2ph = [] |
| | for group in ph_groups: |
| | w = "".join(group) |
| | phone_len = 0 |
| | word_len = len(group) |
| | if w.upper() in eng_dict: |
| | phns, tns = refine_syllables(eng_dict[w.upper()]) |
| | phones += phns |
| | tones += tns |
| | phone_len += len(phns) |
| | else: |
| | phone_list = list(filter(lambda p: p != " ", _g2p(w))) |
| | for ph in phone_list: |
| | if ph in arpa: |
| | ph, tn = refine_ph(ph) |
| | phones.append(ph) |
| | tones.append(tn) |
| | else: |
| | phones.append(ph) |
| | tones.append(0) |
| | phone_len += 1 |
| | aaa = distribute_phone(phone_len, word_len) |
| | word2ph += aaa |
| | phones = [post_replace_ph(i) for i in phones] |
| |
|
| | if pad_start_end: |
| | phones = ["_"] + phones + ["_"] |
| | tones = [0] + tones + [0] |
| | word2ph = [1] + word2ph + [1] |
| | return phones, tones, word2ph |
| |
|
| | def get_bert_feature(text, word2ph, device=None): |
| | from text import english_bert |
| |
|
| | return english_bert.get_bert_feature(text, word2ph, device=device) |
| |
|
| | if __name__ == "__main__": |
| | |
| | |
| | from text.english_bert import get_bert_feature |
| | text = "In this paper, we propose 1 DSPGAN, a N-F-T GAN-based universal vocoder." |
| | text = text_normalize(text) |
| | phones, tones, word2ph = g2p(text) |
| | import pdb; pdb.set_trace() |
| | bert = get_bert_feature(text, word2ph) |
| | |
| | print(phones, tones, word2ph, bert.shape) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| |
|