| | import re |
| |
|
| | import inflect |
| | import torch |
| | from tokenizers import Tokenizer |
| |
|
| |
|
| | |
| | from unidecode import unidecode |
| |
|
| | _whitespace_re = re.compile(r'\s+') |
| |
|
| |
|
| | |
| | _abbreviations = [(re.compile('\\b%s\\.' % x[0], re.IGNORECASE), x[1]) for x in [ |
| | ('mrs', 'misess'), |
| | ('mr', 'mister'), |
| | ('dr', 'doctor'), |
| | ('st', 'saint'), |
| | ('co', 'company'), |
| | ('jr', 'junior'), |
| | ('maj', 'major'), |
| | ('gen', 'general'), |
| | ('drs', 'doctors'), |
| | ('rev', 'reverend'), |
| | ('lt', 'lieutenant'), |
| | ('hon', 'honorable'), |
| | ('sgt', 'sergeant'), |
| | ('capt', 'captain'), |
| | ('esq', 'esquire'), |
| | ('ltd', 'limited'), |
| | ('col', 'colonel'), |
| | ('ft', 'fort'), |
| | ]] |
| |
|
| |
|
| | def expand_abbreviations(text): |
| | for regex, replacement in _abbreviations: |
| | text = re.sub(regex, replacement, text) |
| | return text |
| |
|
| |
|
| | _inflect = inflect.engine() |
| | _comma_number_re = re.compile(r'([0-9][0-9\,]+[0-9])') |
| | _decimal_number_re = re.compile(r'([0-9]+\.[0-9]+)') |
| | _pounds_re = re.compile(r'£([0-9\,]*[0-9]+)') |
| | _dollars_re = re.compile(r'\$([0-9\.\,]*[0-9]+)') |
| | _ordinal_re = re.compile(r'[0-9]+(st|nd|rd|th)') |
| | _number_re = re.compile(r'[0-9]+') |
| |
|
| |
|
| | def _remove_commas(m): |
| | return m.group(1).replace(',', '') |
| |
|
| |
|
| | def _expand_decimal_point(m): |
| | return m.group(1).replace('.', ' point ') |
| |
|
| |
|
| | def _expand_dollars(m): |
| | match = m.group(1) |
| | parts = match.split('.') |
| | if len(parts) > 2: |
| | return match + ' dollars' |
| | dollars = int(parts[0]) if parts[0] else 0 |
| | cents = int(parts[1]) if len(parts) > 1 and parts[1] else 0 |
| | if dollars and cents: |
| | dollar_unit = 'dollar' if dollars == 1 else 'dollars' |
| | cent_unit = 'cent' if cents == 1 else 'cents' |
| | return '%s %s, %s %s' % (dollars, dollar_unit, cents, cent_unit) |
| | elif dollars: |
| | dollar_unit = 'dollar' if dollars == 1 else 'dollars' |
| | return '%s %s' % (dollars, dollar_unit) |
| | elif cents: |
| | cent_unit = 'cent' if cents == 1 else 'cents' |
| | return '%s %s' % (cents, cent_unit) |
| | else: |
| | return 'zero dollars' |
| |
|
| |
|
| | def _expand_ordinal(m): |
| | return _inflect.number_to_words(m.group(0)) |
| |
|
| |
|
| | def _expand_number(m): |
| | num = int(m.group(0)) |
| | if num > 1000 and num < 3000: |
| | if num == 2000: |
| | return 'two thousand' |
| | elif num > 2000 and num < 2010: |
| | return 'two thousand ' + _inflect.number_to_words(num % 100) |
| | elif num % 100 == 0: |
| | return _inflect.number_to_words(num // 100) + ' hundred' |
| | else: |
| | return _inflect.number_to_words(num, andword='', zero='oh', group=2).replace(', ', ' ') |
| | else: |
| | return _inflect.number_to_words(num, andword='') |
| |
|
| |
|
| | def normalize_numbers(text): |
| | text = re.sub(_comma_number_re, _remove_commas, text) |
| | text = re.sub(_pounds_re, r'\1 pounds', text) |
| | text = re.sub(_dollars_re, _expand_dollars, text) |
| | text = re.sub(_decimal_number_re, _expand_decimal_point, text) |
| | text = re.sub(_ordinal_re, _expand_ordinal, text) |
| | text = re.sub(_number_re, _expand_number, text) |
| | return text |
| |
|
| |
|
| | def expand_numbers(text): |
| | return normalize_numbers(text) |
| |
|
| |
|
| | def lowercase(text): |
| | return text.lower() |
| |
|
| |
|
| | def collapse_whitespace(text): |
| | return re.sub(_whitespace_re, ' ', text) |
| |
|
| |
|
| | def convert_to_ascii(text): |
| | return unidecode(text) |
| |
|
| |
|
| | def basic_cleaners(text): |
| | '''Basic pipeline that lowercases and collapses whitespace without transliteration.''' |
| | text = lowercase(text) |
| | text = collapse_whitespace(text) |
| | return text |
| |
|
| |
|
| | def transliteration_cleaners(text): |
| | '''Pipeline for non-English text that transliterates to ASCII.''' |
| | text = convert_to_ascii(text) |
| | text = lowercase(text) |
| | text = collapse_whitespace(text) |
| | return text |
| |
|
| |
|
| | def english_cleaners(text): |
| | '''Pipeline for English text, including number and abbreviation expansion.''' |
| | text = convert_to_ascii(text) |
| | text = lowercase(text) |
| | text = expand_numbers(text) |
| | text = expand_abbreviations(text) |
| | text = collapse_whitespace(text) |
| | text = text.replace('"', '') |
| | return text |
| |
|
| | def lev_distance(s1, s2): |
| | if len(s1) > len(s2): |
| | s1, s2 = s2, s1 |
| |
|
| | distances = range(len(s1) + 1) |
| | for i2, c2 in enumerate(s2): |
| | distances_ = [i2 + 1] |
| | for i1, c1 in enumerate(s1): |
| | if c1 == c2: |
| | distances_.append(distances[i1]) |
| | else: |
| | distances_.append(1 + min((distances[i1], distances[i1 + 1], distances_[-1]))) |
| | distances = distances_ |
| | return distances[-1] |
| |
|
| | class VoiceBpeTokenizer: |
| | def __init__(self, vocab_file='data/tokenizer.json'): |
| | if vocab_file is not None: |
| | self.tokenizer = Tokenizer.from_file(vocab_file) |
| |
|
| | def preprocess_text(self, txt): |
| | txt = english_cleaners(txt) |
| | return txt |
| |
|
| | def encode(self, txt): |
| | txt = self.preprocess_text(txt) |
| | txt = txt.replace(' ', '[SPACE]') |
| | return self.tokenizer.encode(txt).ids |
| |
|
| | def decode(self, seq): |
| | if isinstance(seq, torch.Tensor): |
| | seq = seq.cpu().numpy() |
| | txt = self.tokenizer.decode(seq, skip_special_tokens=False).replace(' ', '') |
| | txt = txt.replace('[SPACE]', ' ') |
| | txt = txt.replace('[STOP]', '') |
| | txt = txt.replace('[UNK]', '') |
| | return txt |