| | """ |
| | Tatoeba (https://tatoeba.org/) is a collection of sentences and translation, mainly aiming for language learning. |
| | It is available for more than 300 languages. |
| | |
| | This script downloads the Tatoeba corpus and extracts the sentences & translations in the languages you like |
| | """ |
| | import os |
| | import sentence_transformers |
| | import tarfile |
| | import gzip |
| |
|
| | |
| | |
| | |
| | |
| | source_languages = set(['eng']) |
| | target_languages = set(['deu', 'ara', 'tur', 'spa', 'ita', 'fra']) |
| |
|
| | num_dev_sentences = 1000 |
| |
|
| |
|
| | tatoeba_folder = "../datasets/tatoeba" |
| | output_folder = "parallel-sentences/" |
| |
|
| |
|
| |
|
| |
|
| | sentences_file_bz2 = os.path.join(tatoeba_folder, 'sentences.tar.bz2') |
| | sentences_file = os.path.join(tatoeba_folder, 'sentences.csv') |
| | links_file_bz2 = os.path.join(tatoeba_folder, 'links.tar.bz2') |
| | links_file = os.path.join(tatoeba_folder, 'links.csv') |
| |
|
| | download_url = "https://downloads.tatoeba.org/exports/" |
| |
|
| |
|
| | os.makedirs(tatoeba_folder, exist_ok=True) |
| | os.makedirs(output_folder, exist_ok=True) |
| |
|
| | |
| | for filepath in [sentences_file_bz2, links_file_bz2]: |
| | if not os.path.exists(filepath): |
| | url = download_url+os.path.basename(filepath) |
| | print("Download", url) |
| | sentence_transformers.util.http_get(url, filepath) |
| |
|
| | |
| | if not os.path.exists(sentences_file): |
| | print("Extract", sentences_file_bz2) |
| | tar = tarfile.open(sentences_file_bz2, "r:bz2") |
| | tar.extract('sentences.csv', path=tatoeba_folder) |
| | tar.close() |
| |
|
| | if not os.path.exists(links_file): |
| | print("Extract", links_file_bz2) |
| | tar = tarfile.open(links_file_bz2, "r:bz2") |
| | tar.extract('links.csv', path=tatoeba_folder) |
| | tar.close() |
| |
|
| |
|
| | |
| | sentences = {} |
| | all_langs = target_languages.union(source_languages) |
| | print("Read sentences.csv file") |
| | with open(sentences_file, encoding='utf8') as fIn: |
| | for line in fIn: |
| | id, lang, sentence = line.strip().split('\t') |
| | if lang in all_langs: |
| | sentences[id] = (lang, sentence) |
| |
|
| | |
| | print("Read links.csv") |
| | translations = {src_lang: {trg_lang: {} for trg_lang in target_languages} for src_lang in source_languages} |
| | with open(links_file, encoding='utf8') as fIn: |
| | for line in fIn: |
| | src_id, target_id = line.strip().split() |
| |
|
| | if src_id in sentences and target_id in sentences: |
| | src_lang, src_sent = sentences[src_id] |
| | trg_lang, trg_sent = sentences[target_id] |
| |
|
| | if src_lang in source_languages and trg_lang in target_languages: |
| | if src_sent not in translations[src_lang][trg_lang]: |
| | translations[src_lang][trg_lang][src_sent] = [] |
| | translations[src_lang][trg_lang][src_sent].append(trg_sent) |
| |
|
| | |
| | print("Write output files") |
| | for src_lang in source_languages: |
| | for trg_lang in target_languages: |
| | source_sentences = list(translations[src_lang][trg_lang]) |
| | train_sentences = source_sentences[num_dev_sentences:] |
| | dev_sentences = source_sentences[0:num_dev_sentences] |
| |
|
| | print("{}-{} has {} sentences".format(src_lang, trg_lang, len(source_sentences))) |
| | if len(dev_sentences) > 0: |
| | with gzip.open(os.path.join(output_folder, 'Tatoeba-{}-{}-dev.tsv.gz'.format(src_lang, trg_lang)), 'wt', encoding='utf8') as fOut: |
| | for sent in dev_sentences: |
| | fOut.write("\t".join([sent]+translations[src_lang][trg_lang][sent])) |
| | fOut.write("\n") |
| |
|
| | if len(train_sentences) > 0: |
| | with gzip.open(os.path.join(output_folder, 'Tatoeba-{}-{}-train.tsv.gz'.format(src_lang, trg_lang)), 'wt', encoding='utf8') as fOut: |
| | for sent in train_sentences: |
| | fOut.write("\t".join([sent]+translations[src_lang][trg_lang][sent])) |
| | fOut.write("\n") |
| |
|
| |
|
| | print("---DONE---") |
| |
|