Zerpal
Collection
The largest open-source Udmurt monolingual corpora and pre-trained language models • 12 items • Updated • 1
How to use udmurtNLP/zerpal-mbert with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("fill-mask", model="udmurtNLP/zerpal-mbert") # Load model directly
from transformers import AutoTokenizer, AutoModelForMaskedLM
tokenizer = AutoTokenizer.from_pretrained("udmurtNLP/zerpal-mbert")
model = AutoModelForMaskedLM.from_pretrained("udmurtNLP/zerpal-mbert")You can use this model directly with a pipeline for masked language modeling:
from transformers import pipeline
unmasker = pipeline('fill-mask', model='udmurtNLP/zerpal-mbert', tokenizer='udmurtNLP/zerpal-mbert-tokenizer')
unmasker("Ӟечбур! Мынам нимы [MASK].")
Here is how to use this model to get the features of a given text in PyTorch:
from transformers import BertTokenizer, BertModel
tokenizer = BertTokenizer.from_pretrained('udmurtNLP/zerpal-mbert-tokenizer')
model = BertModel.from_pretrained("udmurtNLP/zerpal-mbert")
text = "Яратон, яратон, мар меда сыӵе тон?"
encoded_input = tokenizer(text, return_tensors='pt')
output = model(**encoded_input)