XLM-R-SQuAD-sr-lat

This is XLM-R-based model finetuned on synthetic question answering dataset which is created by translating SQuAD 1.1. This model is the result of my thesis.

Usage

from transformers import pipeline

model_name = 'aleksahet/xlm-r-squad-sr-lat'

pipe = pipeline('question-answering', model=model_name, tokenizer=model_name)

sample = {
  'question': 'U kom gradu je rođen Željko Obradović?',
  'context': 'Željko Obradović (Čačak, 9. mart 1960) bivši je srpski i jugoslovenski košarkaš. Najuspešniji je trener u istoriji košarke.'
}

res = pipe(sample)

Performance

Model was tested on synthetic question answering dataset, created by automatic translation of SQuAD 1.1 dev split. The model achieved the following results:

  • Exact Match: 71.04
  • F1: 81.62

Source Code

Source code for synthetic dataset generation and model finetuning can be found on this GitHub repository.

Downloads last month
17
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.