mabornea commited on
Commit
d2d2ac1
1 Parent(s): 69ddf33

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +81 -0
README.md CHANGED
@@ -1,3 +1,84 @@
 
1
  ---
 
 
 
 
 
 
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
  ---
3
+ tags:
4
+ - MRC
5
+ - TyDiQA
6
+ - xlm-roberta-large
7
+ language:
8
+ - multilingual
9
  license: apache-2.0
10
  ---
11
+
12
+ # Model description
13
+
14
+ An XLM-RoBERTa reading comprehension model for [SQuAD 1.1](https://aclanthology.org/D16-1264/).
15
+
16
+ The model is initialized with [xlm-roberta-large](https://huggingface.co/xlm-roberta-large/) and fine-tuned on the [SQuAD 1.1 train data](https://huggingface.co/datasets/squad).
17
+
18
+ ## Intended uses & limitations
19
+
20
+ You can use the raw model for the reading comprehension task. Biases associated with the pre-existing language model, xlm-roberta-large, that we used may be present in our fine-tuned model, squad-v1-xlm-roberta-large. This model is used for zero-shot decoding of [MLQA](https://huggingface.co/datasets/mlqa) and [XQuAD](https://huggingface.co/datasets/xquad) datasets.
21
+
22
+
23
+ ## Usage
24
+
25
+ You can use this model directly with the [PrimeQA](https://github.com/primeqa/primeqa) pipeline for reading comprehension [squad.ipynb](https://github.com/primeqa/primeqa/blob/main/notebooks/mrc/squad.ipynb).
26
+
27
+ ```bibtex
28
+ @article{2016arXiv160605250R,
29
+ author = {{Rajpurkar}, Pranav and {Zhang}, Jian and {Lopyrev},
30
+ Konstantin and {Liang}, Percy},
31
+ title = "{SQuAD: 100,000+ Questions for Machine Comprehension of Text}",
32
+ journal = {arXiv e-prints},
33
+ year = 2016,
34
+ eid = {arXiv:1606.05250},
35
+ pages = {arXiv:1606.05250},
36
+ archivePrefix = {arXiv},
37
+ eprint = {1606.05250},
38
+ }
39
+ ```
40
+ ```bibtex
41
+ @article{lewis2019mlqa,
42
+ title={MLQA: Evaluating Cross-lingual Extractive Question Answering},
43
+ author={Lewis, Patrick and Oguz, Barlas and Rinott, Ruty and Riedel, Sebastian and Schwenk, Holger},
44
+ journal={arXiv preprint arXiv:1910.07475},
45
+ year={2019}
46
+ }
47
+ ```
48
+
49
+ ```bibtex
50
+ @article{Artetxe:etal:2019,
51
+ author = {Mikel Artetxe and Sebastian Ruder and Dani Yogatama},
52
+ title = {On the cross-lingual transferability of monolingual representations},
53
+ journal = {CoRR},
54
+ volume = {abs/1910.11856},
55
+ year = {2019},
56
+ archivePrefix = {arXiv},
57
+ eprint = {1910.11856}
58
+ }
59
+ ```
60
+
61
+ ```bibtex
62
+ @article{DBLP:journals/corr/abs-1911-02116,
63
+ author = {Alexis Conneau and
64
+ Kartikay Khandelwal and
65
+ Naman Goyal and
66
+ Vishrav Chaudhary and
67
+ Guillaume Wenzek and
68
+ Francisco Guzm{\'{a}}n and
69
+ Edouard Grave and
70
+ Myle Ott and
71
+ Luke Zettlemoyer and
72
+ Veselin Stoyanov},
73
+ title = {Unsupervised Cross-lingual Representation Learning at Scale},
74
+ journal = {CoRR},
75
+ volume = {abs/1911.02116},
76
+ year = {2019},
77
+ url = {http://arxiv.org/abs/1911.02116},
78
+ eprinttype = {arXiv},
79
+ eprint = {1911.02116},
80
+ timestamp = {Mon, 11 Nov 2019 18:38:09 +0100},
81
+ biburl = {https://dblp.org/rec/journals/corr/abs-1911-02116.bib},
82
+ bibsource = {dblp computer science bibliography, https://dblp.org}
83
+ }
84
+ ```