Papers
arxiv:1909.05658

UER: An Open-Source Toolkit for Pre-training Models

Published on Sep 12, 2019
Authors:
,
,
,
,
,
,
,
,
,

Abstract

Existing works, including ELMO and BERT, have revealed the importance of pre-training for NLP tasks. While there does not exist a single pre-training model that works best in all cases, it is of necessity to develop a framework that is able to deploy various pre-training models efficiently. For this purpose, we propose an assemble-on-demand <PRE_TAG>pre-training toolkit</POST_TAG>, namely Universal Encoder Representations (UER). UER is loosely coupled, and encapsulated with rich modules. By assembling modules on demand, users can either reproduce a state-of-the-art pre-training model or develop a pre-training model that remains unexplored. With UER, we have built a model zoo, which contains pre-trained models based on different corpora, encoders, and targets (objectives). With proper pre-trained models, we could achieve new state-of-the-art results on a range of downstream datasets.

Community

Sign up or log in to comment

Models citing this paper 64

Browse 64 models citing this paper

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/1909.05658 in a dataset README.md to link it from this page.

Spaces citing this paper 48

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.