ptaszynski's picture
Update README.md
e9ee0da
|
raw
history blame
1.18 kB
metadata
language: ja
license: cc-by-sa-4.0
datasets:
  - YACIS corpus

yacis-electra-small

This is ELECTRA Small model for Japanese pretrained on 354 million sentences / 5.6 billion words of YACIS blog corpus.

The corpus was tokenized for pretraining with MeCab. Subword tokenization was peroformed with WordPiece.

Model architecture

This model uses the original ELECTRA Small model; 12 layers, 128 dimensions of hidden states, and 12 attention heads.

Vocabulary size was 32,000 tokens.

Licenses

The pretrained model with all attached files is distributed under the terms of the CC BY-SA 4.0 license. Creative Commons License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.