gpt2-base-fa / model_config.yaml
arxyzan's picture
Hezar: Upload model_config.yaml
f140525
raw
history blame contribute delete
645 Bytes
name: gpt2_text_generation
config_type: model
add_cross_attention: false
vocab_size: 42001
attn_pdrop: 0.1
bos_token_id: 5
embd_pdrop: 0.1
eos_token_id: 5
gradient_checkpointing: false
initializer_range: 0.02
layer_norm_epsilon: 1.0e-05
model_type: gpt2
n_ctx: 1024
n_embd: 768
n_head: 12
n_layer: 12
n_positions: 1024
resid_pdrop: 0.1
summary_activation: false
summary_first_dropout: 0.1
use_cache: true
generation:
config_type: model
bos_token_id: 0
do_sample: true
decoder_start_token_id: 0
early_stopping: true
eos_token_id: 2
length_penalty: 2.0
max_new_tokens: 50
no_repeat_ngram_size: 3
num_beams: 4
pad_token_id: 1