BMO_3.5 / config.json
PeepDaSlan9's picture
Update.config.json
6c4fcb5 verified
raw
history blame contribute delete
657 Bytes
_name_or_path: ./models/dolphin-2_6-phi-2
activation_function: gelu_new
architectures:
- PhiForCausalLM
attn_pdrop: 0.0
auto_map:
AutoConfig: configuration_phi.PhiConfig
AutoModel: modeling_phi.PhiForCausalLM
AutoModelForCausalLM: modeling_phi.PhiForCausalLM
embd_pdrop: 0.0
flash_attn: false
flash_rotary: false
fused_dense: false
img_processor: null
initializer_range: 0.02
layer_norm_epsilon: 1e-05
model_type: phi-msft
n_embd: 2560
n_head: 32
n_head_kv: null
n_inner: null
n_layer: 32
n_positions: 2048
resid_pdrop: 0.1
rotary_dim: 32
tie_word_embeddings: false
torch_dtype: float16
transformers_version: 4.36.2
use_cache: true
vocab_size: 51200