|
_name_or_path: ./models/dolphin-2_6-phi-2 |
|
activation_function: gelu_new |
|
architectures: |
|
- PhiForCausalLM |
|
attn_pdrop: 0.0 |
|
auto_map: |
|
AutoConfig: configuration_phi.PhiConfig |
|
AutoModel: modeling_phi.PhiForCausalLM |
|
AutoModelForCausalLM: modeling_phi.PhiForCausalLM |
|
embd_pdrop: 0.0 |
|
flash_attn: false |
|
flash_rotary: false |
|
fused_dense: false |
|
img_processor: null |
|
initializer_range: 0.02 |
|
layer_norm_epsilon: 1e-05 |
|
model_type: phi-msft |
|
n_embd: 2560 |
|
n_head: 32 |
|
n_head_kv: null |
|
n_inner: null |
|
n_layer: 32 |
|
n_positions: 2048 |
|
resid_pdrop: 0.1 |
|
rotary_dim: 32 |
|
tie_word_embeddings: false |
|
torch_dtype: float16 |
|
transformers_version: 4.36.2 |
|
use_cache: true |
|
vocab_size: 51200 |