Upload BlenderbotForConditionalGeneration
Browse files- config.json +1 -3
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"activation_dropout": 0.0,
|
4 |
"activation_function": "gelu",
|
5 |
"add_bias_logits": false,
|
@@ -56,11 +56,9 @@
|
|
56 |
"pad_token_id": 0,
|
57 |
"scale_embedding": true,
|
58 |
"static_position_embeddings": false,
|
59 |
-
"task": "conversational",
|
60 |
"torch_dtype": "float32",
|
61 |
"transformers_version": "4.46.2",
|
62 |
"unk_token_id": 3,
|
63 |
"use_cache": true,
|
64 |
-
"use_safetensors": true,
|
65 |
"vocab_size": 8009
|
66 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/content/drive/My Drive/CMPE258DL/fine_tuned_blenderbot_1_epoch_temp-0.4/",
|
3 |
"activation_dropout": 0.0,
|
4 |
"activation_function": "gelu",
|
5 |
"add_bias_logits": false,
|
|
|
56 |
"pad_token_id": 0,
|
57 |
"scale_embedding": true,
|
58 |
"static_position_embeddings": false,
|
|
|
59 |
"torch_dtype": "float32",
|
60 |
"transformers_version": "4.46.2",
|
61 |
"unk_token_id": 3,
|
62 |
"use_cache": true,
|
|
|
63 |
"vocab_size": 8009
|
64 |
}
|