ayjays132 commited on
Commit
d61d2cc
•
1 Parent(s): 0e01298

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -38
README.md CHANGED
@@ -1,15 +1,50 @@
1
  ---
2
- model_type: gpt2
3
  architectures:
4
  - GPT2LMHeadModel
5
- config:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  adaptation_rate: 0.05
7
  complexity_metric: null
8
  desired_improvement_rate: 0.02
9
  ecosystem_dynamics:
10
  environmental_volatility: 0.1
11
  resource_pool: 1
12
- embedding_dim: 768
13
  growth_improvement_threshold: 0.01
14
  hidden_dim: 2048
15
  initial_neuron_count: 5000
@@ -18,49 +53,27 @@ config:
18
  complexity_metric: null
19
  initial_capacity: 250000
20
  input_size: 2048
21
- input_dimension: 768
22
  low_stability_threshold: 0.01
23
- max_complexity: 10000
24
  max_neurons: 250000
25
- max_sequence_length: 200
26
  min_epochs_before_growth: 5
27
  model_filename: pytorch_model.bin
28
- num_embeddings: 25000
29
  pruning_improvement_threshold: 0.005
30
  some_adaptation_rate: 0.05
31
  stability_threshold: 0.02
32
  start_token_index: 2
33
- transformers_version: 4.34.0
34
- attn_pdrop: 0.1
35
- bos_token_id: 50256
36
- embd_pdrop: 0.1
37
- eos_token_id: 50256
38
- initializer_range: 0.02
39
- layer_norm_epsilon: 0.00001
40
- n_ctx: 2048
41
- n_embd: 2048
42
- n_head: 16
43
- n_layer: 24
44
- n_positions: 2048
45
- n_special: 0
46
- predict_special_tokens: true
47
- resid_pdrop: 0.1
48
- summary_activation: null
49
- summary_first_dropout: 0.1
50
- summary_proj_to_labels: true
51
- summary_type: cls_index
52
- summary_use_proj: true
53
- license: apache-2.0
54
- datasets:
55
- - vicgalle/alpaca-gpt4
56
- language:
57
- - en
58
- tags:
59
- - text-generation-inference
60
- metrics:
61
- - accuracy
62
- pipeline_tag: text-generation
63
- library_name: transformers
64
  ---
65
  ---
66
  ## Phillnet Large 🚀
 
1
  ---
2
+ activation_function: gelu_new
3
  architectures:
4
  - GPT2LMHeadModel
5
+ attn_pdrop: 0.1
6
+ bos_token_id: 50256
7
+ embd_pdrop: 0.1
8
+ eos_token_id: 50256
9
+ initializer_range: 0.02
10
+ layer_norm_epsilon: 1e-05
11
+ model_type: gpt2
12
+ n_ctx: 2048
13
+ n_embd: 2048
14
+ n_head: 16
15
+ n_layer: 24
16
+ n_positions: 2048
17
+ n_special: 0
18
+ predict_special_tokens: true
19
+ resid_pdrop: 0.1
20
+ summary_activation: null
21
+ summary_first_dropout: 0.1
22
+ summary_proj_to_labels: true
23
+ summary_type: cls_index
24
+ summary_use_proj: true
25
+ task_specific_params:
26
+ text-generation:
27
+ do_sample: true
28
+ max_length: 2048
29
+ transformers_version: 4.34.0
30
+ language:
31
+ - en
32
+ tags:
33
+ - text-generation-inference
34
+ metrics:
35
+ - accuracy
36
+ pipeline_tag: text-generation
37
+ library_name: transformers
38
+ datasets:
39
+ - vicgalle/alpaca-gpt4
40
+ license: apache-2.0
41
+ custom_params:
42
  adaptation_rate: 0.05
43
  complexity_metric: null
44
  desired_improvement_rate: 0.02
45
  ecosystem_dynamics:
46
  environmental_volatility: 0.1
47
  resource_pool: 1
 
48
  growth_improvement_threshold: 0.01
49
  hidden_dim: 2048
50
  initial_neuron_count: 5000
 
53
  complexity_metric: null
54
  initial_capacity: 250000
55
  input_size: 2048
56
+ input_dimension: 2048
57
  low_stability_threshold: 0.01
58
+ max_complexity: 50000
59
  max_neurons: 250000
60
+ max_sequence_length: 2048
61
  min_epochs_before_growth: 5
62
  model_filename: pytorch_model.bin
63
+ num_embeddings: 100000
64
  pruning_improvement_threshold: 0.005
65
  some_adaptation_rate: 0.05
66
  stability_threshold: 0.02
67
  start_token_index: 2
68
+ max_input_length: 1024
69
+ max_total_tokens: 1512
70
+ max_concurrent_requests: 128
71
+ max_best_of: 2
72
+ max_stop_sequences: 4
73
+ max_top_n_tokens: 5
74
+ waiting_served_ratio: 1.2
75
+ max_batch_prefill_tokens: 2048
76
+ max_waiting_tokens: 20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  ---
78
  ---
79
  ## Phillnet Large 🚀