This view is limited to 50 files because it contains too many changes.  See the raw diff here.
Files changed (50) hide show
  1. README.md +6 -7
  2. config.json +0 -40
  3. configuration.json +0 -1
  4. configuration_telechat.py +0 -94
  5. generation_config.json +0 -14
  6. generation_utils.py +0 -162
  7. model-00001-of-00050.safetensors +0 -3
  8. model-00002-of-00050.safetensors +0 -3
  9. model-00003-of-00050.safetensors +0 -3
  10. model-00004-of-00050.safetensors +0 -3
  11. model-00005-of-00050.safetensors +0 -3
  12. model-00006-of-00050.safetensors +0 -3
  13. model-00007-of-00050.safetensors +0 -3
  14. model-00008-of-00050.safetensors +0 -3
  15. model-00009-of-00050.safetensors +0 -3
  16. model-00010-of-00050.safetensors +0 -3
  17. model-00011-of-00050.safetensors +0 -3
  18. model-00012-of-00050.safetensors +0 -3
  19. model-00013-of-00050.safetensors +0 -3
  20. model-00014-of-00050.safetensors +0 -3
  21. model-00015-of-00050.safetensors +0 -3
  22. model-00016-of-00050.safetensors +0 -3
  23. model-00017-of-00050.safetensors +0 -3
  24. model-00018-of-00050.safetensors +0 -3
  25. model-00019-of-00050.safetensors +0 -3
  26. model-00020-of-00050.safetensors +0 -3
  27. model-00021-of-00050.safetensors +0 -3
  28. model-00022-of-00050.safetensors +0 -3
  29. model-00023-of-00050.safetensors +0 -3
  30. model-00024-of-00050.safetensors +0 -3
  31. model-00025-of-00050.safetensors +0 -3
  32. model-00026-of-00050.safetensors +0 -3
  33. model-00027-of-00050.safetensors +0 -3
  34. model-00028-of-00050.safetensors +0 -3
  35. model-00029-of-00050.safetensors +0 -3
  36. model-00030-of-00050.safetensors +0 -3
  37. model-00031-of-00050.safetensors +0 -3
  38. model-00032-of-00050.safetensors +0 -3
  39. model-00033-of-00050.safetensors +0 -3
  40. model-00034-of-00050.safetensors +0 -3
  41. model-00035-of-00050.safetensors +0 -3
  42. model-00036-of-00050.safetensors +0 -3
  43. model-00037-of-00050.safetensors +0 -3
  44. model-00038-of-00050.safetensors +0 -3
  45. model-00039-of-00050.safetensors +0 -3
  46. model-00040-of-00050.safetensors +0 -3
  47. model-00041-of-00050.safetensors +0 -3
  48. model-00042-of-00050.safetensors +0 -3
  49. model-00043-of-00050.safetensors +0 -3
  50. model-00044-of-00050.safetensors +0 -3
README.md CHANGED
@@ -164,13 +164,12 @@ [email protected],提交《TeleChat模型社区许可协议》要求的
164
  如需引用我们的工作,请使用如下 reference:
165
 
166
  ```
167
- @misc{wang2025technicalreporttelechat2telechat25,
168
- title={Technical Report of TeleChat2, TeleChat2.5 and T1},
169
- author={Zihan Wang and Xinzhang Liu and Yitong Yao and Chao Wang and Yu Zhao and Zhihao Yang and Wenmin Deng and Kaipeng Jia and Jiaxin Peng and Yuyao Huang and Sishi Xiong and Zhuo Jiang and Kaidong Yu and Xiaohui Hu and Fubei Yao and Ruiyu Fang and Zhuoru Jiang and Ruiting Song and Qiyi Xie and Rui Xue and Xuewei He and Yanlei Xue and Zhu Yuan and Zhaoxi Zhang and Zilu Huang and Shiquan Wang and Xin Wang and Hanming Wu and Mingyuan Wang and Xufeng Zhan and Yuhan Sun and Zhaohu Xing and Yuhao Jiang and Bingkai Yang and Shuangyong Song and Yongxiang Li and Zhongjiang He and Xuelong Li},
170
- year={2025},
171
- eprint={2507.18013},
172
  archivePrefix={arXiv},
173
- primaryClass={cs.CL},
174
- url={https://arxiv.org/abs/2507.18013},
175
  }
176
  ```
 
164
  如需引用我们的工作,请使用如下 reference:
165
 
166
  ```
167
+ @misc{wang2024telechat,
168
+ title={TeleChat Technical Report},
169
+ author={Zihan Wang and Xinzhang Liu and Shixuan Liu and Yitong Yao and Yuyao Huang and Zhongjiang He and Xuelong Li and Yongxiang Li and Zhonghao Che and Zhaoxi Zhang and Yan Wang and Xin Wang and Luwen Pu and Huihan Xu and Ruiyu Fang and Yu Zhao and Jie Zhang and Xiaomeng Huang and Zhilong Lu and Jiaxin Peng and Wenjun Zheng and Shiquan Wang and Bingkai Yang and Xuewei he and Zhuoru Jiang and Qiyi Xie and Yanhan Zhang and Zhongqiu Li and Lingling Shi and Weiwei Fu and Yin Zhang and Zilu Huang and Sishi Xiong and Yuxiang Zhang and Chao Wang and Shuangyong Song},
170
+ year={2024},
171
+ eprint={2401.03804},
172
  archivePrefix={arXiv},
173
+ primaryClass={cs.CL}
 
174
  }
175
  ```
config.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "apply_residual_connection_post_layernorm": false,
3
- "architectures": [
4
- "TelechatForCausalLM"
5
- ],
6
- "auto_map": {
7
- "AutoConfig": "configuration_telechat.TelechatConfig",
8
- "AutoModelForCausalLM": "modeling_telechat.TelechatForCausalLM"
9
- },
10
- "attention_dropout": 0.0,
11
- "attention_softmax_in_fp32": true,
12
- "bias_dropout_fusion": true,
13
- "bos_token_id": 1,
14
- "eos_token_id": 2,
15
- "hidden_dropout": 0.0,
16
- "hidden_size": 8192,
17
- "initializer_range": 0.02,
18
- "layer_norm_epsilon": 1e-05,
19
- "masked_softmax_fusion": true,
20
- "model_type": "telechat",
21
- "n_head": 64,
22
- "n_inner": null,
23
- "num_key_value_heads": 8,
24
- "n_layer": 96,
25
- "pad_token_id": 3,
26
- "pretraining_tp": 2,
27
- "skip_bias_add": false,
28
- "skip_bias_add_qkv": false,
29
- "slow_but_exact": false,
30
- "unk_token_id": 0,
31
- "use_cache": true,
32
- "vocab_size": 131072,
33
- "ffn_hidden_size": 40960,
34
- "flash_attn":true,
35
- "tie_word_embeddings":false,
36
- "training_seqlen":8192,
37
- "base_seqlen":8192,
38
- "seq_length": 8192
39
- }
40
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
configuration.json DELETED
@@ -1 +0,0 @@
1
- {"task":"text-generation"}
 
 
configuration_telechat.py DELETED
@@ -1,94 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2022 the Big Science Workshop and HuggingFace Inc. team. All rights reserved.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- """ Telechat configuration"""
17
-
18
- from packaging import version
19
- from collections import OrderedDict
20
- from transformers.utils import is_torch_available, logging
21
- from transformers.configuration_utils import PretrainedConfig
22
- from typing import TYPE_CHECKING, Any, List, Mapping, Optional
23
-
24
- logger = logging.get_logger(__name__)
25
-
26
- class TelechatConfig(PretrainedConfig):
27
- """
28
- Args:
29
- vocab_size (`int`, *optional*, defaults to 160256): Vocabulary size of the Telechat model.
30
- hidden_size (`int`, *optional*, defaults to 4096): Dimensionality of the embeddings and hidden states.
31
- ffn_hidden_size (`int`, *optional*, defaults to 12288): Dimensionality of the feed-forward hidden states.
32
- n_layer (`int`, *optional*, defaults to 30): Number of hidden layers in the Transformer
33
- n_head (`int`, *optional*, defaults to 32): Number of attention heads for each attention layer.
34
- layer_norm_epsilon (`float`, *optional*, defaults to 1e-5): The epsilon to use in the layer normalization layers.
35
- initializer_range (`float`, *optional*, defaults to 0.02): The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
36
- apply_residual_connection_post_layernorm (`bool`, *optional*, defaults to `False`): If enabled, use the layer norm of the hidden states as the residual in the transformer blocks
37
- hidden_dropout (`float`, *optional*, defaults to 0.0): Dropout rate of the dropout function on the bias dropout.
38
- attention_dropout (`float`, *optional*, defaults to 0.0): Dropout rate applied to the attention probs
39
- use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions.
40
- training_seqlen (`int`, *optional*, defaults to 8192): Sequence length during last finetuning.
41
- logn (`bool`, *optional*, defaults to `True`): Whether or not to use logN during extrapolation.
42
- embed_layernorm (`bool`, *optional*, defaults to `True`): Whether or not to use embedding layernorm.
43
-
44
- """
45
-
46
- model_type = "telechat"
47
- keys_to_ignore_at_inference = ["past_key_values"]
48
- attribute_map = {
49
- "num_hidden_layers": "n_layer",
50
- "num_attention_heads": "n_head",
51
- }
52
-
53
- def __init__(
54
- self,
55
- vocab_size=160256,
56
- hidden_size=4096,
57
- n_layer=30,
58
- n_head=32,
59
- layer_norm_epsilon=1e-5,
60
- initializer_range=0.02,
61
- use_cache=True,
62
- bos_token_id=1,
63
- eos_token_id=2,
64
- apply_residual_connection_post_layernorm=False,
65
- hidden_dropout=0.0,
66
- attention_dropout=0.0,
67
- ffn_hidden_size=12288,
68
- training_seqlen = 8192,
69
- logn = True,
70
- embed_layernorm = False,
71
- **kwargs,
72
- ):
73
- self.vocab_size = vocab_size
74
- n_embed = kwargs.pop("n_embed", None)
75
- self.hidden_size = hidden_size if n_embed is None else n_embed
76
- self.n_layer = n_layer
77
- self.n_head = n_head
78
- self.layer_norm_epsilon = layer_norm_epsilon
79
- self.initializer_range = initializer_range
80
- self.use_cache = use_cache
81
- self.apply_residual_connection_post_layernorm = apply_residual_connection_post_layernorm
82
- self.hidden_dropout = hidden_dropout
83
- self.attention_dropout = attention_dropout
84
- self.bos_token_id = bos_token_id
85
- self.eos_token_id = eos_token_id
86
- self.logn = logn
87
- self.ffn_hidden_size = ffn_hidden_size
88
- self.training_seqlen = training_seqlen
89
- self.embed_layernorm = embed_layernorm
90
- self.num_key_value_heads= kwargs.pop("num_key_value_heads", None)
91
-
92
-
93
- super().__init__(bos_token_id=bos_token_id, eos_token_id=eos_token_id, **kwargs)
94
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generation_config.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "max_length": 8192,
3
- "do_sample": false,
4
- "use_cache": true,
5
- "temperature": 0.3,
6
- "top_k": 5,
7
- "top_p": 0.85,
8
- "repetition_penalty": 1.03,
9
- "pad_token_id": 3,
10
- "bos_token_id": 1,
11
- "eos_token_id": 2,
12
- "user_token_id": 4,
13
- "bot_token_id": 5
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generation_utils.py DELETED
@@ -1,162 +0,0 @@
1
- from typing import Optional
2
- from collections import deque
3
- from queue import Queue
4
- import copy
5
-
6
-
7
- class History:
8
-
9
- def __init__(self, tokenizer, history):
10
- '''
11
- init from a list of dict
12
- '''
13
- # use deque to meet some special situation
14
- self.input_history = deque()
15
- self.tokenizer = tokenizer
16
- if history:
17
- self._transfer_from_list(history)
18
-
19
- def _transfer_from_list(self, history):
20
- for message in history:
21
- content = message.get("content")
22
- # the token result may not be equal to the result model gen
23
- message.update(self.tokenizer(content))
24
- self.input_history.append(message)
25
-
26
- def append(self, message):
27
- content = message.get("content")
28
- if "input_ids" not in message or "attention_mask" not in message:
29
- message.update(self.tokenizer(content))
30
- self.input_history.append(message)
31
-
32
- def append_left(self, message):
33
- content = message.get("content")
34
- if "input_ids" not in message or "attention_mask" not in message:
35
- message.update(self.tokenizer(content))
36
- self.input_history.appendleft(message)
37
-
38
- def pop(self):
39
- x = self.input_history.pop()
40
- return x
41
-
42
- def pop_left(self):
43
- x = self.input_history.pop_left()
44
- return x
45
-
46
- def update(self, message):
47
- self.input_history.pop()
48
- self.append(message)
49
-
50
- def __len__(self):
51
- return self.input_history.__len__()
52
-
53
- def __str__(self):
54
- return self.input_history.__str__()
55
-
56
- def __copy__(self):
57
- new_instance = type(self)(self.tokenizer, [])
58
- new_instance.input_history = copy.copy(self.input_history)
59
- return new_instance
60
-
61
- def __deepcopy__(self, memodict={}):
62
- new_instance = type(self)(self.tokenizer, [])
63
- new_instance.input_history = copy.deepcopy(self.input_history)
64
- return new_instance
65
-
66
-
67
- class TelechatIterTextStreamer:
68
- """
69
- With reference to the TextIterStreamers in transformers, we have rewritten this class
70
- """
71
-
72
- def __init__(
73
- self, tokenizer, history: History = None, skip_prompt: bool = False, timeout: Optional[float] = None,
74
- **decode_kwargs
75
- ):
76
-
77
- self.tokenizer = tokenizer
78
- self.history = history
79
- self.skip_prompt = skip_prompt
80
- self.timeout = timeout
81
- self.decode_kwargs = decode_kwargs
82
-
83
- self.text_queue = Queue()
84
- self.cache_time = 0
85
- self.text_until = ""
86
- self.token_until = []
87
- self.stop_signal = None
88
- self.next_tokens_are_prompt = True
89
-
90
- self.history.append({"role": "bot", "content": self.text_until})
91
-
92
- def put(self, value):
93
- """
94
- put printable text into queue
95
- """
96
- if len(value.shape) > 1 and value.shape[0] > 1:
97
- raise ValueError("TextStreamer only supports batch size 1")
98
- elif len(value.shape) > 1:
99
- value = value[0]
100
-
101
- if self.skip_prompt and self.next_tokens_are_prompt:
102
- self.next_tokens_are_prompt = False
103
- return
104
-
105
- if value[-1] == self.tokenizer.eos_token_id:
106
- return
107
-
108
- # there may be some smart way to decode.
109
- self.token_until.extend(value.tolist())
110
- text = self.tokenizer.decode(self.token_until, **self.decode_kwargs)
111
-
112
-
113
- if self._is_printable(text) or self.cache_time >= 6:
114
- output_text = text[len(self.text_until):]
115
- self.text_until = text
116
-
117
- else:
118
- self.cache_time+=1
119
- return
120
-
121
- self.on_finalized_text(output_text)
122
-
123
- def end(self):
124
- """Flushes any remaining cache and prints a newline to stdout."""
125
- # Flush the cache, if it exists
126
- text = self.tokenizer.decode(self.token_until, **self.decode_kwargs)
127
- output_text = text[len(self.text_until):]
128
- self.text_until = text
129
- self.on_finalized_text(output_text, stream_end=True)
130
- self.clear_cache()
131
-
132
- def clear_cache(self):
133
- self.cache_time = 0
134
- self.token_until = []
135
- self.text_until = ""
136
- self.history = None
137
- self.next_tokens_are_prompt = True
138
-
139
- def on_finalized_text(self, text: str, stream_end: bool = False):
140
- """Put the text tuple in the queue."""
141
- self.history.update({"role": "bot", "content": self.text_until, "input_ids": self.token_until,
142
- "attention_mask": [1] * len(self.token_until)})
143
- self.text_queue.put((text, self.history), timeout=self.timeout)
144
- if stream_end:
145
- self.text_queue.put((self.stop_signal, self.history), timeout=self.timeout)
146
-
147
- @staticmethod
148
- def _is_printable(cp):
149
- """Checks whether tokens can be decoded or not"""
150
- if "�" in cp:
151
- return False
152
- return True
153
-
154
- def __iter__(self):
155
- return self
156
-
157
- def __next__(self):
158
- value_now, history_until = self.text_queue.get(timeout=self.timeout)
159
- if value_now == self.stop_signal:
160
- raise StopIteration()
161
- else:
162
- return value_now, history_until
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
model-00001-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:50a4549e91bf98dcb213f6455f92440454fdd012cd0369e0fdb2f019c83fe4c7
3
- size 8925611272
 
 
 
 
model-00002-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:548f67873b1f538bf325590f30dbe267600356ccd1bef73c1f1cc7a2902c2aa7
3
- size 9261287712
 
 
 
 
model-00003-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e855d3528914e522dcad5348c2229dc2e88fa55e87416f7d5106437d3679a8d3
3
- size 9261287712
 
 
 
 
model-00004-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:509b7084b8dcb308c7c7d22fa988e805b6f6947e356526f60cccdd40ceb2a11d
3
- size 9261287712
 
 
 
 
model-00005-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb044a2e88e6e7a0817459f7f409eeb5abd00836346f69f8ebfd9d0c47028ebf
3
- size 9261287712
 
 
 
 
model-00006-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3e410d45493b53ba09cdb8d368ecad07df8c1db96a58d928cd7045c8ed7a48f
3
- size 9261287728
 
 
 
 
model-00007-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9cdcab0486ef69e5eedbb99b08f0b55c918a83f227194cb8b8937f010c950a7
3
- size 9261287736
 
 
 
 
model-00008-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2a7cf122475960568ea086f9b98b2283dbf9344dda2720726a52bff7f31edbc
3
- size 9261287736
 
 
 
 
model-00009-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc5d19f0add344ac71fb3855ca8f04e588fd4d5263e05709ec4dd668ebcf8ee4
3
- size 9261287736
 
 
 
 
model-00010-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d9763dc2004ea25ca751d271f92a1cce636f92c6991077ea391788428cfc6c1
3
- size 9261287736
 
 
 
 
model-00011-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca7e26c5ed4c0fedec09a4cceef1ac7e03d037c921c18d1df0b782ca4fe7b71b
3
- size 9261287736
 
 
 
 
model-00012-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0390b0e55acc47cebac2bc81e57e1e7d3f02f72a0cad2b01b5165caf976ab0b
3
- size 9261287736
 
 
 
 
model-00013-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbb3ccd22fc048c59e67dab8df54264acf53d7d62fc98fb49ffeafe7ea181739
3
- size 9261287736
 
 
 
 
model-00014-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e2c53dae773a55d42d3265d0d85ae02c8f6f70543a90d923823cab68d44d5ce
3
- size 9261287736
 
 
 
 
model-00015-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b813a2adab5ff776a06986e60d22e917c596122c417a7d6ae1930b8ed5d61a4f
3
- size 9261287736
 
 
 
 
model-00016-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:33976aac9f40e0ff430435fe0295ba1f4443dc4b78233c077de95453eff51d01
3
- size 9261287736
 
 
 
 
model-00017-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:096850a1c187e804c370956830ef94d20731f79ee1078fa0613604c93015d305
3
- size 9261287736
 
 
 
 
model-00018-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7ed47b73f1171ee30de651924ab58aaa7088037b9f96203be23b4b2514dfb27
3
- size 9261287736
 
 
 
 
model-00019-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe1e87f96e41804e06c966605e6b00e0a9c6b51c504ee5e22dfdb28e90d1308f
3
- size 9261287736
 
 
 
 
model-00020-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:93bc200e651f67d4910170de85235af15b3228b23192afb5c32d5126daaaf996
3
- size 9261287736
 
 
 
 
model-00021-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:645df22333bb21b8cef574bdadcaecaf0efcf43d3e747a0b0c0d462740d0b7c4
3
- size 9261287736
 
 
 
 
model-00022-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a288d0dafad8ca11e21c869a31d62057f6bc13036934e306cb68cd7fdd5861f
3
- size 9261287736
 
 
 
 
model-00023-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:94e110e13a08711bfd02cbb3d20e32baecd675f2593a8ae66ba1fcf039e34073
3
- size 9261287736
 
 
 
 
model-00024-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:56894143f2c2a3ff2af49e7f3ec2af9929ff8d95fe7e8e63b709280ba4babe52
3
- size 9261287736
 
 
 
 
model-00025-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:72d4aee125d0a41ceb2afc164c78d25b9edc89cf845da5f91e9c60b743b2f3a7
3
- size 9261287736
 
 
 
 
model-00026-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:72fdd61ca5c065b386092a0701f6fe8ed7ae8d99b0eaba87e50e60acff058876
3
- size 9261287736
 
 
 
 
model-00027-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:764f995de7385ab9c14e7393ee6adc73b557c1a182efdcbbf143c3fe22f353c3
3
- size 9261287736
 
 
 
 
model-00028-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba9bd63f6e97a17e03b467629d9a4bafd581da1b9445b134423c75d5aa409c08
3
- size 9261287736
 
 
 
 
model-00029-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:21c3f8b212fef1678419b95693121290430930dd2930bd531aa7a3faa40fac41
3
- size 9261287736
 
 
 
 
model-00030-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b861e871ac1d1ae14b6fcadaefe876c45d694f715eb91d68bacf114c2c426adf
3
- size 9261287736
 
 
 
 
model-00031-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:495ecf892bce401191ce699032b3d95b957d935987ee462df76de2b5928c1127
3
- size 9261287736
 
 
 
 
model-00032-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2613dc2c31320d4c300e260aac5a313dc183b8fdd140f6b3aea338667126ffae
3
- size 9261287736
 
 
 
 
model-00033-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:67348068c452d63d6ea33f468f810157993b1431c823fc4fc68d1e4c7cf1b653
3
- size 9261287736
 
 
 
 
model-00034-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e28056f9391f029ddea9743445dbf763c265feb31a184ba73a7e8de91280d981
3
- size 9261287736
 
 
 
 
model-00035-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c966308e58af945da780ddf143e6f436f21a8019ddb1824200822ac8ecf72c8
3
- size 9261287736
 
 
 
 
model-00036-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:998076e5d150bea2630b24af7bf1da793a951f5c5b9ddfeb7b9400eb1188eea6
3
- size 9261287736
 
 
 
 
model-00037-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1facd7663e7db14ec43876cc727f3d8e75b40bb025cb7a25fb7d2f9a60de68f9
3
- size 9261287736
 
 
 
 
model-00038-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c06aafaf597e562ebce7672650df8ae3fcabf22977c4d47004f817ca4f8fe536
3
- size 9261287736
 
 
 
 
model-00039-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3bda7f14ffa81266bd9ec07e05e9c3bc16726f8fcd75281d4af4002508bfceb
3
- size 9261287736
 
 
 
 
model-00040-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf19d0fbf517593390ccc10b1c00c1410b54286ca5c651c1fe3ee8de7b02c4d1
3
- size 9261287736
 
 
 
 
model-00041-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f98f5ea063046592be25906b7dc84f0096e67f104f1bcd847858ee778ba4e1eb
3
- size 9261287736
 
 
 
 
model-00042-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:033ba66aecb88ed950b1c49bd2c6f6d5e5173ed4882b22ff189f4caef542ec9f
3
- size 9261287736
 
 
 
 
model-00043-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a46195067f9f791eee221c2195d30136f0f8eb723dd0c1674d68c6c2409fa68
3
- size 9261287736
 
 
 
 
model-00044-of-00050.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5becd94d0a4e65b2b0496f21ca679ba4b171ffb32851eabd15af17004bda901
3
- size 9261287736