File size: 620 Bytes
ed5db8f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 |
from transformers import PretrainedConfig
from typing import List
class ExGPTConfig(PretrainedConfig):
model_type = "ExGPT"
def __init__(
self,
block_size: int = 1024, # Ctx length?
vocab_size: int = 50527, # 50,000 BPE merges + 256 bytes tokens + 1 <|endoftext|> token
n_layer: int = 12,
n_head: int = 12,
n_embd: int = 768,
**kwargs
):
self.block_size = block_size
self.vocab_size = vocab_size
self.n_layer = n_layer
self.n_head = n_head
self.n_embd = n_embd
super().__init__(**kwargs) |