from transformers import PretrainedConfig class MapperConfig(PretrainedConfig): model_type = "embedding_mapper" def __init__(self, d_in: int = 64, d_hidden: int = 1024, n_layers: int = 6, d_out: int = 64, n_out: int = 2, dropout: float = 0.1, layer_norm_eps: float = 1e-12, **kwargs ): self.d_in = d_in self.d_hidden = d_hidden self.n_layers = n_layers self.d_out = d_out self.n_out = n_out self.dropout = dropout self.layer_norm_eps = layer_norm_eps super().__init__(**kwargs)