from transformers import PretrainedConfig from typing import List class AbLangConfig(PretrainedConfig): model_type = "bert" def __init__( self, max_position_embeddings: int=160, hidden_size: int=768, num_hidden_layers: int=12, num_attention_heads: int=12, attention_probs_dropout_prob: float=0.1, intermediate_size: int=3072, hidden_act: str="gelu", hidden_dropout_prob: float=0.1, initializer_range: float=0.02, layer_norm_eps: float=1e-12, chain: str="heavy", **kwargs, ): self.ptid = 21 self.vocab_size=24 self.max_position_embeddings=max_position_embeddings self.hidden_size=hidden_size self.num_hidden_layers=num_hidden_layers self.num_attention_heads=num_attention_heads self.attention_probs_dropout_prob=attention_probs_dropout_prob self.intermediate_size=intermediate_size self.hidden_act=hidden_act self.hidden_dropout_prob=hidden_dropout_prob self.initializer_range=initializer_range self.layer_norm_eps=layer_norm_eps self.chain=chain super().__init__(**kwargs)