File size: 1,356 Bytes
b6d5f2a 88153b4 b6d5f2a 4309e8e b6d5f2a 93ed73c 0a0a61e 93ed73c b6d5f2a 93ed73c b6d5f2a 4309e8e b6d5f2a 4309e8e b6d5f2a 93ed73c 0a0a61e 93ed73c b6d5f2a 93ed73c b6d5f2a 4309e8e b6d5f2a 93ed73c 0a0a61e 93ed73c b6d5f2a 93ed73c b6d5f2a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 |
from transformers import PretrainedConfig
class BilmaConfig(PretrainedConfig):
model_type = "bilma"
def __init__(
self,
weights="spanish",
include_top=True,
num_attention_heads: int = 4,
num_hidden_layers: int = 2,
seq_max_length: int = 280,
hidden_size: int = 512,
vocab_size: int = 28949,
hidden_dropout_prob: float = 0.1,
**kwargs,
):
if weights not in ["spanish", ""]:
raise ValueError(f"`weights` must be 'spanish', got {weights}.")
if weights == "spanish":
self.weights = weights
self.include_top = include_top
self.num_attention_heads = 4
self.num_hidden_layers = 2
self.seq_max_length = 280
self.hidden_size = 512
self.vocab_size = 28949
self.hidden_dropout_prob = 0.1
super().__init__(**kwargs)
return
self.weights = weights
self.include_top = include_top
self.num_attention_heads = num_attention_heads
self.num_hidden_layers = num_hidden_layers
self.seq_max_length = seq_max_length
self.hidden_size = hidden_size
self.vocab_size = vocab_size
self.hidden_dropout_prob = hidden_dropout_prob
super().__init__(**kwargs) |