File size: 1,356 Bytes
b6d5f2a
 
 
88153b4
b6d5f2a
 
 
 
4309e8e
b6d5f2a
93ed73c
0a0a61e
93ed73c
b6d5f2a
93ed73c
b6d5f2a
 
 
4309e8e
b6d5f2a
 
4309e8e
b6d5f2a
93ed73c
0a0a61e
93ed73c
b6d5f2a
93ed73c
b6d5f2a
 
 
 
4309e8e
b6d5f2a
93ed73c
0a0a61e
93ed73c
b6d5f2a
93ed73c
b6d5f2a
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
from transformers import PretrainedConfig

class BilmaConfig(PretrainedConfig):
    model_type = "bilma"

    def __init__(
        self,
        weights="spanish",
        include_top=True,
        num_attention_heads: int = 4,
        num_hidden_layers: int = 2,
        seq_max_length: int = 280,
        hidden_size: int = 512,
        vocab_size: int = 28949,
        hidden_dropout_prob: float = 0.1,        
        **kwargs,
    ):
        if weights not in ["spanish", ""]:
            raise ValueError(f"`weights` must be 'spanish', got {weights}.")
        if weights == "spanish":
            self.weights = weights
            self.include_top = include_top
            self.num_attention_heads = 4
            self.num_hidden_layers = 2
            self.seq_max_length = 280
            self.hidden_size = 512
            self.vocab_size = 28949
            self.hidden_dropout_prob = 0.1
            super().__init__(**kwargs)
            return

        self.weights = weights
        self.include_top = include_top
        self.num_attention_heads = num_attention_heads
        self.num_hidden_layers = num_hidden_layers
        self.seq_max_length = seq_max_length
        self.hidden_size = hidden_size
        self.vocab_size = vocab_size
        self.hidden_dropout_prob = hidden_dropout_prob        
        super().__init__(**kwargs)