File size: 579 Bytes
0ae6e17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
from .version_check import check_transformers_version

check_transformers_version()

from .transformers_v4_35_2__configuration_llama import LlamaConfig


class DeciLMConfig(LlamaConfig):
    r"""
    Args:
        num_key_value_heads_per_layer (`List[int]`):
            The number of key-value heads per layer.
    """
    model_type = "deci_lm"

    def __init__(
            self,
            num_key_value_heads_per_layer: list = None,
            **kwargs,
    ):
        self.num_key_value_heads_per_layer = num_key_value_heads_per_layer
        super().__init__(**kwargs)