Update modeling_diff_llama.py
Browse files- modeling_diff_llama.py +0 -8
modeling_diff_llama.py
CHANGED
@@ -39,14 +39,6 @@ class DiffLLaMAConfig(LlamaConfig):
|
|
39 |
self.rope_scaling = rope_scaling or {"type": "linear", "factor": 1.0}
|
40 |
# Add any custom configuration parameters here
|
41 |
|
42 |
-
@classmethod
|
43 |
-
def from_pretrained(cls, pretrained_model_name_or_path, **kwargs):
|
44 |
-
"""
|
45 |
-
Load configuration from a pretrained model.
|
46 |
-
"""
|
47 |
-
config_dict = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
|
48 |
-
return cls(**config_dict)
|
49 |
-
|
50 |
|
51 |
def init_method(tensor):
|
52 |
"""Initialize tensor with Kaiming uniform initialization."""
|
|
|
39 |
self.rope_scaling = rope_scaling or {"type": "linear", "factor": 1.0}
|
40 |
# Add any custom configuration parameters here
|
41 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
42 |
|
43 |
def init_method(tensor):
|
44 |
"""Initialize tensor with Kaiming uniform initialization."""
|