Upload ChatGLMForConditionalGeneration
Browse files- README.md +1 -1
- config.json +2 -2
- generation_config.json +1 -1
- model-00001-of-00004.safetensors +1 -1
- model-00002-of-00004.safetensors +1 -1
- model-00003-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +1 -1
- modeling_chatglm.py +9 -5
README.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
---
|
2 |
-
library_name: transformers
|
3 |
language:
|
4 |
- ko
|
|
|
5 |
---
|
6 |
|
7 |
# Model Card for Model ID
|
|
|
1 |
---
|
|
|
2 |
language:
|
3 |
- ko
|
4 |
+
library_name: transformers
|
5 |
---
|
6 |
|
7 |
# Model Card for Model ID
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "/workspace/new_korean_llama/src/
|
3 |
"add_bias_linear": false,
|
4 |
"add_qkv_bias": true,
|
5 |
"apply_query_key_layer_scaling": true,
|
@@ -44,7 +44,7 @@
|
|
44 |
"seq_length": 131072,
|
45 |
"tie_word_embeddings": false,
|
46 |
"torch_dtype": "bfloat16",
|
47 |
-
"transformers_version": "4.
|
48 |
"use_cache": false,
|
49 |
"vocab_size": 151552
|
50 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/workspace/new_korean_llama/src/checkpoints/checkpoint-600",
|
3 |
"add_bias_linear": false,
|
4 |
"add_qkv_bias": true,
|
5 |
"apply_query_key_layer_scaling": true,
|
|
|
44 |
"seq_length": 131072,
|
45 |
"tie_word_embeddings": false,
|
46 |
"torch_dtype": "bfloat16",
|
47 |
+
"transformers_version": "4.41.2",
|
48 |
"use_cache": false,
|
49 |
"vocab_size": 151552
|
50 |
}
|
generation_config.json
CHANGED
@@ -9,5 +9,5 @@
|
|
9 |
"pad_token_id": 151329,
|
10 |
"temperature": 0.8,
|
11 |
"top_p": 0.8,
|
12 |
-
"transformers_version": "4.
|
13 |
}
|
|
|
9 |
"pad_token_id": 151329,
|
10 |
"temperature": 0.8,
|
11 |
"top_p": 0.8,
|
12 |
+
"transformers_version": "4.41.2"
|
13 |
}
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4984147224
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:64e7837813fb4760a1a02980a5809cbf27fef423994956f67f52355f457f96c8
|
3 |
size 4984147224
|
model-00002-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4895071360
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:25b9bd8a1dde48779d341f11e451318090c9c8e8850407e0a16b7cc60b43a123
|
3 |
size 4895071360
|
model-00003-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4895071384
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:361736b8608890e9af90b90a3a3ec0882259daf5267c9194a771e4735832e9e5
|
3 |
size 4895071384
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4025651256
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:87f3f87e9c2f70809a691a8768dfab04a4540549e8ec3cc9144e42e08d416dc2
|
3 |
size 4025651256
|
modeling_chatglm.py
CHANGED
@@ -21,16 +21,20 @@ from transformers.modeling_outputs import (
|
|
21 |
SequenceClassifierOutputWithPast,
|
22 |
)
|
23 |
from transformers.modeling_utils import PreTrainedModel
|
24 |
-
from transformers.utils import logging, is_torch_npu_available
|
25 |
-
is_flash_attn_2_available
|
26 |
from transformers.generation.logits_process import LogitsProcessor
|
27 |
from transformers.generation.utils import LogitsProcessorList, StoppingCriteriaList, GenerationConfig, ModelOutput
|
28 |
|
29 |
from .configuration_chatglm import ChatGLMConfig
|
30 |
|
31 |
-
|
32 |
-
from
|
33 |
-
|
|
|
|
|
|
|
|
|
|
|
34 |
|
35 |
# flags required to enable jit fusion kernels
|
36 |
|
|
|
21 |
SequenceClassifierOutputWithPast,
|
22 |
)
|
23 |
from transformers.modeling_utils import PreTrainedModel
|
24 |
+
from transformers.utils import logging, is_torch_npu_available
|
|
|
25 |
from transformers.generation.logits_process import LogitsProcessor
|
26 |
from transformers.generation.utils import LogitsProcessorList, StoppingCriteriaList, GenerationConfig, ModelOutput
|
27 |
|
28 |
from .configuration_chatglm import ChatGLMConfig
|
29 |
|
30 |
+
try:
|
31 |
+
from transformers.utils import is_flash_attn_greater_or_equal_2_10, is_flash_attn_2_available
|
32 |
+
if is_flash_attn_2_available():
|
33 |
+
from flash_attn import flash_attn_func, flash_attn_varlen_func
|
34 |
+
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
|
35 |
+
except:
|
36 |
+
pass
|
37 |
+
|
38 |
|
39 |
# flags required to enable jit fusion kernels
|
40 |
|