4n3mone commited on
Commit
dd8ea72
1 Parent(s): 25227b5

Upload ChatGLMForConditionalGeneration

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
- library_name: transformers
3
  language:
4
  - ko
 
5
  ---
6
 
7
  # Model Card for Model ID
 
1
  ---
 
2
  language:
3
  - ko
4
+ library_name: transformers
5
  ---
6
 
7
  # Model Card for Model ID
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/workspace/new_korean_llama/src/result/4n3mone/glm-4-ko-9b-chat",
3
  "add_bias_linear": false,
4
  "add_qkv_bias": true,
5
  "apply_query_key_layer_scaling": true,
@@ -44,7 +44,7 @@
44
  "seq_length": 131072,
45
  "tie_word_embeddings": false,
46
  "torch_dtype": "bfloat16",
47
- "transformers_version": "4.40.1",
48
  "use_cache": false,
49
  "vocab_size": 151552
50
  }
 
1
  {
2
+ "_name_or_path": "/workspace/new_korean_llama/src/checkpoints/checkpoint-600",
3
  "add_bias_linear": false,
4
  "add_qkv_bias": true,
5
  "apply_query_key_layer_scaling": true,
 
44
  "seq_length": 131072,
45
  "tie_word_embeddings": false,
46
  "torch_dtype": "bfloat16",
47
+ "transformers_version": "4.41.2",
48
  "use_cache": false,
49
  "vocab_size": 151552
50
  }
generation_config.json CHANGED
@@ -9,5 +9,5 @@
9
  "pad_token_id": 151329,
10
  "temperature": 0.8,
11
  "top_p": 0.8,
12
- "transformers_version": "4.40.1"
13
  }
 
9
  "pad_token_id": 151329,
10
  "temperature": 0.8,
11
  "top_p": 0.8,
12
+ "transformers_version": "4.41.2"
13
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a887940e5c42a8589fe8554dac25e23ebedf9ff44836ac52570a4aa0b197dcf
3
  size 4984147224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64e7837813fb4760a1a02980a5809cbf27fef423994956f67f52355f457f96c8
3
  size 4984147224
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a20f6c8376be09812240d2617e1c6555750d30c5799f2c841c0d9ab5beb9a41
3
  size 4895071360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25b9bd8a1dde48779d341f11e451318090c9c8e8850407e0a16b7cc60b43a123
3
  size 4895071360
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88d08031e5a7c351d207a09a9ed557fb28df038fbc03ea5f2d4034e89c256f61
3
  size 4895071384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:361736b8608890e9af90b90a3a3ec0882259daf5267c9194a771e4735832e9e5
3
  size 4895071384
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e211c47324ba455c4680b973b7f04bd60e9c4c0005e1bf5805c0d6aefa7dd3e9
3
  size 4025651256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87f3f87e9c2f70809a691a8768dfab04a4540549e8ec3cc9144e42e08d416dc2
3
  size 4025651256
modeling_chatglm.py CHANGED
@@ -21,16 +21,20 @@ from transformers.modeling_outputs import (
21
  SequenceClassifierOutputWithPast,
22
  )
23
  from transformers.modeling_utils import PreTrainedModel
24
- from transformers.utils import logging, is_torch_npu_available, is_flash_attn_greater_or_equal_2_10, \
25
- is_flash_attn_2_available
26
  from transformers.generation.logits_process import LogitsProcessor
27
  from transformers.generation.utils import LogitsProcessorList, StoppingCriteriaList, GenerationConfig, ModelOutput
28
 
29
  from .configuration_chatglm import ChatGLMConfig
30
 
31
- if is_flash_attn_2_available():
32
- from flash_attn import flash_attn_func, flash_attn_varlen_func
33
- from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
 
 
 
 
 
34
 
35
  # flags required to enable jit fusion kernels
36
 
 
21
  SequenceClassifierOutputWithPast,
22
  )
23
  from transformers.modeling_utils import PreTrainedModel
24
+ from transformers.utils import logging, is_torch_npu_available
 
25
  from transformers.generation.logits_process import LogitsProcessor
26
  from transformers.generation.utils import LogitsProcessorList, StoppingCriteriaList, GenerationConfig, ModelOutput
27
 
28
  from .configuration_chatglm import ChatGLMConfig
29
 
30
+ try:
31
+ from transformers.utils import is_flash_attn_greater_or_equal_2_10, is_flash_attn_2_available
32
+ if is_flash_attn_2_available():
33
+ from flash_attn import flash_attn_func, flash_attn_varlen_func
34
+ from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
35
+ except:
36
+ pass
37
+
38
 
39
  # flags required to enable jit fusion kernels
40