wei01 commited on
Commit
add9f3d
1 Parent(s): f2e9c01

first commit

Browse files
config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./checkpoints/qwen14b-finetune_all-3m-ocr-448-1024-zero3/checkpoint-4500",
3
+ "architectures": [
4
+ "HuazangQWenForCausalLM"
5
+ ],
6
+ "attn_dropout_prob": 0.0,
7
+ "auto_map": {
8
+ "AutoConfig": "configuration_qwen.QWenConfig",
9
+ "AutoModelForCausalLM": "modeling_qwen.QWenLMHeadModel"
10
+ },
11
+ "bf16": true,
12
+ "emb_dropout_prob": 0.0,
13
+ "fp16": false,
14
+ "fp32": false,
15
+ "hidden_size": 5120,
16
+ "image_aspect_ratio": "pad",
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 27392,
19
+ "kv_channels": 128,
20
+ "layer_norm_epsilon": 1e-06,
21
+ "max_position_embeddings": 8192,
22
+ "mm_hidden_size": 1024,
23
+ "mm_projector_lr": null,
24
+ "mm_projector_type": "mlp2x_gelu",
25
+ "mm_use_im_patch_token": false,
26
+ "mm_use_im_start_end": false,
27
+ "mm_vision_select_feature": "patch",
28
+ "mm_vision_select_layer": -2,
29
+ "mm_vision_tower": "./huazang_sxp/pretrain_weight/qwen_vit_448",
30
+ "model_type": "Huazang_Qwen",
31
+ "no_bias": true,
32
+ "num_attention_heads": 40,
33
+ "num_hidden_layers": 40,
34
+ "onnx_safe": null,
35
+ "rotary_emb_base": 10000,
36
+ "rotary_pct": 1.0,
37
+ "scale_attn_weights": true,
38
+ "seq_length": 2048,
39
+ "softmax_in_fp32": false,
40
+ "tie_word_embeddings": false,
41
+ "tokenizer_class": "QWenTokenizer",
42
+ "tokenizer_model_max_length": 2048,
43
+ "tokenizer_padding_side": "right",
44
+ "torch_dtype": "bfloat16",
45
+ "transformers_version": "4.34.0",
46
+ "tune_mm_mlp_adapter": true,
47
+ "use_cache": true,
48
+ "use_cache_kernel": false,
49
+ "use_cache_quantization": false,
50
+ "use_dynamic_ntk": true,
51
+ "use_flash_attn": true,
52
+ "use_logn_attn": true,
53
+ "use_mm_proj": true,
54
+ "vocab_size": 152064
55
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chat_format": "chatml",
3
+ "do_sample": true,
4
+ "eos_token_id": 151643,
5
+ "max_new_tokens": 2048,
6
+ "max_window_size": 6144,
7
+ "pad_token_id": 151643,
8
+ "top_k": 0,
9
+ "top_p": 0.8,
10
+ "transformers_version": "4.34.0"
11
+ }
mm_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dde34621edf02e129ab42ccb7e7816e7b9bd02260280b9f9ab8a163fcba016a
3
+ size 295040401
pytorch_model-00001-of-00004.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9b4c35aa4ab02144ea37fb7cc9b90178e01da6902b8b9004f501142822632bd
3
+ size 9963536445
pytorch_model-00002-of-00004.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8486db2d1c4adbd5b44e03eca48cd3266b26ba717aab215fb6ec21377b20cd0a
3
+ size 9878405831
pytorch_model-00003-of-00004.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da1adf96deefa5a251afe38cd5fe07c3d14edd6d2de25526d3c58a94de6350db
3
+ size 9980282713
pytorch_model-00004-of-00004.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fed7074c0709dc7af38d85e468e6d8d6ad60cfdd2a9259c633a202f998bd2da
3
+ size 2684467664
pytorch_model.bin.index.json ADDED
The diff for this file is too large to render. See raw diff
 
qwen.tiktoken ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "pad_token": "<|endoftext|>"
3
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {},
3
+ "additional_special_tokens": [],
4
+ "auto_map": {
5
+ "AutoTokenizer": [
6
+ "tokenization_qwen.QWenTokenizer",
7
+ null
8
+ ]
9
+ },
10
+ "clean_up_tokenization_spaces": true,
11
+ "model_max_length": 2048,
12
+ "pad_token": "<|endoftext|>",
13
+ "padding_side": "right",
14
+ "tokenizer_class": "QWenTokenizer",
15
+ "tokenizer_file": null,
16
+ "use_fast": false
17
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8577a1072be38aa38801cb46d1a8e8a39148a8322caad8fed0cdcb369bf21f04
3
+ size 6139