mmaaz60 commited on
Commit
8dc70ee
1 Parent(s): a74ff76

Updates the full-scope model - better performance across all tasks.

Browse files
config.json CHANGED
@@ -1,17 +1,20 @@
1
  {
2
- "_name_or_path": "./trained_checkpoints/exp1_grand_pretrain_v15_gs4488_ckpt_25000_hf",
3
  "architectures": [
4
- "LISAForCausalLM"
5
  ],
6
  "bbox_token_idx": 32002,
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
- "freeze_mm_mlp_adapter": true,
 
10
  "freeze_mm_vision_resampler": false,
11
  "hidden_act": "silu",
12
  "hidden_size": 4096,
 
13
  "image_aspect_ratio": "square",
14
  "image_grid_pinpoints": null,
 
15
  "initializer_range": 0.02,
16
  "intermediate_size": 11008,
17
  "max_length": 4096,
@@ -21,6 +24,8 @@
21
  "mm_resampler_type": null,
22
  "mm_use_im_patch_token": false,
23
  "mm_use_im_start_end": true,
 
 
24
  "mm_vision_select_feature": "patch",
25
  "mm_vision_select_layer": -2,
26
  "mm_vision_tower": "openai/clip-vit-large-patch14-336",
@@ -29,21 +34,26 @@
29
  "num_hidden_layers": 32,
30
  "num_key_value_heads": 32,
31
  "num_level_reg_features": 4,
 
32
  "out_dim": 256,
33
  "pad_token_id": 0,
34
  "pretrain_mm_mlp_adapter": null,
35
  "pretraining_tp": 1,
36
  "rms_norm_eps": 1e-05,
37
  "rope_scaling": null,
 
38
  "tie_word_embeddings": false,
39
  "torch_dtype": "bfloat16",
40
  "train_mask_decoder": true,
41
  "transformers_version": "4.28.0.dev0",
 
42
  "tune_mm_mlp_adapter": false,
43
  "tune_mm_vision_resampler": false,
44
  "unfreeze_mm_vision_tower": false,
45
  "use_cache": false,
 
46
  "use_mm_proj": true,
 
47
  "vision_tower": "openai/clip-vit-large-patch14-336",
48
  "vocab_size": 32007,
49
  "with_region": true
 
1
  {
2
+ "_name_or_path": "./checkpoints/GLaMM-GranD-Pretrained",
3
  "architectures": [
4
+ "GLaMMForCausalLM"
5
  ],
6
  "bbox_token_idx": 32002,
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
+ "freeze_mlp_adapter": true,
10
+ "freeze_mm_mlp_adapter": false,
11
  "freeze_mm_vision_resampler": false,
12
  "hidden_act": "silu",
13
  "hidden_size": 4096,
14
+ "image_aspect": "square",
15
  "image_aspect_ratio": "square",
16
  "image_grid_pinpoints": null,
17
+ "image_grid_points": null,
18
  "initializer_range": 0.02,
19
  "intermediate_size": 11008,
20
  "max_length": 4096,
 
24
  "mm_resampler_type": null,
25
  "mm_use_im_patch_token": false,
26
  "mm_use_im_start_end": true,
27
+ "mm_use_image_start_end": true,
28
+ "mm_vision_module": "openai/clip-vit-large-patch14-336",
29
  "mm_vision_select_feature": "patch",
30
  "mm_vision_select_layer": -2,
31
  "mm_vision_tower": "openai/clip-vit-large-patch14-336",
 
34
  "num_hidden_layers": 32,
35
  "num_key_value_heads": 32,
36
  "num_level_reg_features": 4,
37
+ "num_reg_features": 4,
38
  "out_dim": 256,
39
  "pad_token_id": 0,
40
  "pretrain_mm_mlp_adapter": null,
41
  "pretraining_tp": 1,
42
  "rms_norm_eps": 1e-05,
43
  "rope_scaling": null,
44
+ "select_feature_type": "patch",
45
  "tie_word_embeddings": false,
46
  "torch_dtype": "bfloat16",
47
  "train_mask_decoder": true,
48
  "transformers_version": "4.28.0.dev0",
49
+ "tune_mlp_adapter": false,
50
  "tune_mm_mlp_adapter": false,
51
  "tune_mm_vision_resampler": false,
52
  "unfreeze_mm_vision_tower": false,
53
  "use_cache": false,
54
+ "use_image_patch_token": false,
55
  "use_mm_proj": true,
56
+ "vision_module": "openai/clip-vit-large-patch14-336",
57
  "vision_tower": "openai/clip-vit-large-patch14-336",
58
  "vocab_size": 32007,
59
  "with_region": true
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5852b29e2afb0734d4c767e24e167f9332b6797a3e599b6b84dc2acba354001a
3
  size 9976691902
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dcdc7b329c5aaf02265d2eff7bdec0cc64418c0ef2310f567b442af11ce2f89
3
  size 9976691902
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:971e39574ff21400760f10803e846d1356489171d7649a3ae025b38b0107fee3
3
  size 6776538784
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04c6e1fda509b04b701ee021ba81cef0240451d138367c51bd6658482704419f
3
  size 6776538784
pytorch_model.bin.index.json CHANGED
The diff for this file is too large to render. See raw diff