wuxiaojun commited on
Commit
503ee33
1 Parent(s): 3502b52

update ZiyaBlip2ForCausalLM

Browse files
README.md CHANGED
@@ -106,7 +106,7 @@ First load the Ziya-Visual model: it should be noted that the model repository o
106
 
107
  ```python
108
  from transformers import LlamaForCausalLM, LlamaTokenizer, BlipImageProcessor
109
- from modeling_ziya_blip2 import ZiyaBLIP2ForCausalLM
110
  from PIL import Image
111
 
112
  # model path of IDEA-CCNL/Ziya-LLaMA-13B-v1
@@ -118,7 +118,7 @@ tokenizer = LlamaTokenizer.from_pretrained(LM_MODEL_PATH)
118
  OPENAI_CLIP_MEAN = [0.48145466, 0.4578275, 0.40821073]
119
  OPENAI_CLIP_STD = [0.26862954, 0.26130258, 0.27577711]
120
  # demo.py is in the project path, so we can use local path ".". Otherwise you should use "IDEA-CCNL/Ziya-BLIP2-14B-Visual-v1"
121
- model = ZiyaBLIP2ForCausalLM.from_pretrained(".", language_model=lm_model)
122
  image_size = model.config.vision_config.image_size
123
  image_processor = BlipImageProcessor(
124
  size={"height": image_size, "width": image_size},
 
106
 
107
  ```python
108
  from transformers import LlamaForCausalLM, LlamaTokenizer, BlipImageProcessor
109
+ from modeling_ziya_blip2 import ZiyaBlip2ForCausalLM
110
  from PIL import Image
111
 
112
  # model path of IDEA-CCNL/Ziya-LLaMA-13B-v1
 
118
  OPENAI_CLIP_MEAN = [0.48145466, 0.4578275, 0.40821073]
119
  OPENAI_CLIP_STD = [0.26862954, 0.26130258, 0.27577711]
120
  # demo.py is in the project path, so we can use local path ".". Otherwise you should use "IDEA-CCNL/Ziya-BLIP2-14B-Visual-v1"
121
+ model = ZiyaBlip2ForCausalLM.from_pretrained(".", language_model=lm_model)
122
  image_size = model.config.vision_config.image_size
123
  image_processor = BlipImageProcessor(
124
  size={"height": image_size, "width": image_size},
__pycache__/modeling_ziya_blip2.cpython-310.pyc CHANGED
Binary files a/__pycache__/modeling_ziya_blip2.cpython-310.pyc and b/__pycache__/modeling_ziya_blip2.cpython-310.pyc differ
 
modeling_ziya_blip2.py CHANGED
@@ -20,7 +20,7 @@ from transformers import (
20
  logger = logging.get_logger(__name__)
21
 
22
 
23
- class ZiyaBLIP2ForCausalLM(Blip2PreTrainedModel):
24
  config_class = Blip2Config
25
  main_input_name = "pixel_values"
26
  _keys_to_ignore_on_load_missing = [
 
20
  logger = logging.get_logger(__name__)
21
 
22
 
23
+ class ZiyaBlip2ForCausalLM(Blip2PreTrainedModel):
24
  config_class = Blip2Config
25
  main_input_name = "pixel_values"
26
  _keys_to_ignore_on_load_missing = [
test.py CHANGED
@@ -1,5 +1,5 @@
1
  from transformers import LlamaForCausalLM, LlamaTokenizer, BlipImageProcessor
2
- from modeling_ziya_blip2 import ZiyaBLIP2ForCausalLM
3
  from PIL import Image
4
 
5
  # 请注意目前https://huggingface.co/IDEA-CCNL/Ziya-LLaMA-13B-v1是delta权重(即差值权重)
@@ -14,7 +14,7 @@ tokenizer = LlamaTokenizer.from_pretrained(LM_MODEL_PATH)
14
  OPENAI_CLIP_MEAN = [0.48145466, 0.4578275, 0.40821073]
15
  OPENAI_CLIP_STD = [0.26862954, 0.26130258, 0.27577711]
16
  # demo.py is in the project path, so we can use local path ".". Otherwise you should use "IDEA-CCNL/Ziya-BLIP2-14B-Visual-v1"
17
- model = ZiyaBLIP2ForCausalLM.from_pretrained(".", language_model=lm_model)
18
  image_size = model.config.vision_config.image_size
19
  image_processor = BlipImageProcessor(
20
  size={"height": image_size, "width": image_size},
 
1
  from transformers import LlamaForCausalLM, LlamaTokenizer, BlipImageProcessor
2
+ from modeling_ziya_blip2 import ZiyaBlip2ForCausalLM
3
  from PIL import Image
4
 
5
  # 请注意目前https://huggingface.co/IDEA-CCNL/Ziya-LLaMA-13B-v1是delta权重(即差值权重)
 
14
  OPENAI_CLIP_MEAN = [0.48145466, 0.4578275, 0.40821073]
15
  OPENAI_CLIP_STD = [0.26862954, 0.26130258, 0.27577711]
16
  # demo.py is in the project path, so we can use local path ".". Otherwise you should use "IDEA-CCNL/Ziya-BLIP2-14B-Visual-v1"
17
+ model = ZiyaBlip2ForCausalLM.from_pretrained(".", language_model=lm_model)
18
  image_size = model.config.vision_config.image_size
19
  image_processor = BlipImageProcessor(
20
  size={"height": image_size, "width": image_size},