ariG23498 HF Staff commited on
Commit
d2fe5d8
·
verified ·
1 Parent(s): 4547f49

Upload zai-org_AutoGLM-Phone-9B-Multilingual_0.txt with huggingface_hub

Browse files
zai-org_AutoGLM-Phone-9B-Multilingual_0.txt ADDED
@@ -0,0 +1,151 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Use a pipeline as a high-level helper
3
+ from transformers import pipeline
4
+
5
+ pipe = pipeline("image-text-to-text", model="zai-org/AutoGLM-Phone-9B-Multilingual")
6
+ messages = [
7
+ {
8
+ "role": "user",
9
+ "content": [
10
+ {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/p-blog/candy.JPG"},
11
+ {"type": "text", "text": "What animal is on the candy?"}
12
+ ]
13
+ },
14
+ ]
15
+ pipe(text=messages)
16
+ ```
17
+
18
+ ERROR:
19
+ Traceback (most recent call last):
20
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/utils/generic.py", line 1079, in wrapper
21
+ outputs = func(self, *args, **kwargs_without_recordable)
22
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/models/glm4v/modeling_glm4v.py", line 858, in forward
23
+ layer_outputs = decoder_layer(
24
+ hidden_states,
25
+ ...<5 lines>...
26
+ **kwargs,
27
+ )
28
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/modeling_layers.py", line 94, in __call__
29
+ return super().__call__(*args, **kwargs)
30
+ ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
31
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
32
+ return self._call_impl(*args, **kwargs)
33
+ ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
34
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
35
+ return forward_call(*args, **kwargs)
36
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/models/glm4v/modeling_glm4v.py", line 611, in forward
37
+ hidden_states, _ = self.self_attn(
38
+ ~~~~~~~~~~~~~~^
39
+ hidden_states=hidden_states,
40
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
41
+ ...<7 lines>...
42
+ **kwargs,
43
+ ^^^^^^^^^
44
+ )
45
+ ^
46
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
47
+ return self._call_impl(*args, **kwargs)
48
+ ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
49
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
50
+ return forward_call(*args, **kwargs)
51
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/models/glm4v/modeling_glm4v.py", line 538, in forward
52
+ query_states, key_states, cos, sin, self.rope_scaling["mrope_section"]
53
+ ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
54
+ TypeError: 'NoneType' object is not subscriptable
55
+
56
+ During handling of the above exception, another exception occurred:
57
+
58
+ Traceback (most recent call last):
59
+ File "/tmp/zai-org_AutoGLM-Phone-9B-Multilingual_0gOWvqY.py", line 36, in <module>
60
+ pipe(text=messages)
61
+ ~~~~^^^^^^^^^^^^^^^
62
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/pipelines/image_text_to_text.py", line 346, in __call__
63
+ return super().__call__(Chat(text, images), **kwargs)
64
+ ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
65
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/pipelines/base.py", line 1467, in __call__
66
+ return self.run_single(inputs, preprocess_params, forward_params, postprocess_params)
67
+ ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
68
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/pipelines/base.py", line 1474, in run_single
69
+ model_outputs = self.forward(model_inputs, **forward_params)
70
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/pipelines/base.py", line 1374, in forward
71
+ model_outputs = self._forward(model_inputs, **forward_params)
72
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/pipelines/image_text_to_text.py", line 437, in _forward
73
+ generated_sequence = self.model.generate(**model_inputs, **generate_kwargs)
74
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context
75
+ return func(*args, **kwargs)
76
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/generation/utils.py", line 2564, in generate
77
+ result = decoding_method(
78
+ self,
79
+ ...<5 lines>...
80
+ **model_kwargs,
81
+ )
82
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/generation/utils.py", line 2784, in _sample
83
+ outputs = self(**model_inputs, return_dict=True)
84
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
85
+ return self._call_impl(*args, **kwargs)
86
+ ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
87
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
88
+ return forward_call(*args, **kwargs)
89
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/utils/generic.py", line 918, in wrapper
90
+ output = func(self, *args, **kwargs)
91
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/models/glm4v/modeling_glm4v.py", line 1418, in forward
92
+ outputs = self.model(
93
+ input_ids=input_ids,
94
+ ...<9 lines>...
95
+ **kwargs,
96
+ )
97
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
98
+ return self._call_impl(*args, **kwargs)
99
+ ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
100
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
101
+ return forward_call(*args, **kwargs)
102
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/utils/generic.py", line 918, in wrapper
103
+ output = func(self, *args, **kwargs)
104
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/models/glm4v/modeling_glm4v.py", line 1267, in forward
105
+ outputs = self.language_model(
106
+ input_ids=None,
107
+ ...<5 lines>...
108
+ **kwargs,
109
+ )
110
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
111
+ return self._call_impl(*args, **kwargs)
112
+ ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
113
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
114
+ return forward_call(*args, **kwargs)
115
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/utils/generic.py", line 1081, in wrapper
116
+ raise original_exception
117
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/utils/generic.py", line 1072, in wrapper
118
+ outputs = func(self, *args, **kwargs)
119
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/models/glm4v/modeling_glm4v.py", line 858, in forward
120
+ layer_outputs = decoder_layer(
121
+ hidden_states,
122
+ ...<5 lines>...
123
+ **kwargs,
124
+ )
125
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/modeling_layers.py", line 94, in __call__
126
+ return super().__call__(*args, **kwargs)
127
+ ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
128
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
129
+ return self._call_impl(*args, **kwargs)
130
+ ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
131
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
132
+ return forward_call(*args, **kwargs)
133
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/models/glm4v/modeling_glm4v.py", line 611, in forward
134
+ hidden_states, _ = self.self_attn(
135
+ ~~~~~~~~~~~~~~^
136
+ hidden_states=hidden_states,
137
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
138
+ ...<7 lines>...
139
+ **kwargs,
140
+ ^^^^^^^^^
141
+ )
142
+ ^
143
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
144
+ return self._call_impl(*args, **kwargs)
145
+ ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
146
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
147
+ return forward_call(*args, **kwargs)
148
+ File "/tmp/.cache/uv/environments-v2/9da7231b583fbadc/lib/python3.13/site-packages/transformers/models/glm4v/modeling_glm4v.py", line 538, in forward
149
+ query_states, key_states, cos, sin, self.rope_scaling["mrope_section"]
150
+ ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
151
+ TypeError: 'NoneType' object is not subscriptable