ekolasky commited on
Commit
35e0f29
1 Parent(s): 3191346

Upload LSGRobertaForMaskedLM

Browse files
Files changed (3) hide show
  1. config.json +47 -0
  2. model.safetensors +3 -0
  3. modeling_lsg_roberta.py +1206 -0
config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "roberta-base",
3
+ "adaptive": true,
4
+ "architectures": [
5
+ "LSGRobertaForMaskedLM"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "auto_map": {
9
+ "AutoConfig": "modeling_lsg_roberta.LSGRobertaConfig",
10
+ "AutoModel": "modeling_lsg_roberta.LSGRobertaModel",
11
+ "AutoModelForCausalLM": "modeling_lsg_roberta.LSGRobertaForCausalLM",
12
+ "AutoModelForMaskedLM": "modeling_lsg_roberta.LSGRobertaForMaskedLM",
13
+ "AutoModelForMultipleChoice": "modeling_lsg_roberta.LSGRobertaForMultipleChoice",
14
+ "AutoModelForQuestionAnswering": "modeling_lsg_roberta.LSGRobertaForQuestionAnswering",
15
+ "AutoModelForSequenceClassification": "modeling_lsg_roberta.LSGRobertaForSequenceClassification",
16
+ "AutoModelForTokenClassification": "modeling_lsg_roberta.LSGRobertaForTokenClassification"
17
+ },
18
+ "base_model_prefix": "lsg",
19
+ "block_size": 128,
20
+ "bos_token_id": 0,
21
+ "classifier_dropout": null,
22
+ "eos_token_id": 2,
23
+ "hidden_act": "gelu",
24
+ "hidden_dropout_prob": 0.1,
25
+ "hidden_size": 768,
26
+ "initializer_range": 0.02,
27
+ "intermediate_size": 3072,
28
+ "layer_norm_eps": 1e-05,
29
+ "lsh_num_pre_rounds": 1,
30
+ "mask_first_token": false,
31
+ "max_position_embeddings": 16386,
32
+ "model_type": "roberta",
33
+ "num_attention_heads": 12,
34
+ "num_global_tokens": 1,
35
+ "num_hidden_layers": 12,
36
+ "pad_token_id": 1,
37
+ "pool_with_global": true,
38
+ "position_embedding_type": "absolute",
39
+ "sparse_block_size": 128,
40
+ "sparsity_factor": 2,
41
+ "sparsity_type": "norm",
42
+ "torch_dtype": "float32",
43
+ "transformers_version": "4.36.2",
44
+ "type_vocab_size": 1,
45
+ "use_cache": true,
46
+ "vocab_size": 50265
47
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9214e08f9bb944e3a5563eb81bea3fb21975e138a293bd431e4828c16eac1e0
3
+ size 549145716
modeling_lsg_roberta.py ADDED
@@ -0,0 +1,1206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from logging import warn
2
+ from transformers.models.roberta.modeling_roberta import *
3
+ import torch
4
+ import torch.nn as nn
5
+ from transformers.models.roberta.configuration_roberta import RobertaConfig
6
+ import sys
7
+
8
+ AUTO_MAP = {
9
+ "AutoModel": "modeling_lsg_roberta.LSGRobertaModel",
10
+ "AutoModelForCausalLM": "modeling_lsg_roberta.LSGRobertaForCausalLM",
11
+ "AutoModelForMaskedLM": "modeling_lsg_roberta.LSGRobertaForMaskedLM",
12
+ "AutoModelForMultipleChoice": "modeling_lsg_roberta.LSGRobertaForMultipleChoice",
13
+ "AutoModelForQuestionAnswering": "modeling_lsg_roberta.LSGRobertaForQuestionAnswering",
14
+ "AutoModelForSequenceClassification": "modeling_lsg_roberta.LSGRobertaForSequenceClassification",
15
+ "AutoModelForTokenClassification": "modeling_lsg_roberta.LSGRobertaForTokenClassification"
16
+ }
17
+
18
+ class LSGRobertaConfig(RobertaConfig):
19
+ """
20
+ This class overrides :class:`~transformers.RobertaConfig`. Please check the superclass for the appropriate
21
+ documentation alongside usage examples.
22
+ """
23
+
24
+ base_model_prefix = "lsg"
25
+ model_type = "roberta"
26
+
27
+ def __init__(
28
+ self,
29
+ adaptive=True,
30
+ base_model_prefix="lsg",
31
+ block_size=128,
32
+ lsh_num_pre_rounds=1,
33
+ mask_first_token=False,
34
+ num_global_tokens=1,
35
+ pool_with_global=True,
36
+ sparse_block_size=128,
37
+ sparsity_factor=2,
38
+ sparsity_type="norm",
39
+ **kwargs
40
+ ):
41
+ """Constructs LSGRobertaConfig."""
42
+ super().__init__(**kwargs)
43
+
44
+ self.adaptive = adaptive
45
+ self.auto_map = AUTO_MAP
46
+ self.base_model_prefix = base_model_prefix
47
+ self.block_size = block_size
48
+ self.lsh_num_pre_rounds = lsh_num_pre_rounds
49
+ self.mask_first_token = mask_first_token
50
+ self.num_global_tokens = num_global_tokens
51
+ self.pool_with_global = pool_with_global
52
+ self.sparse_block_size = sparse_block_size
53
+ self.sparsity_factor = sparsity_factor
54
+ self.sparsity_type = sparsity_type
55
+
56
+ if sparsity_type not in [None, "none", "norm", "lsh", "pooling", "stride", "block_stride", "bos_pooling"]:
57
+ logger.warning(
58
+ "[WARNING CONFIG]: sparsity_mode not in [None, 'none', 'norm', 'lsh', 'pooling', 'stride', 'block_stride', 'bos_pooling'], \
59
+ setting sparsity_type=None, computation will skip sparse attention")
60
+ self.sparsity_type = None
61
+
62
+ if self.sparsity_type in ["stride", "block_stride"]:
63
+ if self.sparsity_factor > self.num_attention_heads:
64
+ logger.warning(
65
+ "[WARNING CONFIG]: sparsity_factor > num_attention_heads is not recommended for stride/block_stride sparsity"
66
+ )
67
+
68
+ if self.num_global_tokens < 1:
69
+ logger.warning(
70
+ "[WARNING CONFIG]: num_global_tokens < 1 is not compatible, setting num_global_tokens=1"
71
+ )
72
+ self.num_global_tokens = 1
73
+ elif self.num_global_tokens > 512:
74
+ logger.warning(
75
+ "[WARNING CONFIG]: num_global_tokens > 512 is not allowed, setting num_global_tokens=512"
76
+ )
77
+ self.num_global_tokens = 512
78
+
79
+ if self.sparsity_factor > 0:
80
+ assert self.block_size % self.sparsity_factor == 0, "[ERROR CONFIG]: block_size must be divisible by sparsity_factor"
81
+ assert self.block_size//self.sparsity_factor >= 1, "[ERROR CONFIG]: make sure block_size >= sparsity_factor"
82
+
83
+ if self.mask_first_token and not pool_with_global:
84
+ logger.warning(
85
+ "[WARNING CONFIG]: pool_with_global==False is not compatible with mask_first_token==True. Setting pool_with_global to True.")
86
+ self.pool_with_global = True
87
+
88
+ if hasattr(self, "position_embedding_type"):
89
+ if self.position_embedding_type != "absolute":
90
+ logger.warning(
91
+ "[WARNING CONFIG]: LSG Attention is not compatible with relative positional embedding and will skip its computation. Set position_embedding_type='absolute' to remove this warning.")
92
+
93
+
94
+ class BaseSelfAttention(nn.Module):
95
+
96
+ def init_modules(self, config):
97
+ if config.hidden_size % config.num_attention_heads != 0 and not hasattr(
98
+ config, "embedding_size"
99
+ ):
100
+ raise ValueError(
101
+ "The hidden size (%d) is not a multiple of the number of attention "
102
+ "heads (%d)" % (config.hidden_size, config.num_attention_heads)
103
+ )
104
+
105
+ self.num_attention_heads = config.num_attention_heads
106
+ self.attention_head_size = int(config.hidden_size / config.num_attention_heads)
107
+ self.all_head_size = self.num_attention_heads * self.attention_head_size
108
+
109
+ self.query = nn.Linear(config.hidden_size, self.all_head_size)
110
+ self.key = nn.Linear(config.hidden_size, self.all_head_size)
111
+ self.value = nn.Linear(config.hidden_size, self.all_head_size)
112
+
113
+ self.dropout = nn.Dropout(config.attention_probs_dropout_prob)
114
+
115
+ def transpose_for_scores(self, x):
116
+ new_x_shape = x.size()[:-1] + (
117
+ self.num_attention_heads,
118
+ self.attention_head_size,
119
+ )
120
+ x = x.view(*new_x_shape)
121
+ return x.permute(0, 2, 1, 3)
122
+
123
+ def reshape_output(self, context_layer):
124
+ context_layer = context_layer.permute(0, 2, 1, 3).contiguous()
125
+ new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,)
126
+ return context_layer.view(*new_context_layer_shape)
127
+
128
+ def project_QKV(self, hidden_states):
129
+
130
+ query_layer = self.transpose_for_scores(self.query(hidden_states))
131
+ key_layer = self.transpose_for_scores(self.key(hidden_states))
132
+ value_layer = self.transpose_for_scores(self.value(hidden_states))
133
+ return query_layer, key_layer, value_layer
134
+
135
+
136
+ class BaseAttentionProduct(nn.Module):
137
+
138
+ def __init__(self, config):
139
+ """
140
+ Compute attention: softmax(Q @ K.T) @ V
141
+ """
142
+ super().__init__()
143
+ self.dropout = nn.Dropout(config.attention_probs_dropout_prob)
144
+
145
+ def forward(self, query_layer, key_layer, value_layer, attention_mask=None):
146
+
147
+ d = query_layer.shape[-1]
148
+
149
+ # Take the dot product between "query" and "key" to get the raw attention scores.
150
+ attention_scores = query_layer @ key_layer.transpose(-1, -2) / math.sqrt(d)
151
+
152
+ del query_layer
153
+ del key_layer
154
+
155
+ if attention_mask is not None:
156
+ # Apply the attention mask is (precomputed for all layers in RobertaModel forward() function)
157
+ attention_scores = attention_scores + attention_mask
158
+ del attention_mask
159
+
160
+ # Normalize the attention scores to probabilities.
161
+ attention_probs = nn.Softmax(dim=-1)(attention_scores)
162
+
163
+ # This is actually dropping out entire tokens to attend to, which might
164
+ # seem a bit unusual, but is taken from the original Transformer paper.
165
+ context_layer = self.dropout(attention_probs) @ value_layer
166
+
167
+ return context_layer
168
+
169
+
170
+ class CausalAttentionProduct(nn.Module):
171
+
172
+ def __init__(self, config):
173
+ """
174
+ Compute attention: softmax(Q @ K.T) @ V
175
+ """
176
+ super().__init__()
177
+ self.dropout = nn.Dropout(config.attention_probs_dropout_prob)
178
+ self.block_size = config.block_size
179
+
180
+ def forward(self, query_layer, key_layer, value_layer, attention_mask=None, causal_shape=None):
181
+
182
+ d = query_layer.shape[-1]
183
+
184
+ # Take the dot product between "query" and "key" to get the raw attention scores.
185
+ attention_scores = query_layer @ key_layer.transpose(-1, -2) / math.sqrt(d)
186
+
187
+ del query_layer
188
+ del key_layer
189
+
190
+ if attention_mask is not None:
191
+ # Add causal mask
192
+ causal_shape = (self.block_size, self.block_size) if causal_shape is None else causal_shape
193
+ causal_mask = torch.tril(
194
+ torch.ones(*causal_shape, device=attention_mask.device, dtype=attention_scores.dtype),
195
+ diagonal=-1
196
+ )
197
+
198
+ # Min value
199
+ dtype_min = torch.tensor(
200
+ torch.finfo(attention_scores.dtype).min, device=attention_scores.device, dtype=attention_scores.dtype
201
+ )
202
+
203
+ # Build causal + attention_mask
204
+ causal_mask = torch.nn.functional.pad(causal_mask.T * dtype_min, (attention_mask.size()[-1] - self.block_size, 0), value=0)
205
+ attention_mask = torch.max(attention_mask + causal_mask.unsqueeze(0).unsqueeze(0).unsqueeze(0), dtype_min)
206
+
207
+ attention_scores = attention_scores + attention_mask
208
+ del attention_mask
209
+ del causal_mask
210
+
211
+ # Normalize the attention scores to probabilities.
212
+ attention_probs = nn.Softmax(dim=-1)(attention_scores)
213
+
214
+ # This is actually dropping out entire tokens to attend to, which might
215
+ # seem a bit unusual, but is taken from the original Transformer paper.
216
+ context_layer = self.dropout(attention_probs) @ value_layer
217
+
218
+ return context_layer
219
+
220
+
221
+ class LSGAttentionProduct(nn.Module):
222
+
223
+ def __init__(self, config, block_size=None, sparse_block_size=None, sparsity_factor=4, is_causal=False):
224
+ """
225
+ Compute block or overlapping blocks attention products
226
+ """
227
+ super().__init__()
228
+
229
+ self.block_size = block_size
230
+ self.sparse_block_size = sparse_block_size
231
+ self.sparsity_factor = sparsity_factor
232
+ self.is_causal = is_causal
233
+
234
+ if self.block_size is None:
235
+ self.block_size = config.block_size
236
+
237
+ if self.sparse_block_size is None:
238
+ self.sparse_block_size = config.sparse_block_size
239
+
240
+ # Shape of blocks
241
+ self.local_shapes = (self.block_size*3, self.block_size)
242
+ if self.sparse_block_size and self.sparsity_factor > 0:
243
+ self.sparse_shapes = (self.sparse_block_size*3, self.block_size//self.sparsity_factor)
244
+
245
+ if is_causal:
246
+ self.attention = CausalAttentionProduct(config)
247
+ else:
248
+ self.attention = BaseAttentionProduct(config)
249
+
250
+ def build_lsg_inputs(self, hidden_states, sparse_hidden_states, global_hidden_states, is_attn_mask=False):
251
+
252
+ # Build local tokens
253
+ local_hidden_states = self.reshape_to_local_block(hidden_states, is_attn_mask)
254
+ del hidden_states
255
+
256
+ # Build sparse tokens
257
+ if sparse_hidden_states is not None:
258
+ sparse_hidden_states = self.reshape_to_sparse_block(sparse_hidden_states, is_attn_mask)
259
+
260
+ return self.cat_global_sparse_local_tokens(global_hidden_states, sparse_hidden_states, local_hidden_states)
261
+
262
+ def forward(
263
+ self,
264
+ query_layer,
265
+ key_layer,
266
+ value_layer,
267
+ attention_mask=None,
268
+ sparse_key=None,
269
+ sparse_value=None,
270
+ sparse_mask=None,
271
+ global_key=None,
272
+ global_value=None,
273
+ global_mask=None
274
+ ):
275
+
276
+ # Input batch, heads, length, hidden_size
277
+ n, h, t, d = query_layer.size()
278
+ n_blocks = t // self.block_size
279
+ assert t % self.block_size == 0
280
+
281
+ key_layer = self.build_lsg_inputs(
282
+ key_layer,
283
+ sparse_key,
284
+ global_key
285
+ )
286
+ del sparse_key
287
+ del global_key
288
+
289
+ value_layer = self.build_lsg_inputs(
290
+ value_layer,
291
+ sparse_value,
292
+ global_value
293
+ )
294
+ del sparse_value
295
+ del global_value
296
+
297
+ attention_mask = self.build_lsg_inputs(
298
+ attention_mask,
299
+ sparse_mask,
300
+ global_mask.transpose(-1, -2),
301
+ is_attn_mask=True
302
+ ).transpose(-1, -2)
303
+ del sparse_mask
304
+ del global_mask
305
+
306
+ # expect (..., t, d) shape
307
+ # Compute attention
308
+ context_layer = self.attention(
309
+ query_layer=self.chunk(query_layer, n_blocks),
310
+ key_layer=key_layer,
311
+ value_layer=value_layer,
312
+ attention_mask=attention_mask
313
+ )
314
+
315
+ return context_layer.reshape(n, h, -1, d)
316
+
317
+ def reshape_to_local_block(self, hidden_states, is_attn_mask=False):
318
+
319
+ size, step = self.local_shapes
320
+ s = (size - step) // 2
321
+
322
+ # Pad before block reshaping
323
+ if is_attn_mask:
324
+ pad_value = torch.finfo(hidden_states.dtype).min
325
+ hidden_states = hidden_states.transpose(-1, -2)
326
+ else:
327
+ pad_value = 0
328
+
329
+ hidden_states = torch.nn.functional.pad(
330
+ hidden_states.transpose(-1, -2),
331
+ pad=(s, s),
332
+ value=pad_value
333
+ ).transpose(-1, -2)
334
+
335
+ # Make blocks
336
+ hidden_states = hidden_states.unfold(-2, size=size, step=step).transpose(-1, -2)
337
+
338
+ # Skip third block if causal
339
+ if self.is_causal:
340
+ return hidden_states[..., :size*2//3, :]
341
+
342
+ return hidden_states
343
+
344
+ def reshape_to_sparse_block(self, hidden_states, is_attn_mask=False):
345
+
346
+ size, step = self.sparse_shapes
347
+
348
+ # In case of odd case
349
+ odd_offset = (step % 2)
350
+
351
+ # n, h, t, d*2 + 1
352
+ size = size*2
353
+ s = (size - step) // 2 + odd_offset
354
+
355
+ # Pad before block reshaping
356
+ if is_attn_mask:
357
+ pad_value = torch.finfo(hidden_states.dtype).min
358
+ hidden_states = hidden_states.transpose(-1, -2)
359
+ else:
360
+ pad_value = 0
361
+
362
+ hidden_states = torch.nn.functional.pad(
363
+ hidden_states.transpose(-1, -2),
364
+ pad=(s, s),
365
+ value=pad_value
366
+ ).transpose(-1, -2)
367
+
368
+ # Make blocks
369
+ hidden_states = hidden_states.unfold(-2, size=size, step=step).transpose(-1, -2)
370
+
371
+ # Fix case where block_size == sparsify_factor
372
+ if odd_offset:
373
+ hidden_states = hidden_states[..., :-1, :, :]
374
+
375
+ # Indexes for selection
376
+ u = (size - self.block_size * 3 // self.sparsity_factor) // 2 + odd_offset
377
+ s = self.sparse_block_size
378
+
379
+ # Skip right block if causal
380
+ if self.is_causal:
381
+ return hidden_states[..., u-s:u, :]
382
+
383
+ u_ = u + odd_offset
384
+ return torch.cat([hidden_states[..., u-s:u, :], hidden_states[..., -u_:-u_+s, :]], dim=-2)
385
+
386
+ def cat_global_sparse_local_tokens(self, x_global, x_sparse=None, x_local=None, dim=-2):
387
+
388
+ n, h, b, t, d = x_local.size()
389
+ x_global = x_global.unsqueeze(-3).expand(-1, -1, b, -1, -1)
390
+ if x_sparse is not None:
391
+ return torch.cat([x_global, x_sparse, x_local], dim=dim)
392
+ return torch.cat([x_global, x_local], dim=dim)
393
+
394
+ def chunk(self, x, n_blocks):
395
+
396
+ t, d = x.size()[-2:]
397
+ return x.reshape(*x.size()[:-2], n_blocks, -1, d)
398
+
399
+
400
+ class LSGRobertaEmbeddings(RobertaEmbeddings):
401
+
402
+ def __init__(self, config):
403
+ super().__init__(config)
404
+
405
+ self.num_global_tokens = config.num_global_tokens
406
+
407
+ # Hardcoded but partially trained
408
+ self.global_embeddings = nn.Embedding(512, embedding_dim=config.hidden_size, )
409
+
410
+ self.block_size = config.block_size
411
+
412
+ def forward(
413
+ self, input_ids=None, token_type_ids=None, position_ids=None, inputs_embeds=None, past_key_values_length=0
414
+ ):
415
+ if position_ids is None:
416
+ if input_ids is not None:
417
+ # Create the position ids from the input token ids. Any padded tokens remain padded.
418
+ position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length)
419
+ else:
420
+ position_ids = self.create_position_ids_from_inputs_embeds(inputs_embeds)
421
+
422
+ if input_ids is not None:
423
+ input_shape = input_ids.size()
424
+ else:
425
+ input_shape = inputs_embeds.size()[:-1]
426
+
427
+ seq_length = input_shape[1]
428
+
429
+ # Setting the token_type_ids to the registered buffer in constructor where it is all zeros, which usually occurs
430
+ # when its auto-generated, registered buffer helps users when tracing the model without passing token_type_ids, solves
431
+ # issue #5664
432
+ if token_type_ids is None:
433
+ if hasattr(self, "token_type_ids"):
434
+ buffered_token_type_ids = self.token_type_ids[:, :seq_length]
435
+ buffered_token_type_ids_expanded = buffered_token_type_ids.expand(input_shape[0], seq_length)
436
+ token_type_ids = buffered_token_type_ids_expanded
437
+ else:
438
+ token_type_ids = torch.zeros(input_shape, dtype=torch.long, device=self.position_ids.device)
439
+
440
+ if inputs_embeds is None:
441
+ inputs_embeds = self.word_embeddings(input_ids)
442
+ token_type_embeddings = self.token_type_embeddings(token_type_ids[:, :seq_length])
443
+
444
+ embeddings = inputs_embeds + token_type_embeddings
445
+ if self.position_embedding_type == "absolute":
446
+ position_embeddings = self.position_embeddings(position_ids[:, :seq_length])
447
+ embeddings += position_embeddings
448
+
449
+ #if self.num_global_tokens < 0:
450
+ n, t, d = embeddings.size()
451
+
452
+ # Add global_tokens
453
+ indexes = torch.arange(self.num_global_tokens, device=embeddings.device).reshape(1, -1)
454
+ global_embeddings = self.global_embeddings(indexes)
455
+ embeddings = torch.cat([global_embeddings.expand(n, -1, d), embeddings], dim=-2)
456
+
457
+ embeddings = self.LayerNorm(embeddings)
458
+ embeddings = self.dropout(embeddings)
459
+ return embeddings
460
+
461
+
462
+ class LSGAttention(RobertaAttention):
463
+
464
+ def __init__(self, config):
465
+
466
+ super().__init__(config)
467
+
468
+ self.self = LSGSelfAttention(config)
469
+
470
+
471
+ class LSGSelfAttention(BaseSelfAttention):
472
+ '''
473
+ Compute local attention with overlapping blocs
474
+ Use global attention for tokens with highest norm
475
+ '''
476
+ def __init__(self, config):
477
+ super().__init__()
478
+
479
+ self.init_modules(config)
480
+
481
+ self.block_size = config.block_size
482
+ self.sparse_block_size = config.sparse_block_size
483
+ self.num_global_tokens = config.num_global_tokens
484
+ self.sparsity_factor = config.sparsity_factor
485
+ self.is_causal = config.is_decoder
486
+ self.is_decoder = config.is_decoder
487
+
488
+ self.attention = LSGAttentionProduct(
489
+ config,
490
+ block_size=config.block_size,
491
+ sparse_block_size=config.sparse_block_size,
492
+ sparsity_factor=self.sparsity_factor,
493
+ is_causal=self.is_causal
494
+ )
495
+
496
+ if self.is_causal:
497
+ self.causal_attention = CausalAttentionProduct(config)
498
+ self.full_attention = BaseAttentionProduct(config)
499
+
500
+ sparse_functions = {
501
+ "norm": self.get_sparse_tokens_with_norm,
502
+ "pooling": self.get_sparse_tokens_with_pooling,
503
+ "lsh": self.get_sparse_tokens_with_lsh,
504
+ "stride": self.get_sparse_tokens_with_stride,
505
+ "block_stride": self.get_sparse_tokens_with_block_stride,
506
+ "bos_pooling": self.get_sparse_tokens_with_bos_pooling
507
+ }
508
+
509
+ self.sparsity_type = config.sparsity_type
510
+ self.get_sparse_elements = sparse_functions.get(self.sparsity_type, lambda w, x, y, z: (None, None, None))
511
+
512
+ if config.sparsity_type == "lsh":
513
+ self.lsh_num_pre_rounds = config.lsh_num_pre_rounds
514
+
515
+ def get_sparse_tokens_with_norm(self, queries, keys, values, mask):
516
+
517
+ if self.sparsity_factor == 1:
518
+ return keys, values, mask.expand(-1, keys.size()[1], -1, -1)
519
+
520
+ with torch.no_grad():
521
+
522
+ block_size = min(self.block_size, self.sparse_block_size)
523
+ key_norm = keys.detach().norm(dim=-1, keepdim=True)
524
+ key_norm = key_norm * ~mask.transpose(-1, -2).bool()
525
+ key_norm = self.chunk(key_norm, block_size)
526
+
527
+ n, h, b, t, d = key_norm.size()
528
+
529
+ idx = key_norm.argsort(dim=-2)
530
+ del key_norm
531
+ idx += (torch.arange(b, device=keys.device)*t).reshape(1, 1, b, 1, 1)
532
+
533
+ split = (t - block_size // self.sparsity_factor, block_size // self.sparsity_factor)
534
+ sparse_idx = idx.split(split, -2)[-1].reshape(n, h, -1, 1)
535
+
536
+ d = keys.size()[-1]
537
+ keys = keys.gather(dim=-2, index=sparse_idx.expand(-1, -1, -1, d))
538
+ values = values.gather(dim=-2, index=sparse_idx.expand(-1, -1, -1, d))
539
+ mask = mask.expand(-1, h, -1, -1).transpose(-1, -2).gather(dim=-2, index=sparse_idx).transpose(-1, -2)
540
+
541
+ return keys, values, mask
542
+
543
+ def get_sparse_tokens_with_pooling(self, queries, keys, values, mask):
544
+
545
+ if self.sparsity_factor == 1:
546
+ return keys, values, mask.expand(-1, keys.size()[1], -1, -1)
547
+
548
+ keys = self.chunk(keys, self.sparsity_factor)
549
+ values = self.chunk(values, self.sparsity_factor)
550
+
551
+ n, h, b, t, d = keys.size()
552
+ mask = mask.reshape(n, 1, b, 1, t)
553
+ mask = ~mask.transpose(-1, -2).bool()
554
+
555
+ keys = keys * mask
556
+ values = values * mask
557
+
558
+ mask = mask.sum(dim=-2)
559
+ keys = keys.sum(dim=-2) / (mask + 1e-6)
560
+ values = values.sum(dim=-2) / (mask + 1e-6)
561
+
562
+ mask = (1. - mask.clamp(0, 1))
563
+ mask *= torch.finfo(mask.dtype).min
564
+ return keys.reshape(n, h, -1, d), values.reshape(n, h, -1, d), mask.expand(-1, h, -1, -1).transpose(-1, -2)
565
+
566
+ def get_sparse_tokens_with_stride(self, queries, keys, values, mask):
567
+
568
+ if self.sparsity_factor == 1:
569
+ return keys, values, mask.expand(-1, keys.size()[1], -1, -1)
570
+
571
+ n, h, t, d = keys.size()
572
+ sparse_idx = torch.arange(t // self.sparsity_factor, device=keys.device) * self.sparsity_factor
573
+ sparse_idx = sparse_idx.reshape(1, 1, -1, 1) + (torch.arange(h, device=keys.device) % self.sparsity_factor).reshape(1, h, 1, 1)
574
+ sparse_idx = sparse_idx.expand(n, h, -1, 1)
575
+
576
+ keys = keys.gather(dim=-2, index=sparse_idx.expand(-1, -1, -1, d))
577
+ values = values.gather(dim=-2, index=sparse_idx.expand(-1, -1, -1, d))
578
+ mask = mask.expand(-1, h, -1, -1).transpose(-1, -2).gather(dim=-2, index=sparse_idx).transpose(-1, -2)
579
+
580
+ return keys, values, mask
581
+
582
+ def get_sparse_tokens_with_block_stride(self, queries, keys, values, mask):
583
+
584
+ if self.sparsity_factor == 1:
585
+ return keys, values, mask.expand(-1, keys.size()[1], -1, -1)
586
+
587
+ n, h, t, d = keys.size()
588
+
589
+ t, b = self.block_size, t // self.block_size
590
+ sparse_idx = torch.arange(t // self.sparsity_factor, device=keys.device)
591
+ sparse_idx = sparse_idx.reshape(1, 1, 1, -1, 1) + torch.arange(h, device=keys.device).reshape(1, h, 1, 1, 1) * (t // self.sparsity_factor)
592
+ sparse_idx = (sparse_idx % t)
593
+ sparse_idx = sparse_idx + torch.arange(b, device=keys.device).reshape(1, 1, -1, 1, 1) * t
594
+ sparse_idx = sparse_idx.reshape(1, h, -1, 1).expand(n, h, -1, 1)
595
+
596
+ keys = keys.gather(dim=-2, index=sparse_idx.expand(-1, -1, -1, d))
597
+ values = values.gather(dim=-2, index=sparse_idx.expand(-1, -1, -1, d))
598
+ mask = mask.expand(-1, h, -1, -1).transpose(-1, -2).gather(dim=-2, index=sparse_idx).transpose(-1, -2)
599
+
600
+ return keys, values, mask
601
+
602
+ def get_sparse_tokens_with_lsh(self, queries, keys, values, mask):
603
+
604
+ if self.sparsity_factor == 1:
605
+ return keys, values, mask.expand(-1, keys.size()[1], -1, -1)
606
+
607
+ if self.sparsity_factor == self.sparse_block_size:
608
+ return self.get_sparse_tokens_with_bos_pooling(queries, keys, values, mask)
609
+
610
+ block_size = min(self.block_size, self.sparse_block_size)
611
+ keys = self.chunk(keys, block_size)
612
+ values = self.chunk(values, block_size)
613
+
614
+ n, h, b, t, d = keys.size()
615
+ mask = mask.reshape(n, 1, b, 1, t)
616
+ mask = ~mask.transpose(-1, -2).bool()
617
+
618
+ keys = keys * mask
619
+ values = values * mask
620
+ mask = mask.expand(-1, h, -1, -1, -1).float()
621
+
622
+ extra_factor = 1
623
+
624
+ for _ in range(self.lsh_num_pre_rounds):
625
+ keys, values, mask = self.lsh_round(keys, values, mask, t*extra_factor)
626
+
627
+ keys, values, mask = self.lsh_round(keys, values, mask, t//self.sparsity_factor)
628
+ keys /= mask + 1e-8
629
+ values /= mask + 1e-8
630
+
631
+ mask = (1. - mask.clamp(0, 1))
632
+ mask *= torch.finfo(mask.dtype).min
633
+
634
+ return keys.reshape(n, h, -1, d), values.reshape(n, h, -1, d), mask.transpose(-1, -2).reshape(n, h, 1, -1)
635
+
636
+ def lsh_round(self, keys, values, mask, output_size):
637
+
638
+ with torch.no_grad():
639
+
640
+ n_hashes = output_size // 2
641
+ n, h, b, t, d = keys.size()
642
+ binary_mask = mask.clamp(0, 1)
643
+
644
+ indexes = (torch.nn.functional.normalize(keys, dim=-1) * binary_mask) @ torch.randn(1, h, 1, d, n_hashes, device=keys.device)
645
+ indexes = torch.cat([indexes, -indexes], dim=-1).argmax(dim=-1, keepdim=True)
646
+
647
+ n, h, b, t, d = keys.size()
648
+
649
+ x_ = torch.zeros(n, h, b, output_size, d, device=keys.device)
650
+ mask_ = torch.zeros(n, h, b, output_size, 1, device=keys.device)
651
+ keys = torch.scatter_add(x_, dim=-2, index=indexes.expand(-1, -1, -1, -1, d), src=keys)
652
+ values = torch.scatter_add(x_, dim=-2, index=indexes.expand(-1, -1, -1, -1, d), src=values)
653
+ mask = torch.scatter_add(mask_, dim=-2, index=indexes, src=mask)
654
+
655
+ return keys[..., :output_size, :], values[..., :output_size, :], mask[..., :output_size, :]
656
+
657
+ def get_sparse_tokens_with_bos_pooling(self, queries, keys, values, mask):
658
+
659
+ if self.sparsity_factor == 1:
660
+ return keys, values, mask.expand(-1, keys.size()[1], -1, -1)
661
+
662
+ queries = queries.unsqueeze(-3)
663
+ mask = self.chunk(mask.transpose(-1, -2), self.sparsity_factor).transpose(-1, -2)
664
+ keys = self.chunk(keys, self.sparsity_factor)
665
+ values = self.chunk(values, self.sparsity_factor)
666
+
667
+ n, h, b, t, d = keys.size()
668
+ scores = (queries[..., :1, :] @ keys.transpose(-1, -2)) / math.sqrt(d)
669
+ if mask is not None:
670
+ scores = scores + mask
671
+
672
+ scores = torch.softmax(scores, dim=-1)
673
+ keys = scores @ keys
674
+ values = scores @ values
675
+ mask = mask.mean(dim=-1)
676
+ mask[mask != torch.finfo(mask.dtype).min] = 0
677
+
678
+ return keys.reshape(n, h, -1, d), values.reshape(n, h, -1, d), mask.expand(-1, h, -1, -1).transpose(-1, -2)
679
+
680
+ def forward(
681
+ self,
682
+ hidden_states,
683
+ attention_mask=None,
684
+ head_mask=None,
685
+ encoder_hidden_states=None,
686
+ encoder_attention_mask=None,
687
+ past_key_value=None,
688
+ output_attentions=False,
689
+ ):
690
+
691
+ query_layer = self.query(hidden_states)
692
+
693
+ # If this is instantiated as a cross-attention module, the keys
694
+ # and values come from an encoder; the attention mask needs to be
695
+ # such that the encoder's padding tokens are not attended to.
696
+ is_cross_attention = encoder_hidden_states is not None
697
+
698
+ if is_cross_attention and past_key_value is not None:
699
+ # reuse k,v, cross_attentions
700
+ key_layer = past_key_value[0]
701
+ value_layer = past_key_value[1]
702
+ attention_mask = encoder_attention_mask
703
+ elif is_cross_attention:
704
+ key_layer = self.transpose_for_scores(self.key(encoder_hidden_states))
705
+ value_layer = self.transpose_for_scores(self.value(encoder_hidden_states))
706
+ attention_mask = encoder_attention_mask
707
+ elif past_key_value is not None:
708
+ key_layer = self.transpose_for_scores(self.key(hidden_states))
709
+ value_layer = self.transpose_for_scores(self.value(hidden_states))
710
+ key_layer = torch.cat([past_key_value[0], key_layer], dim=2)
711
+ value_layer = torch.cat([past_key_value[1], value_layer], dim=2)
712
+ else:
713
+ key_layer = self.transpose_for_scores(self.key(hidden_states))
714
+ value_layer = self.transpose_for_scores(self.value(hidden_states))
715
+
716
+ query_layer = self.transpose_for_scores(query_layer)
717
+
718
+ if self.is_decoder:
719
+ # if cross_attention save Tuple(torch.Tensor, torch.Tensor) of all cross attention key/value_states.
720
+ # Further calls to cross_attention layer can then reuse all cross-attention
721
+ # key/value_states (first "if" case)
722
+ # if uni-directional self-attention (decoder) save Tuple(torch.Tensor, torch.Tensor) of
723
+ # all previous decoder key/value_states. Further calls to uni-directional self-attention
724
+ # can concat previous decoder key/value_states to current projected key/value_states (third "elif" case)
725
+ # if encoder bi-directional self-attention `past_key_value` is always `None`
726
+ past_key_value = (key_layer, value_layer)
727
+
728
+ if is_cross_attention:
729
+ outputs = self.cross_attention_forward(
730
+ query_layer=query_layer,
731
+ key_layer=key_layer,
732
+ value_layer=value_layer,
733
+ attention_mask=attention_mask,
734
+ output_attentions=output_attentions
735
+ )
736
+ else:
737
+ outputs = self.causal_forward(
738
+ query_layer,
739
+ key_layer,
740
+ value_layer,
741
+ attention_mask=attention_mask,
742
+ output_attentions=output_attentions,
743
+ )
744
+
745
+ outputs = outputs + ((key_layer, value_layer),)
746
+
747
+ else:
748
+ outputs = self.not_causal_forward(
749
+ query_layer,
750
+ key_layer,
751
+ value_layer,
752
+ attention_mask=attention_mask,
753
+ output_attentions=output_attentions
754
+ )
755
+
756
+ return outputs
757
+
758
+ def causal_forward(
759
+ self,
760
+ query_layer,
761
+ key_layer,
762
+ value_layer,
763
+ attention_mask=None,
764
+ output_attentions=False,
765
+ ):
766
+
767
+ n, h, t, d = key_layer.size()
768
+
769
+ # Cat global mask
770
+ attention_mask = torch.nn.functional.pad(attention_mask, (self.num_global_tokens, 0), value=0)
771
+
772
+ # Split input into global tokens and other tokens
773
+ split = (self.num_global_tokens, t - self.num_global_tokens)
774
+ global_query, query_layer = query_layer.split(split, dim=-2)
775
+
776
+ # Use normal causal attention if local attention covers every tokens
777
+ if t <= 2 * self.block_size + self.num_global_tokens:
778
+ context_layer = self.causal_attention(
779
+ query_layer=query_layer,
780
+ key_layer=key_layer,
781
+ value_layer=value_layer,
782
+ attention_mask=attention_mask,
783
+ causal_shape=(t - self.num_global_tokens, t - self.num_global_tokens)
784
+ )
785
+
786
+ context_layer = torch.cat([global_query, context_layer], dim=-2)
787
+ return (self.reshape_output(context_layer), )
788
+
789
+ # Split K Q M on global and non global
790
+ global_key, key_layer = key_layer.split(split, dim=-2)
791
+ global_value, value_layer = value_layer.split(split, dim=-2)
792
+ global_mask, attention_mask = attention_mask.split(split, dim=-1)
793
+
794
+ n, h, t, d = key_layer.size()
795
+
796
+ # Get sparse idx
797
+ sparse_key, sparse_value, sparse_mask = (None, None, None)
798
+ if self.sparse_block_size and self.sparsity_factor > 0:
799
+ sparse_key, sparse_value, sparse_mask = self.get_sparse_elements(query_layer, key_layer, value_layer, attention_mask)
800
+
801
+ # Expand masks on heads
802
+ attention_mask = attention_mask.expand(-1, h, -1, -1)
803
+ global_mask = global_mask.expand(-1, h, -1, -1)
804
+
805
+ # Compute dot product attention
806
+ context_layer = self.attention(
807
+ query_layer,
808
+ key_layer,
809
+ value_layer,
810
+ attention_mask,
811
+ sparse_key=sparse_key,
812
+ sparse_value=sparse_value,
813
+ sparse_mask=sparse_mask,
814
+ global_key=global_key,
815
+ global_value=global_value,
816
+ global_mask=global_mask
817
+ )
818
+
819
+ # Merge pseudo global (causal) and local-sparse tokens
820
+ context_layer = torch.cat([global_query, context_layer], dim=-2)
821
+ context_layer = self.reshape_output(context_layer)
822
+
823
+ return (context_layer,)
824
+
825
+ def not_causal_forward(
826
+ self,
827
+ query_layer,
828
+ key_layer,
829
+ value_layer,
830
+ attention_mask=None,
831
+ output_attentions=False,
832
+ ):
833
+
834
+ n, h, t, d = query_layer.size()
835
+
836
+ # Cat global mask
837
+ attention_mask = torch.nn.functional.pad(attention_mask, (self.num_global_tokens, 0), value=0)
838
+
839
+ # Use normal attention if local attention covers every tokens
840
+ if t <= 2 * self.block_size + self.num_global_tokens:
841
+ context_layer = self.full_attention(
842
+ query_layer=query_layer,
843
+ key_layer=key_layer,
844
+ value_layer=value_layer,
845
+ attention_mask=attention_mask
846
+ )
847
+ return (self.reshape_output(context_layer), )
848
+
849
+ # Split input into global tokens and other tokens
850
+ split = (self.num_global_tokens, t - self.num_global_tokens)
851
+ global_query, query_layer = query_layer.split(split, dim=-2)
852
+
853
+ # Get global_attention
854
+ bos = self.full_attention(
855
+ query_layer=global_query,
856
+ key_layer=key_layer,
857
+ value_layer=value_layer,
858
+ attention_mask=attention_mask
859
+ )
860
+
861
+ # Split K Q M on global and non global
862
+ global_key, key_layer = key_layer.split(split, dim=-2)
863
+ global_value, value_layer = value_layer.split(split, dim=-2)
864
+ global_mask, attention_mask = attention_mask.split(split, dim=-1)
865
+
866
+ n, h, t, d = key_layer.size()
867
+
868
+ # Get sparse idx
869
+ sparse_key, sparse_value, sparse_mask = (None, None, None)
870
+
871
+ if self.sparse_block_size and self.sparsity_factor > 0:
872
+ sparse_key, sparse_value, sparse_mask = self.get_sparse_elements(query_layer, key_layer, value_layer, attention_mask)
873
+
874
+ # Expand masks on heads
875
+ attention_mask = attention_mask.expand(-1, h, -1, -1)
876
+ global_mask = global_mask.expand(-1, h, -1, -1)
877
+
878
+ # Compute dot product attention
879
+ context_layer = self.attention(
880
+ query_layer,
881
+ key_layer,
882
+ value_layer,
883
+ attention_mask,
884
+ sparse_key=sparse_key,
885
+ sparse_value=sparse_value,
886
+ sparse_mask=sparse_mask,
887
+ global_key=global_key,
888
+ global_value=global_value,
889
+ global_mask=global_mask
890
+ )
891
+
892
+ # Merge global and local-sparse tokens
893
+ context_layer = torch.cat([bos, context_layer], dim=-2)
894
+ context_layer = self.reshape_output(context_layer)
895
+
896
+ return (context_layer,)
897
+
898
+ def cross_attention_forward(
899
+ self,
900
+ query_layer,
901
+ key_layer,
902
+ value_layer,
903
+ attention_mask=None,
904
+ output_attentions=False,
905
+ ):
906
+
907
+ context_layer = self.full_attention(
908
+ query_layer=query_layer,
909
+ key_layer=key_layer,
910
+ value_layer=value_layer,
911
+ attention_mask=attention_mask
912
+ )
913
+ return (self.reshape_output(context_layer), )
914
+
915
+ def chunk(self, x, chunk_size):
916
+
917
+ n, h, t, d = x.size()
918
+ return x.reshape(n, h, -1, chunk_size, d)
919
+
920
+
921
+ class LSGRobertaLayer(RobertaLayer):
922
+
923
+ def __init__(self, config):
924
+
925
+ super().__init__(config)
926
+
927
+ self.attention = LSGAttention(config)
928
+ if self.add_cross_attention:
929
+ assert self.is_decoder, f"{self} should be used as a decoder model if cross attention is added"
930
+ self.crossattention = LSGAttention(config)
931
+
932
+
933
+ class LSGRobertaEncoder(RobertaEncoder):
934
+
935
+ def __init__(self, config):
936
+
937
+ super().__init__(config)
938
+
939
+ self.layer = nn.ModuleList([LSGRobertaLayer(config) for _ in range(config.num_hidden_layers)])
940
+
941
+ assert hasattr(config, "num_global_tokens")
942
+ self.num_global_tokens = config.num_global_tokens
943
+ self.pad_idx = config.pad_token_id
944
+
945
+ assert hasattr(config, "block_size") and hasattr(config, "adaptive")
946
+ self.block_size = config.block_size
947
+ self.adaptive = config.adaptive
948
+ self.mask_first_token = config.mask_first_token
949
+ self.pool_with_global = config.pool_with_global
950
+
951
+ def forward(
952
+ self,
953
+ hidden_states: torch.Tensor,
954
+ attention_mask: Optional[torch.FloatTensor] = None,
955
+ head_mask: Optional[torch.FloatTensor] = None,
956
+ encoder_hidden_states: Optional[torch.FloatTensor] = None,
957
+ encoder_attention_mask: Optional[torch.FloatTensor] = None,
958
+ past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None,
959
+ use_cache: Optional[bool] = None,
960
+ output_attentions: Optional[bool] = False,
961
+ output_hidden_states: Optional[bool] = False,
962
+ return_dict: Optional[bool] = True,
963
+ ) -> Union[Tuple[torch.Tensor], BaseModelOutputWithPastAndCrossAttentions]:
964
+
965
+ mask_value = torch.finfo(attention_mask.dtype).min
966
+ n, _, __, t = attention_mask.size()
967
+
968
+ if not (self.config.is_decoder and encoder_hidden_states is not None):
969
+ b = self.block_size * 2
970
+ pad = t % self.block_size
971
+
972
+ # Check if t is multiple of block_size and pad
973
+ if self.adaptive and t > b and pad > 0:
974
+ pad_length = self.block_size - pad
975
+ hidden_states = torch.nn.functional.pad(hidden_states.transpose(-1, -2), (0, pad_length), value=0.).transpose(-1, -2)
976
+ attention_mask = torch.nn.functional.pad(attention_mask, (0, pad_length), value=mask_value)
977
+
978
+ if self.mask_first_token:
979
+ attention_mask[..., 0] = mask_value
980
+
981
+ encoder_outputs = super().forward(
982
+ hidden_states=hidden_states,
983
+ attention_mask=attention_mask,
984
+ head_mask=head_mask,
985
+ encoder_hidden_states=encoder_hidden_states,
986
+ encoder_attention_mask=encoder_attention_mask,
987
+ past_key_values=past_key_values,
988
+ use_cache=use_cache,
989
+ output_attentions=output_attentions,
990
+ output_hidden_states=output_hidden_states,
991
+ return_dict=return_dict
992
+ )
993
+
994
+ sequence_output = encoder_outputs[0]
995
+ if self.pool_with_global:
996
+ sequence_output[:, self.num_global_tokens] = sequence_output[:, 0]
997
+
998
+ # Adapt sequence to initial shape
999
+ sequence_output = sequence_output[..., self.num_global_tokens: t + self.num_global_tokens, :]
1000
+
1001
+ if not return_dict:
1002
+ return (sequence_output, ) + encoder_outputs[1:]
1003
+
1004
+ encoder_outputs.last_hidden_state = sequence_output
1005
+ return encoder_outputs
1006
+
1007
+
1008
+ class LSGRobertaPreTrainedModel(RobertaPreTrainedModel):
1009
+ """
1010
+ An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained
1011
+ models.
1012
+ """
1013
+
1014
+ config_class = LSGRobertaConfig
1015
+ base_model_prefix = "roberta"
1016
+ supports_gradient_checkpointing = True
1017
+ _no_split_modules = []
1018
+
1019
+ def _set_gradient_checkpointing(self, module, value=False):
1020
+ if isinstance(module, (RobertaEncoder, LSGRobertaEncoder)):
1021
+ module.gradient_checkpointing = value
1022
+
1023
+
1024
+ class LSGRobertaModel(LSGRobertaPreTrainedModel, RobertaModel):
1025
+ """
1026
+ This class overrides :class:`~transformers.RobertaModel`. Please check the superclass for the appropriate
1027
+ documentation alongside usage examples.
1028
+ """
1029
+
1030
+ def __init__(self, config, add_pooling_layer=True):
1031
+
1032
+ LSGRobertaPreTrainedModel.__init__(self, config)
1033
+
1034
+ self.embeddings = LSGRobertaEmbeddings(config)
1035
+ self.encoder = LSGRobertaEncoder(config)
1036
+ self.pooler = RobertaPooler(config) if add_pooling_layer else None
1037
+
1038
+ if config.add_cross_attention:
1039
+ logger.warning(
1040
+ "Cross attention is computed using full attention since it is not LSG compatible."
1041
+ )
1042
+
1043
+ self._use_flash_attention_2 = config._attn_implementation == "flash_attention_2"
1044
+ if self._use_flash_attention_2:
1045
+ logger.warning(
1046
+ "[WARNING flash-attention]: LSG doesnt support flash-attention currently"
1047
+ )
1048
+
1049
+ # Initialize weights and apply final processing
1050
+ self.post_init()
1051
+
1052
+ def get_extended_attention_mask(self, attention_mask, input_shape, device=None):
1053
+
1054
+ # Do not rely on original triangular mask from BERT/RoBERTa for causalLM
1055
+ if attention_mask.dim() == 3:
1056
+ extended_attention_mask = attention_mask[:, None, :, :]
1057
+ elif attention_mask.dim() == 2:
1058
+ extended_attention_mask = attention_mask[:, None, None, :]
1059
+ else:
1060
+ raise ValueError(
1061
+ f"Wrong shape for input_ids (shape {input_shape}) or attention_mask (shape {attention_mask.shape})"
1062
+ )
1063
+
1064
+ extended_attention_mask = extended_attention_mask.to(dtype=self.dtype) # fp16 compatibility
1065
+ extended_attention_mask = (1.0 - extended_attention_mask) * torch.finfo(extended_attention_mask.dtype).min
1066
+
1067
+ return extended_attention_mask
1068
+
1069
+
1070
+ class LSGRobertaForCausalLM(LSGRobertaPreTrainedModel, RobertaForCausalLM):
1071
+
1072
+ _tied_weights_keys = ["lm_head.decoder.weight", "lm_head.decoder.bias"]
1073
+
1074
+ def __init__(self, config):
1075
+
1076
+ LSGRobertaPreTrainedModel.__init__(self, config)
1077
+
1078
+ if not config.is_decoder:
1079
+ logger.warning("If you want to use `LSGRobertaLMHeadModel` as a standalone, add `is_decoder=True.`")
1080
+
1081
+ self.roberta = LSGRobertaModel(config, add_pooling_layer=False)
1082
+ self.lm_head = RobertaLMHead(config)
1083
+
1084
+ # Initialize weights and apply final processing
1085
+ self.post_init()
1086
+
1087
+
1088
+ class LSGRobertaForMaskedLM(LSGRobertaPreTrainedModel, RobertaForMaskedLM):
1089
+ """
1090
+ This class overrides :class:`~transformers.RobertaForMaskedLM`. Please check the superclass for the appropriate
1091
+ documentation alongside usage examples.
1092
+ """
1093
+
1094
+ _tied_weights_keys = ["lm_head.decoder.weight", "lm_head.decoder.bias"]
1095
+
1096
+ def __init__(self, config):
1097
+
1098
+ LSGRobertaPreTrainedModel.__init__(self, config)
1099
+
1100
+ if config.is_decoder:
1101
+ logger.warning(
1102
+ "If you want to use `LSGRobertaForMaskedLM` make sure `config.is_decoder=False` for "
1103
+ "bi-directional self-attention."
1104
+ )
1105
+
1106
+ self.roberta = LSGRobertaModel(config, add_pooling_layer=False)
1107
+ self.lm_head = RobertaLMHead(config)
1108
+
1109
+ # Initialize weights and apply final processing
1110
+ self.post_init()
1111
+
1112
+
1113
+ class LSGRobertaForSequenceClassification(LSGRobertaPreTrainedModel, RobertaForSequenceClassification):
1114
+ """
1115
+ This class overrides :class:`~transformers.RobertaForSequenceClassification`. Please check the superclass for the
1116
+ appropriate documentation alongside usage examples.
1117
+ """
1118
+
1119
+ def __init__(self, config):
1120
+
1121
+ LSGRobertaPreTrainedModel.__init__(self, config)
1122
+
1123
+ self.num_labels = config.num_labels
1124
+ self.config = config
1125
+
1126
+ self.roberta = LSGRobertaModel(config, add_pooling_layer=False)
1127
+ self.classifier = RobertaClassificationHead(config)
1128
+
1129
+ # Initialize weights and apply final processing
1130
+ self.post_init()
1131
+
1132
+
1133
+ class LSGRobertaForMultipleChoice(LSGRobertaPreTrainedModel, RobertaForMultipleChoice):
1134
+ """
1135
+ This class overrides :class:`~transformers.RobertaForMultipleChoice`. Please check the superclass for the
1136
+ appropriate documentation alongside usage examples.
1137
+ """
1138
+
1139
+ def __init__(self, config):
1140
+
1141
+ LSGRobertaPreTrainedModel.__init__(self, config)
1142
+
1143
+ self.roberta = LSGRobertaModel(config)
1144
+ self.dropout = nn.Dropout(config.hidden_dropout_prob)
1145
+ self.classifier = nn.Linear(config.hidden_size, 1)
1146
+
1147
+ # Initialize weights and apply final processing
1148
+ self.post_init()
1149
+
1150
+
1151
+ class LSGRobertaForTokenClassification(LSGRobertaPreTrainedModel, RobertaForTokenClassification):
1152
+ """
1153
+ This class overrides :class:`~transformers.RobertaForTokenClassification`. Please check the superclass for the
1154
+ appropriate documentation alongside usage examples.
1155
+ """
1156
+
1157
+ _keys_to_ignore_on_load_unexpected = [r"pooler"]
1158
+ _keys_to_ignore_on_load_missing = [r"position_ids"]
1159
+
1160
+ def __init__(self, config):
1161
+
1162
+ LSGRobertaPreTrainedModel.__init__(self, config)
1163
+
1164
+ self.num_labels = config.num_labels
1165
+
1166
+ self.roberta = LSGRobertaModel(config, add_pooling_layer=False)
1167
+ classifier_dropout = (
1168
+ config.classifier_dropout if config.classifier_dropout is not None else config.hidden_dropout_prob
1169
+ )
1170
+ self.dropout = nn.Dropout(classifier_dropout)
1171
+ self.classifier = nn.Linear(config.hidden_size, config.num_labels)
1172
+
1173
+ # Initialize weights and apply final processing
1174
+ self.post_init()
1175
+
1176
+
1177
+ class LSGRobertaForQuestionAnswering(LSGRobertaPreTrainedModel, RobertaForQuestionAnswering):
1178
+ """
1179
+ This class overrides :class:`~transformers.RobertaForQuestionAnswering`. Please check the superclass for the
1180
+ appropriate documentation alongside usage examples.
1181
+ """
1182
+
1183
+ def __init__(self, config):
1184
+
1185
+ LSGRobertaPreTrainedModel.__init__(self, config)
1186
+
1187
+ self.num_labels = config.num_labels
1188
+
1189
+ self.roberta = LSGRobertaModel(config, add_pooling_layer=False)
1190
+ self.qa_outputs = nn.Linear(config.hidden_size, config.num_labels)
1191
+
1192
+ # Initialize weights and apply final processing
1193
+ self.post_init()
1194
+
1195
+
1196
+ def str_to_class(classname):
1197
+ return getattr(sys.modules[__name__], classname)
1198
+
1199
+ # Register model in Auto API
1200
+ try:
1201
+ LSGRobertaConfig.register_for_auto_class()
1202
+ for key, value in AUTO_MAP.items():
1203
+ str_to_class(value.split(".")[-1]).register_for_auto_class(key)
1204
+ except:
1205
+ warn("AutoRegister isn't available, you'll have to manually copy modeling.py after .save_pretrained(...).")
1206
+ warn("Update to transformers >= 4.36.1 to fix.")