| | """ |
| | MAC (Memory-as-Context) 结构集成到 Qwen 的详细实现 |
| | |
| | === MAC 结构核心原理 === |
| | |
| | 1. 将长序列分成多个 segment(如每 128 个 token 一个 segment) |
| | 2. 在每个 segment 的【开头】插入 longterm_mem_tokens(如 16 个) |
| | 3. 这些 memory tokens 会参与 attention 计算 |
| | 4. 使用 NeuralMemory 模块来动态更新这些 memory tokens 的内容 |
| | |
| | 原始序列: [t1, t2, t3, ..., t128, t129, ..., t256, ...] |
| | ↓ |
| | MAC 序列: [M1..M16, t1...t128, M1..M16, t129...t256, ...] |
| | ↑ ↑ |
| | memory tokens memory tokens |
| | |
| | === Qwen2 架构 === |
| | |
| | Qwen2DecoderLayer: |
| | ├── input_layernorm (RMSNorm) |
| | ├── self_attn (Qwen2Attention/Qwen2FlashAttention2) |
| | │ ├── q_proj, k_proj, v_proj |
| | │ ├── RoPE (rotary positional embedding) |
| | │ └── o_proj |
| | ├── post_attention_layernorm (RMSNorm) |
| | └── mlp (Qwen2MLP: gate_proj, up_proj, down_proj with SiLU) |
| | |
| | 我们需要在特定层添加 NeuralMemory 模块 |
| | """ |
| |
|
| | import torch |
| | import torch.nn as nn |
| | import torch.nn.functional as F |
| | from torch import Tensor |
| | from typing import Optional, Tuple, List, Dict, Any |
| | from copy import deepcopy |
| | from functools import partial |
| |
|
| | from einops import rearrange, repeat, pack, unpack |
| |
|
| | |
| | from titans_pytorch import NeuralMemory, MemoryMLP |
| | from titans_pytorch.neural_memory import NeuralMemState |
| |
|
| |
|
| | |
| | |
| | |
| |
|
| | def exists(v): |
| | return v is not None |
| |
|
| | def default(v, d): |
| | return v if exists(v) else d |
| |
|
| | def divisible_by(num, den): |
| | return (num % den) == 0 |
| |
|
| | def round_up_multiple(seq, mult): |
| | return ((seq + mult - 1) // mult) * mult |
| |
|
| |
|
| | |
| | |
| | |
| |
|
| | class QwenMACTransformer(nn.Module): |
| | """ |
| | 将 MAC 结构应用到 Qwen 模型的完整实现 |
| | |
| | 架构图: |
| | |
| | Input IDs |
| | │ |
| | ▼ |
| | ┌─────────────────┐ |
| | │ Token Embed │ |
| | └────────┬────────┘ |
| | │ |
| | ▼ |
| | ┌─────────────────────────────────────────┐ |
| | │ 插入 Memory Tokens 到每个 Segment 开头 │ |
| | │ [M1..Mn, t1..t_seg, M1..Mn, ...] │ |
| | └────────┬────────────────────────────────┘ |
| | │ |
| | ▼ |
| | ╔═════════════════════════════════════════╗ |
| | ║ Qwen Decoder Layer 1 ║ |
| | ║ ┌────────────────────────────────┐ ║ |
| | ║ │ RMSNorm → Self-Attention → Add │ ║ |
| | ║ └────────────────────────────────┘ ║ |
| | ║ ┌────────────────────────────────┐ ║ |
| | ║ │ RMSNorm → MLP → Add │ ║ |
| | ║ └────────────────────────────────┘ ║ |
| | ╚═════════════════════════════════════════╝ |
| | │ |
| | ▼ |
| | ╔═════════════════════════════════════════╗ |
| | ║ Qwen Decoder Layer 2 (with Memory) ║ |
| | ║ ┌────────────────────────────────┐ ║ |
| | ║ │ RMSNorm → Self-Attention → Add │ ║ |
| | ║ └────────────────────────────────┘ ║ |
| | ║ ┌──────────────────────────────────┐ ║ |
| | ║ │ ★ NeuralMemory 记忆增强 ★ │ ║ |
| | ║ │ retrieved = mem(hidden_states) │ ║ |
| | ║ │ hidden += gate * retrieved │ ║ |
| | ║ └──────────────────────────────────┘ ║ |
| | ║ ┌────────────────────────────────┐ ║ |
| | ║ │ RMSNorm → MLP → Add │ ║ |
| | ║ └────────────────────────────────┘ ║ |
| | ╚═════════════════════════════════════════╝ |
| | │ |
| | ▼ |
| | ... 更多层 ... |
| | │ |
| | ▼ |
| | ┌─────────────────┐ |
| | │ Final RMSNorm │ |
| | └────────┬────────┘ |
| | │ |
| | ▼ |
| | ┌─────────────────┐ |
| | │ LM Head │ |
| | └────────┬────────┘ |
| | │ |
| | ▼ |
| | Logits |
| | """ |
| | |
| | def __init__( |
| | self, |
| | qwen_model, |
| | |
| | segment_len: int = 128, |
| | num_longterm_mem_tokens: int = 16, |
| | num_persist_mem_tokens: int = 4, |
| | |
| | |
| | neural_memory_layers: Tuple[int, ...] = (2, 4, 6), |
| | memory_chunk_size: int = 64, |
| | memory_batch_size: int = 128, |
| | memory_depth: int = 2, |
| | |
| | |
| | dim_head: int = 64, |
| | num_heads: int = None, |
| | use_momentum: bool = True, |
| | gate_memory_output: bool = False, |
| | ): |
| | super().__init__() |
| | |
| | |
| | self.qwen = qwen_model |
| | self.config = qwen_model.config |
| | |
| | |
| | self.hidden_size = self.config.hidden_size |
| | self.num_layers = self.config.num_hidden_layers |
| | num_heads = default(num_heads, self.hidden_size // dim_head) |
| | |
| | |
| | self.segment_len = segment_len |
| | self.num_longterm_mem_tokens = num_longterm_mem_tokens |
| | self.num_persist_mem_tokens = num_persist_mem_tokens |
| | self.total_segment_len = segment_len + num_longterm_mem_tokens |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | self.persist_mem_tokens = nn.Parameter( |
| | torch.randn(num_persist_mem_tokens, self.hidden_size) * 0.02 |
| | ) |
| | |
| | |
| | |
| | self.longterm_mem_tokens = nn.Parameter( |
| | torch.randn(num_longterm_mem_tokens, self.hidden_size) * 0.02 |
| | ) |
| | |
| | |
| | |
| | |
| | |
| | self.neural_memory_layers = neural_memory_layers |
| | self.gate_memory_output = gate_memory_output |
| | |
| | |
| | self.neural_memories = nn.ModuleDict() |
| | self.memory_projections = nn.ModuleDict() |
| | self.memory_gates = nn.ModuleDict() |
| | |
| | |
| | memory_model_template = MemoryMLP( |
| | dim=dim_head, |
| | depth=memory_depth, |
| | expansion_factor=2.0 |
| | ) |
| | |
| | for layer_idx in neural_memory_layers: |
| | layer_key = str(layer_idx) |
| | |
| | |
| | self.neural_memories[layer_key] = NeuralMemory( |
| | dim=self.hidden_size, |
| | chunk_size=memory_chunk_size, |
| | batch_size=memory_batch_size, |
| | dim_head=dim_head, |
| | heads=num_heads, |
| | model=deepcopy(memory_model_template), |
| | momentum=use_momentum, |
| | momentum_order=1, |
| | qk_rmsnorm=True, |
| | pre_rmsnorm=True, |
| | default_step_transform_max_lr=0.1, |
| | ) |
| | |
| | |
| | self.memory_gates[layer_key] = nn.Sequential( |
| | nn.Linear(self.hidden_size, self.hidden_size), |
| | nn.Sigmoid() |
| | ) |
| | |
| | print(f"[QwenMAC] 初始化完成:") |
| | print(f" - 隐藏层大小: {self.hidden_size}") |
| | print(f" - 层数: {self.num_layers}") |
| | print(f" - Segment 长度: {segment_len}") |
| | print(f" - Longterm Memory Tokens: {num_longterm_mem_tokens}") |
| | print(f" - Persist Memory Tokens: {num_persist_mem_tokens}") |
| | print(f" - 记忆层: {neural_memory_layers}") |
| | |
| | def _insert_memory_tokens( |
| | self, |
| | hidden_states: Tensor, |
| | batch_size: int, |
| | seq_len: int, |
| | ) -> Tuple[Tensor, int]: |
| | """ |
| | 在序列中插入 memory tokens |
| | |
| | 输入: [batch, seq_len, hidden] |
| | 输出: [batch, new_seq_len, hidden] |
| | |
| | 处理流程: |
| | 原始: [t1, t2, ..., t128, t129, ..., t256] |
| | |
| | 1. 分成 segments: |
| | Seg1: [t1, ..., t128] |
| | Seg2: [t129, ..., t256] |
| | |
| | 2. 每个 segment 前插入 longterm_mem: |
| | Seg1: [M1, ..., M16, t1, ..., t128] |
| | Seg2: [M1, ..., M16, t129, ..., t256] |
| | |
| | 3. 合并 + 前置 persist_mem: |
| | [P1, ..., P4, M1..M16, t1..t128, M1..M16, t129..t256] |
| | """ |
| | segment_len = self.segment_len |
| | num_longterm = self.num_longterm_mem_tokens |
| | num_persist = self.num_persist_mem_tokens |
| | |
| | |
| | num_segments = (seq_len + segment_len - 1) // segment_len |
| | |
| | |
| | padded_len = num_segments * segment_len |
| | if seq_len < padded_len: |
| | padding = padded_len - seq_len |
| | hidden_states = F.pad(hidden_states, (0, 0, 0, padding)) |
| | |
| | |
| | hidden_states = rearrange( |
| | hidden_states, |
| | 'b (s n) d -> b s n d', |
| | s=num_segments, |
| | n=segment_len |
| | ) |
| | |
| | |
| | longterm_mem = repeat( |
| | self.longterm_mem_tokens, |
| | 'n d -> b s n d', |
| | b=batch_size, |
| | s=num_segments |
| | ) |
| | |
| | |
| | |
| | hidden_states = torch.cat([longterm_mem, hidden_states], dim=2) |
| | |
| | |
| | hidden_states = rearrange(hidden_states, 'b s n d -> b (s n) d') |
| | |
| | |
| | persist_mem = repeat( |
| | self.persist_mem_tokens, |
| | 'n d -> b n d', |
| | b=batch_size |
| | ) |
| | hidden_states = torch.cat([persist_mem, hidden_states], dim=1) |
| | |
| | new_seq_len = hidden_states.shape[1] |
| | |
| | return hidden_states, new_seq_len |
| | |
| | def _remove_memory_tokens( |
| | self, |
| | hidden_states: Tensor, |
| | original_seq_len: int, |
| | ) -> Tensor: |
| | """ |
| | 从输出中移除 memory tokens,恢复原始序列长度 |
| | """ |
| | segment_len = self.segment_len |
| | num_longterm = self.num_longterm_mem_tokens |
| | num_persist = self.num_persist_mem_tokens |
| | total_segment_len = segment_len + num_longterm |
| | |
| | batch_size = hidden_states.shape[0] |
| | |
| | |
| | hidden_states = hidden_states[:, num_persist:] |
| | |
| | |
| | num_segments = (original_seq_len + segment_len - 1) // segment_len |
| | |
| | |
| | hidden_states = rearrange( |
| | hidden_states, |
| | 'b (s n) d -> b s n d', |
| | s=num_segments, |
| | n=total_segment_len |
| | ) |
| | |
| | |
| | hidden_states = hidden_states[:, :, num_longterm:, :] |
| | |
| | |
| | hidden_states = rearrange(hidden_states, 'b s n d -> b (s n) d') |
| | hidden_states = hidden_states[:, :original_seq_len, :] |
| | |
| | return hidden_states |
| | |
| | def _create_mac_attention_mask( |
| | self, |
| | seq_len_with_mem: int, |
| | device: torch.device, |
| | dtype: torch.dtype, |
| | ) -> Tensor: |
| | """ |
| | 创建 MAC 风格的 attention mask |
| | |
| | MAC mask 的特点: |
| | 1. Persist memory tokens 对所有位置可见 |
| | 2. 每个 segment 内部是 causal 的 |
| | 3. Memory tokens 可以 attend 到之前的 segment |
| | |
| | 这是一个简化版本,完整版需要考虑更多细节 |
| | """ |
| | |
| | mask = torch.ones(seq_len_with_mem, seq_len_with_mem, device=device, dtype=dtype) |
| | mask = torch.tril(mask) |
| | |
| | |
| | num_persist = self.num_persist_mem_tokens |
| | mask[:, :num_persist] = 1.0 |
| | |
| | return mask |
| | |
| | def forward( |
| | self, |
| | input_ids: Tensor, |
| | attention_mask: Optional[Tensor] = None, |
| | position_ids: Optional[Tensor] = None, |
| | memory_states: Optional[Dict[str, NeuralMemState]] = None, |
| | return_memory_states: bool = True, |
| | **kwargs |
| | ) -> Dict[str, Any]: |
| | """ |
| | 前向传播 |
| | |
| | Args: |
| | input_ids: [batch, seq_len] |
| | attention_mask: [batch, seq_len] |
| | memory_states: 各层的记忆状态(用于增量推理) |
| | |
| | Returns: |
| | dict with 'logits', 'hidden_states', 'memory_states' |
| | """ |
| | batch_size, seq_len = input_ids.shape |
| | device = input_ids.device |
| | |
| | |
| | |
| | |
| | if hasattr(self.qwen.model, 'embed_tokens'): |
| | hidden_states = self.qwen.model.embed_tokens(input_ids) |
| | else: |
| | hidden_states = self.qwen.get_input_embeddings()(input_ids) |
| | |
| | |
| | |
| | |
| | hidden_states, seq_len_with_mem = self._insert_memory_tokens( |
| | hidden_states, batch_size, seq_len |
| | ) |
| | |
| | |
| | |
| | |
| | |
| | |
| | mac_attention_mask = None |
| | |
| | |
| | |
| | |
| | if memory_states is None: |
| | memory_states = {} |
| | next_memory_states = {} |
| | |
| | |
| | for layer_idx, layer in enumerate(self.qwen.model.layers): |
| | layer_key = str(layer_idx) |
| | |
| | |
| | |
| | |
| | |
| | |
| | residual = hidden_states |
| | |
| | |
| | hidden_states = layer.input_layernorm(hidden_states) |
| | |
| | |
| | |
| | attn_output = layer.self_attn( |
| | hidden_states=hidden_states, |
| | attention_mask=mac_attention_mask, |
| | position_ids=None, |
| | ) |
| | |
| | |
| | if isinstance(attn_output, tuple): |
| | attn_output = attn_output[0] |
| | |
| | hidden_states = residual + attn_output |
| | |
| | |
| | |
| | |
| | if layer_key in self.neural_memories: |
| | neural_mem = self.neural_memories[layer_key] |
| | gate_fn = self.memory_gates[layer_key] |
| | |
| | |
| | mem_state = memory_states.get(layer_key) |
| | |
| | |
| | retrieved, next_mem_state = neural_mem( |
| | hidden_states, |
| | state=mem_state |
| | ) |
| | |
| | |
| | gate = gate_fn(hidden_states) |
| | |
| | if self.gate_memory_output: |
| | |
| | hidden_states = hidden_states * (1 + gate * retrieved.sigmoid()) |
| | else: |
| | |
| | hidden_states = hidden_states + gate * retrieved |
| | |
| | |
| | next_memory_states[layer_key] = next_mem_state |
| | |
| | |
| | |
| | |
| | residual = hidden_states |
| | hidden_states = layer.post_attention_layernorm(hidden_states) |
| | hidden_states = layer.mlp(hidden_states) |
| | hidden_states = residual + hidden_states |
| | |
| | |
| | |
| | |
| | hidden_states = self.qwen.model.norm(hidden_states) |
| | |
| | |
| | |
| | |
| | hidden_states = self._remove_memory_tokens(hidden_states, seq_len) |
| | |
| | |
| | |
| | |
| | logits = self.qwen.lm_head(hidden_states) |
| | |
| | |
| | |
| | |
| | result = { |
| | 'logits': logits, |
| | 'hidden_states': hidden_states, |
| | } |
| | |
| | if return_memory_states: |
| | result['memory_states'] = next_memory_states |
| | |
| | return result |
| | |
| | def generate( |
| | self, |
| | input_ids: Tensor, |
| | max_new_tokens: int = 100, |
| | temperature: float = 1.0, |
| | top_p: float = 0.9, |
| | memory_states: Optional[Dict] = None, |
| | **kwargs |
| | ) -> Tensor: |
| | """ |
| | 简单的生成函数 |
| | """ |
| | generated = input_ids.clone() |
| | |
| | for _ in range(max_new_tokens): |
| | outputs = self.forward( |
| | generated, |
| | memory_states=memory_states, |
| | return_memory_states=True |
| | ) |
| | |
| | logits = outputs['logits'][:, -1, :] |
| | memory_states = outputs['memory_states'] |
| | |
| | |
| | if temperature > 0: |
| | probs = F.softmax(logits / temperature, dim=-1) |
| | next_token = torch.multinomial(probs, num_samples=1) |
| | else: |
| | next_token = logits.argmax(dim=-1, keepdim=True) |
| | |
| | generated = torch.cat([generated, next_token], dim=-1) |
| | |
| | |
| | if hasattr(self.config, 'eos_token_id'): |
| | if (next_token == self.config.eos_token_id).all(): |
| | break |
| | |
| | return generated |
| |
|
| |
|
| | |
| | |
| | |
| |
|
| | def main(): |
| | """ |
| | 完整的使用示例 |
| | """ |
| | print("=" * 70) |
| | print("MAC (Memory-as-Context) 集成到 Qwen 的示例") |
| | print("=" * 70) |
| | |
| | |
| | |
| | |
| | |
| | try: |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | |
| | print("\n[1] 加载 Qwen 模型...") |
| | model_name = "Qwen/Qwen2-0.5B" |
| | |
| | tokenizer = AutoTokenizer.from_pretrained( |
| | model_name, |
| | trust_remote_code=True |
| | ) |
| | |
| | qwen_model = AutoModelForCausalLM.from_pretrained( |
| | model_name, |
| | torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32, |
| | device_map="auto" if torch.cuda.is_available() else None, |
| | trust_remote_code=True |
| | ) |
| | |
| | print(f" 模型配置:") |
| | print(f" - hidden_size: {qwen_model.config.hidden_size}") |
| | print(f" - num_layers: {qwen_model.config.num_hidden_layers}") |
| | print(f" - num_heads: {qwen_model.config.num_attention_heads}") |
| | |
| | |
| | print("\n[2] 创建 QwenMAC 模型...") |
| | mac_model = QwenMACTransformer( |
| | qwen_model=qwen_model, |
| | segment_len=64, |
| | num_longterm_mem_tokens=8, |
| | num_persist_mem_tokens=4, |
| | neural_memory_layers=(1, 3, 5), |
| | memory_chunk_size=32, |
| | memory_batch_size=64, |
| | ) |
| | |
| | if torch.cuda.is_available(): |
| | mac_model = mac_model.cuda() |
| | |
| | |
| | print("\n[3] 测试前向传播...") |
| | test_text = "人工智能正在改变世界,它可以" |
| | inputs = tokenizer(test_text, return_tensors="pt") |
| | |
| | device = next(mac_model.parameters()).device |
| | input_ids = inputs.input_ids.to(device) |
| | |
| | with torch.no_grad(): |
| | outputs = mac_model(input_ids) |
| | |
| | print(f" 输入形状: {input_ids.shape}") |
| | print(f" 输出 logits 形状: {outputs['logits'].shape}") |
| | print(f" 记忆状态数量: {len(outputs['memory_states'])}") |
| | |
| | |
| | print("\n[4] 测试文本生成...") |
| | with torch.no_grad(): |
| | generated = mac_model.generate( |
| | input_ids, |
| | max_new_tokens=50, |
| | temperature=0.7 |
| | ) |
| | |
| | generated_text = tokenizer.decode(generated[0], skip_special_tokens=True) |
| | print(f" 生成文本: {generated_text}") |
| | |
| | except ImportError as e: |
| | print(f"\n注意: 需要安装 transformers") |
| | print(f"pip install transformers") |
| | print(f"错误: {e}") |
| | |
| | |
| | |
| | |
| | |
| | print("\n" + "=" * 70) |
| | print("[独立测试] NeuralMemory 组件") |
| | print("=" * 70) |
| | |
| | device = 'cuda' if torch.cuda.is_available() else 'cpu' |
| | |
| | |
| | mem = NeuralMemory( |
| | dim=512, |
| | chunk_size=32, |
| | batch_size=64, |
| | dim_head=64, |
| | heads=8, |
| | model=MemoryMLP(dim=64, depth=2), |
| | momentum=True, |
| | qk_rmsnorm=True, |
| | ).to(device) |
| | |
| | |
| | batch_size = 2 |
| | seq_len = 256 |
| | hidden_dim = 512 |
| | |
| | x = torch.randn(batch_size, seq_len, hidden_dim).to(device) |
| | |
| | print(f"\n输入形状: {x.shape}") |
| | |
| | |
| | retrieved, state = mem(x) |
| | print(f"检索输出形状: {retrieved.shape}") |
| | print(f"记忆状态 seq_index: {state.seq_index}") |
| | |
| | |
| | x2 = torch.randn(batch_size, seq_len, hidden_dim).to(device) |
| | retrieved2, state2 = mem(x2, state=state) |
| | print(f"第二次检索输出形状: {retrieved2.shape}") |
| | print(f"更新后 seq_index: {state2.seq_index}") |
| | |
| | print("\n" + "=" * 70) |
| | print("完成!") |
| | print("=" * 70) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|