| | """ |
| | K-FAC Statistics Collector |
| | |
| | Collects activation covariance (A) and gradient covariance (G) matrices |
| | for MLP layers to approximate the Fisher Information Matrix. |
| | |
| | Based on: "From Memorization to Reasoning in the Spectrum of Loss Curvature" |
| | """ |
| |
|
| | import torch |
| | import torch.nn as nn |
| | from torch import Tensor |
| | from typing import Optional, Callable |
| | from dataclasses import dataclass, field |
| | from tqdm import tqdm |
| |
|
| |
|
| | @dataclass |
| | class LayerStatistics: |
| | """K-FAC statistics for a single layer.""" |
| | |
| | |
| | A: Optional[Tensor] = None |
| | G: Optional[Tensor] = None |
| | |
| | |
| | A_sum: Optional[Tensor] = None |
| | G_sum: Optional[Tensor] = None |
| | |
| | |
| | n_samples_A: int = 0 |
| | n_samples_G: int = 0 |
| | |
| | def finalize(self) -> None: |
| | """Convert running sums to means.""" |
| | if self.A_sum is not None and self.n_samples_A > 0: |
| | self.A = self.A_sum / self.n_samples_A |
| | self.A_sum = None |
| | if self.G_sum is not None and self.n_samples_G > 0: |
| | self.G = self.G_sum / self.n_samples_G |
| | self.G_sum = None |
| |
|
| |
|
| | @dataclass |
| | class KFACConfig: |
| | """Configuration for K-FAC collection.""" |
| | |
| | |
| | target_layers: list[int] = field(default_factory=lambda: [11, 12, 13]) |
| | |
| | |
| | target_projections: list[str] = field(default_factory=lambda: ["gate_proj", "up_proj"]) |
| | |
| | |
| | seq_length: int = 512 |
| | |
| | |
| | exclude_last_position: bool = True |
| | |
| | |
| | sample_labels: bool = True |
| | |
| | |
| | device: str = "cuda" |
| |
|
| |
|
| | class KFACCollector: |
| | """ |
| | Collects K-FAC statistics (activation and gradient covariances) for MLP layers. |
| | |
| | The K-FAC approximation factorizes the Fisher Information Matrix as: |
| | F_W ≈ G ⊗ A = E[gg^T] ⊗ E[aa^T] |
| | |
| | where: |
| | - A is the covariance of activations going into the layer |
| | - G is the covariance of gradients on the layer's output |
| | |
| | Usage: |
| | collector = KFACCollector(model, config) |
| | collector.register_hooks() |
| | |
| | for batch in dataloader: |
| | collector.collect_batch(batch, tokenizer) |
| | |
| | collector.finalize() |
| | collector.save("kfac_stats.pt") |
| | """ |
| | |
| | def __init__( |
| | self, |
| | model: nn.Module, |
| | config: Optional[KFACConfig] = None, |
| | ): |
| | self.model = model |
| | self.config = config or KFACConfig() |
| | |
| | |
| | self.stats: dict[str, LayerStatistics] = {} |
| | |
| | |
| | self._forward_hooks: list = [] |
| | self._backward_hooks: list = [] |
| | |
| | |
| | self._activation_buffer: dict[str, Tensor] = {} |
| | self._gradient_buffer: dict[str, Tensor] = {} |
| | |
| | |
| | self._hooks_registered = False |
| | |
| | def _get_layer_name(self, layer_idx: int, proj_name: str) -> str: |
| | """Generate a unique name for a layer/projection combination.""" |
| | return f"layer_{layer_idx}.{proj_name}" |
| | |
| | def _get_target_modules(self) -> dict[str, nn.Linear]: |
| | """Find all target modules in the model.""" |
| | targets = {} |
| | |
| | |
| | |
| | |
| | layers = None |
| | if hasattr(self.model, "model") and hasattr(self.model.model, "layers"): |
| | |
| | layers = self.model.model.layers |
| | elif hasattr(self.model, "transformer") and hasattr(self.model.transformer, "blocks"): |
| | |
| | layers = self.model.transformer.blocks |
| | elif hasattr(self.model, "layers"): |
| | |
| | layers = self.model.layers |
| | |
| | if layers is None: |
| | raise ValueError( |
| | "Could not find transformer layers. " |
| | "Model architecture not recognized. " |
| | f"Model type: {type(self.model)}" |
| | ) |
| | |
| | for layer_idx in self.config.target_layers: |
| | if layer_idx >= len(layers): |
| | print(f"Warning: Layer {layer_idx} does not exist (model has {len(layers)} layers)") |
| | continue |
| | |
| | layer = layers[layer_idx] |
| | |
| | |
| | mlp = None |
| | if hasattr(layer, "mlp"): |
| | mlp = layer.mlp |
| | elif hasattr(layer, "feed_forward"): |
| | mlp = layer.feed_forward |
| | elif hasattr(layer, "ff"): |
| | mlp = layer.ff |
| | |
| | if mlp is None: |
| | print(f"Warning: Could not find MLP in layer {layer_idx}") |
| | continue |
| | |
| | for proj_name in self.config.target_projections: |
| | if hasattr(mlp, proj_name): |
| | proj = getattr(mlp, proj_name) |
| | if isinstance(proj, nn.Linear): |
| | name = self._get_layer_name(layer_idx, proj_name) |
| | targets[name] = proj |
| | self.stats[name] = LayerStatistics() |
| | else: |
| | print(f"Warning: {proj_name} not found in layer {layer_idx}") |
| | |
| | return targets |
| | |
| | def register_hooks(self) -> None: |
| | """Register forward and backward hooks on target modules.""" |
| | if self._hooks_registered: |
| | print("Hooks already registered") |
| | return |
| | |
| | targets = self._get_target_modules() |
| | |
| | if not targets: |
| | raise ValueError("No target modules found to hook") |
| | |
| | print(f"Registering hooks on {len(targets)} modules:") |
| | for name in targets: |
| | print(f" - {name}") |
| | |
| | for name, module in targets.items(): |
| | |
| | def make_forward_hook(layer_name: str): |
| | def hook(module: nn.Module, input: tuple, output: Tensor) -> None: |
| | |
| | x = input[0] |
| | if x.requires_grad: |
| | |
| | self._activation_buffer[layer_name] = x.detach() |
| | return hook |
| | |
| | |
| | def make_backward_hook(layer_name: str): |
| | def hook(module: nn.Module, grad_input: tuple, grad_output: tuple) -> None: |
| | |
| | g = grad_output[0] |
| | if g is not None: |
| | self._gradient_buffer[layer_name] = g.detach() |
| | return hook |
| | |
| | fh = module.register_forward_hook(make_forward_hook(name)) |
| | bh = module.register_full_backward_hook(make_backward_hook(name)) |
| | |
| | self._forward_hooks.append(fh) |
| | self._backward_hooks.append(bh) |
| | |
| | self._hooks_registered = True |
| | |
| | def remove_hooks(self) -> None: |
| | """Remove all registered hooks.""" |
| | for hook in self._forward_hooks: |
| | hook.remove() |
| | for hook in self._backward_hooks: |
| | hook.remove() |
| | |
| | self._forward_hooks = [] |
| | self._backward_hooks = [] |
| | self._hooks_registered = False |
| | |
| | def _update_statistics(self) -> None: |
| | """Update running statistics from current buffers.""" |
| | for name in self.stats: |
| | if name in self._activation_buffer: |
| | x = self._activation_buffer[name] |
| | |
| | |
| | |
| | if self.config.exclude_last_position and x.shape[1] > 1: |
| | x = x[:, :-1, :] |
| | |
| | |
| | x_flat = x.reshape(-1, x.shape[-1]) |
| | n_positions = x_flat.shape[0] |
| | |
| | |
| | A_batch = x_flat.T @ x_flat |
| | |
| | |
| | if self.stats[name].A_sum is None: |
| | self.stats[name].A_sum = A_batch |
| | else: |
| | self.stats[name].A_sum = self.stats[name].A_sum + A_batch |
| | self.stats[name].n_samples_A += n_positions |
| | |
| | if name in self._gradient_buffer: |
| | g = self._gradient_buffer[name] |
| | |
| | |
| | |
| | if self.config.exclude_last_position and g.shape[1] > 1: |
| | g = g[:, :-1, :] |
| | |
| | |
| | g_flat = g.reshape(-1, g.shape[-1]) |
| | n_positions = g_flat.shape[0] |
| | |
| | |
| | G_batch = g_flat.T @ g_flat |
| | |
| | |
| | if self.stats[name].G_sum is None: |
| | self.stats[name].G_sum = G_batch |
| | else: |
| | self.stats[name].G_sum = self.stats[name].G_sum + G_batch |
| | self.stats[name].n_samples_G += n_positions |
| | |
| | |
| | self._activation_buffer.clear() |
| | self._gradient_buffer.clear() |
| | |
| | @torch.no_grad() |
| | def _sample_labels(self, logits: Tensor) -> Tensor: |
| | """ |
| | Sample labels from model's predicted distribution. |
| | |
| | For proper FIM computation, we sample ŷ ~ p(y|x) rather than |
| | using ground truth labels. |
| | """ |
| | |
| | probs = torch.softmax(logits, dim=-1) |
| | |
| | sampled = torch.multinomial( |
| | probs.view(-1, probs.shape[-1]), |
| | num_samples=1 |
| | ).view(probs.shape[:-1]) |
| | return sampled |
| | |
| | def collect_batch( |
| | self, |
| | input_ids: Tensor, |
| | attention_mask: Optional[Tensor] = None, |
| | labels: Optional[Tensor] = None, |
| | ) -> float: |
| | """ |
| | Collect K-FAC statistics from a single batch. |
| | |
| | Args: |
| | input_ids: Token IDs (batch, seq_len) |
| | attention_mask: Attention mask (batch, seq_len) |
| | labels: Ground truth labels (optional, will be sampled if not provided |
| | or if config.sample_labels is True) |
| | |
| | Returns: |
| | Loss value for this batch |
| | """ |
| | self.model.train() |
| | |
| | |
| | input_ids = input_ids.to(self.config.device) |
| | if attention_mask is not None: |
| | attention_mask = attention_mask.to(self.config.device) |
| | |
| | |
| | with torch.enable_grad(): |
| | outputs = self.model( |
| | input_ids=input_ids, |
| | attention_mask=attention_mask, |
| | use_cache=False, |
| | ) |
| | logits = outputs.logits |
| | |
| | |
| | if self.config.sample_labels or labels is None: |
| | |
| | sampled_labels = self._sample_labels(logits) |
| | |
| | shift_labels = sampled_labels[:, 1:].contiguous() |
| | shift_logits = logits[:, :-1, :].contiguous() |
| | else: |
| | |
| | shift_labels = labels[:, 1:].contiguous().to(self.config.device) |
| | shift_logits = logits[:, :-1, :].contiguous() |
| | |
| | |
| | loss_fn = nn.CrossEntropyLoss() |
| | loss = loss_fn( |
| | shift_logits.view(-1, shift_logits.shape[-1]), |
| | shift_labels.view(-1) |
| | ) |
| | |
| | |
| | loss.backward() |
| | |
| | |
| | self._update_statistics() |
| | |
| | |
| | self.model.zero_grad() |
| | |
| | return loss.item() |
| | |
| | def collect_from_dataloader( |
| | self, |
| | dataloader, |
| | max_tokens: int = 20_000_000, |
| | progress_bar: bool = True, |
| | ) -> dict: |
| | """ |
| | Collect K-FAC statistics from a dataloader. |
| | |
| | Args: |
| | dataloader: PyTorch DataLoader yielding batches with input_ids |
| | max_tokens: Maximum number of tokens to process |
| | progress_bar: Whether to show progress bar |
| | |
| | Returns: |
| | Dictionary with collection statistics |
| | """ |
| | if not self._hooks_registered: |
| | self.register_hooks() |
| | |
| | total_tokens = 0 |
| | total_loss = 0.0 |
| | n_batches = 0 |
| | |
| | iterator = tqdm(dataloader, desc="Collecting K-FAC stats") if progress_bar else dataloader |
| | |
| | for batch in iterator: |
| | if isinstance(batch, dict): |
| | input_ids = batch["input_ids"] |
| | attention_mask = batch.get("attention_mask") |
| | else: |
| | input_ids = batch[0] |
| | attention_mask = batch[1] if len(batch) > 1 else None |
| | |
| | batch_tokens = input_ids.numel() |
| | |
| | loss = self.collect_batch(input_ids, attention_mask) |
| | |
| | total_tokens += batch_tokens |
| | total_loss += loss |
| | n_batches += 1 |
| | |
| | if progress_bar: |
| | iterator.set_postfix({ |
| | "tokens": f"{total_tokens/1e6:.1f}M", |
| | "loss": f"{loss:.3f}" |
| | }) |
| | |
| | if total_tokens >= max_tokens: |
| | break |
| | |
| | return { |
| | "total_tokens": total_tokens, |
| | "n_batches": n_batches, |
| | "avg_loss": total_loss / max(n_batches, 1), |
| | } |
| | |
| | def finalize(self) -> None: |
| | """Finalize statistics by converting sums to means.""" |
| | for name, stat in self.stats.items(): |
| | stat.finalize() |
| | print(f"Finalized {name}: A={stat.A.shape if stat.A is not None else None}, " |
| | f"G={stat.G.shape if stat.G is not None else None}") |
| | |
| | def save(self, path: str) -> None: |
| | """Save K-FAC statistics to file.""" |
| | save_dict = { |
| | "config": { |
| | "target_layers": self.config.target_layers, |
| | "target_projections": self.config.target_projections, |
| | "seq_length": self.config.seq_length, |
| | }, |
| | "statistics": {} |
| | } |
| | |
| | for name, stat in self.stats.items(): |
| | save_dict["statistics"][name] = { |
| | "A": stat.A.cpu() if stat.A is not None else None, |
| | "G": stat.G.cpu() if stat.G is not None else None, |
| | "n_samples_A": stat.n_samples_A, |
| | "n_samples_G": stat.n_samples_G, |
| | } |
| | |
| | torch.save(save_dict, path) |
| | print(f"Saved K-FAC statistics to {path}") |
| | |
| | @classmethod |
| | def load(cls, path: str, model: nn.Module) -> "KFACCollector": |
| | """Load K-FAC statistics from file.""" |
| | data = torch.load(path, map_location="cpu") |
| | |
| | config = KFACConfig( |
| | target_layers=data["config"]["target_layers"], |
| | target_projections=data["config"]["target_projections"], |
| | seq_length=data["config"]["seq_length"], |
| | ) |
| | |
| | collector = cls(model, config) |
| | |
| | for name, stat_data in data["statistics"].items(): |
| | collector.stats[name] = LayerStatistics( |
| | A=stat_data["A"], |
| | G=stat_data["G"], |
| | n_samples_A=stat_data["n_samples_A"], |
| | n_samples_G=stat_data["n_samples_G"], |
| | ) |
| | |
| | print(f"Loaded K-FAC statistics from {path}") |
| | return collector |
| | |
| | def get_statistics(self) -> dict[str, tuple[Tensor, Tensor]]: |
| | """ |
| | Get computed A and G matrices for all layers. |
| | |
| | Returns: |
| | Dictionary mapping layer names to (A, G) tuples |
| | """ |
| | result = {} |
| | for name, stat in self.stats.items(): |
| | if stat.A is not None and stat.G is not None: |
| | result[name] = (stat.A, stat.G) |
| | return result |
| |
|
| |
|
| | def create_dataloader( |
| | dataset_name: str = "allenai/c4", |
| | dataset_config: str = "en", |
| | tokenizer = None, |
| | batch_size: int = 4, |
| | seq_length: int = 512, |
| | max_samples: Optional[int] = None, |
| | streaming: bool = True, |
| | shuffle_buffer: int = 10000, |
| | seed: int = 42, |
| | ): |
| | """ |
| | Create a DataLoader for K-FAC collection. |
| | |
| | Args: |
| | dataset_name: HuggingFace dataset name |
| | dataset_config: Dataset configuration/subset name |
| | tokenizer: Tokenizer for the model |
| | batch_size: Batch size |
| | seq_length: Sequence length for tokenization |
| | max_samples: Maximum number of samples to load |
| | streaming: Whether to use streaming mode |
| | shuffle_buffer: Buffer size for streaming shuffle |
| | seed: Random seed |
| | |
| | Returns: |
| | PyTorch DataLoader |
| | """ |
| | from datasets import load_dataset |
| | from torch.utils.data import DataLoader, IterableDataset |
| | |
| | |
| | if dataset_config: |
| | ds = load_dataset(dataset_name, name=dataset_config, split="train", streaming=streaming) |
| | else: |
| | ds = load_dataset(dataset_name, split="train", streaming=streaming) |
| | |
| | if streaming: |
| | ds = ds.shuffle(buffer_size=shuffle_buffer, seed=seed) |
| | |
| | |
| | def tokenize_fn(examples): |
| | |
| | text_column = "text" if "text" in examples else list(examples.keys())[0] |
| | texts = examples[text_column] |
| | |
| | tokenized = tokenizer( |
| | texts, |
| | truncation=True, |
| | max_length=seq_length, |
| | padding="max_length", |
| | return_tensors="pt", |
| | ) |
| | return tokenized |
| | |
| | |
| | class TokenizedIterableDataset(IterableDataset): |
| | def __init__(self, dataset, tokenizer, seq_length, max_samples): |
| | self.dataset = dataset |
| | self.tokenizer = tokenizer |
| | self.seq_length = seq_length |
| | self.max_samples = max_samples |
| | |
| | def __iter__(self): |
| | count = 0 |
| | for example in self.dataset: |
| | if self.max_samples and count >= self.max_samples: |
| | break |
| | |
| | |
| | text = example.get("text", list(example.values())[0]) |
| | if not text: |
| | continue |
| | |
| | |
| | tokens = self.tokenizer( |
| | text, |
| | truncation=True, |
| | max_length=self.seq_length, |
| | padding="max_length", |
| | return_tensors="pt", |
| | ) |
| | |
| | yield { |
| | "input_ids": tokens["input_ids"].squeeze(0), |
| | "attention_mask": tokens["attention_mask"].squeeze(0), |
| | } |
| | count += 1 |
| | |
| | torch_dataset = TokenizedIterableDataset(ds, tokenizer, seq_length, max_samples) |
| | |
| | dataloader = DataLoader( |
| | torch_dataset, |
| | batch_size=batch_size, |
| | num_workers=0, |
| | ) |
| | |
| | return dataloader |
| |
|