| { | |
| "_attn_implementation_autoset": true, | |
| "architectures": [ | |
| "FIMSDE" | |
| ], | |
| "dropout_rate": 0.1, | |
| "dtype": "float32", | |
| "finetune": false, | |
| "finetune_detach_diffusion": false, | |
| "finetune_em_steps": 1, | |
| "finetune_num_points": -1, | |
| "finetune_on_sampling_mse": false, | |
| "finetune_on_sampling_nll": false, | |
| "finetune_samples_count": 1, | |
| "finetune_samples_steps": 1, | |
| "hidden_act": { | |
| "name": "torch.nn.GELU" | |
| }, | |
| "hidden_layers": [ | |
| 256, | |
| 256 | |
| ], | |
| "learnable_loss_scales": { | |
| "attention": { | |
| "activation": "torch.nn.GELU", | |
| "dim_feedforward": 1024, | |
| "dropout": 0.1, | |
| "nhead": 8 | |
| }, | |
| "num_res_layers": 8, | |
| "paths_block_attention": false, | |
| "projection": { | |
| "dropout": 0.1, | |
| "hidden_act": { | |
| "name": "torch.nn.GELU" | |
| }, | |
| "hidden_layers": [ | |
| 256, | |
| 256 | |
| ], | |
| "name": "fim.models.blocks.base.MLP" | |
| } | |
| }, | |
| "max_dimension": 3, | |
| "model_embedding_size": 256, | |
| "model_type": "fimsde", | |
| "name": "FIMSDE", | |
| "num_locations_on_path": 32, | |
| "operator": { | |
| "attention": { | |
| "activation": "torch.nn.GELU", | |
| "dim_feedforward": 1024, | |
| "dropout": 0.1, | |
| "nhead": 8 | |
| }, | |
| "num_res_layers": 8, | |
| "paths_block_attention": false, | |
| "projection": { | |
| "dropout": 0.1, | |
| "hidden_act": { | |
| "name": "torch.nn.GELU" | |
| }, | |
| "hidden_layers": [ | |
| 256, | |
| 256 | |
| ], | |
| "name": "fim.models.blocks.base.MLP" | |
| } | |
| }, | |
| "phi_0t": { | |
| "name": "torch.nn.Linear" | |
| }, | |
| "phi_0x": { | |
| "name": "torch.nn.Linear" | |
| }, | |
| "phi_1x": { | |
| "name": "torch.nn.Linear" | |
| }, | |
| "psi_1": { | |
| "layer": { | |
| "activation": "torch.nn.GELU", | |
| "attn_method": "linear", | |
| "dim_feedforward": 1024, | |
| "dropout": 0.1, | |
| "lin_feature_map": "softmax", | |
| "lin_normalize": false, | |
| "nhead": 8 | |
| }, | |
| "name": "CombinedPathTransformer", | |
| "num_layers": 2 | |
| }, | |
| "residual_ff_size": 1024, | |
| "states_norm": { | |
| "name": "fim.models.sde.Standardization" | |
| }, | |
| "times_norm": { | |
| "name": "fim.models.sde.DeltaLogCentering" | |
| }, | |
| "transformer_layer_activation": "gelu", | |
| "transformers_version": "4.57.1" | |
| } | |