AbstractPhil commited on
Commit
a98b68c
·
verified ·
1 Parent(s): 5a94b82

Update metrics - Run 20251012_161107

Browse files
Files changed (1) hide show
  1. best_model.json +30 -22
best_model.json CHANGED
@@ -1,35 +1,43 @@
1
  {
2
- "model_name": "David-partial_shared-hierarchical_tree",
3
- "run_id": "20251012_152245",
4
- "timestamp": "2025-10-12T15:40:39.200600",
5
- "best_val_acc": 75.532,
6
- "best_epoch": 8,
7
- "final_train_acc": 89.719294986524,
8
- "final_train_loss": 1.2751819482817055,
9
  "scale_accuracies": {
10
- "256": 74.356,
11
- "512": 75.256,
12
- "768": 74.766,
13
- "1024": 74.7
 
 
 
 
14
  },
15
  "architecture": {
16
- "preset": "balanced",
17
  "sharing_mode": "partial_shared",
18
- "fusion_mode": "hierarchical_tree",
19
  "scales": [
20
- 256,
21
  512,
22
  768,
23
- 1024
 
 
 
 
24
  ],
25
- "feature_dim": 512,
26
  "num_classes": 1000,
27
  "use_belly": true,
28
  "belly_expand": 2.0
29
  },
30
  "training": {
31
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
32
- "model_variant": "clip_vit_laion_b32",
33
  "num_epochs": 10,
34
  "batch_size": 1024,
35
  "learning_rate": 0.001,
@@ -39,10 +47,10 @@
39
  "scheduler": "cosine_restarts"
40
  },
41
  "files": {
42
- "weights_safetensors": "weights/David-partial_shared-hierarchical_tree/20251012_152245/best_model_acc75.53.safetensors",
43
- "weights_pytorch": "weights/David-partial_shared-hierarchical_tree/20251012_152245/best_model.pth",
44
- "config": "weights/David-partial_shared-hierarchical_tree/20251012_152245/david_config.json",
45
- "training_config": "weights/David-partial_shared-hierarchical_tree/20251012_152245/train_config.json",
46
- "tensorboard": "runs/David-partial_shared-hierarchical_tree/20251012_152245/"
47
  }
48
  }
 
1
  {
2
+ "model_name": "David-partial_shared-deep_efficiency",
3
+ "run_id": "20251012_161107",
4
+ "timestamp": "2025-10-12T16:14:46.682522",
5
+ "best_val_acc": 83.714,
6
+ "best_epoch": 0,
7
+ "final_train_acc": 81.03081019102116,
8
+ "final_train_loss": 2.090034657678665,
9
  "scale_accuracies": {
10
+ "384": 83.52,
11
+ "512": 83.54,
12
+ "768": 83.714,
13
+ "1024": 83.484,
14
+ "1280": 83.63,
15
+ "1536": 83.592,
16
+ "1792": 83.622,
17
+ "2048": 83.664
18
  },
19
  "architecture": {
20
+ "preset": "clip_vit_bigg14",
21
  "sharing_mode": "partial_shared",
22
+ "fusion_mode": "deep_efficiency",
23
  "scales": [
24
+ 384,
25
  512,
26
  768,
27
+ 1024,
28
+ 1280,
29
+ 1536,
30
+ 1792,
31
+ 2048
32
  ],
33
+ "feature_dim": 1280,
34
  "num_classes": 1000,
35
  "use_belly": true,
36
  "belly_expand": 2.0
37
  },
38
  "training": {
39
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
40
+ "model_variant": "clip_vit_laion_bigg14",
41
  "num_epochs": 10,
42
  "batch_size": 1024,
43
  "learning_rate": 0.001,
 
47
  "scheduler": "cosine_restarts"
48
  },
49
  "files": {
50
+ "weights_safetensors": "weights/David-partial_shared-deep_efficiency/20251012_161107/best_model_acc83.71.safetensors",
51
+ "weights_pytorch": "weights/David-partial_shared-deep_efficiency/20251012_161107/best_model.pth",
52
+ "config": "weights/David-partial_shared-deep_efficiency/20251012_161107/david_config.json",
53
+ "training_config": "weights/David-partial_shared-deep_efficiency/20251012_161107/train_config.json",
54
+ "tensorboard": "runs/David-partial_shared-deep_efficiency/20251012_161107/"
55
  }
56
  }