OneNorth commited on
Commit
430d0fd
1 Parent(s): 341e764

End of training

Browse files
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_accuracy": 0.38064516129032255,
4
+ "eval_loss": 1.8708232641220093,
5
+ "eval_runtime": 27.0879,
6
+ "eval_samples_per_second": 5.722,
7
+ "eval_steps_per_second": 1.44
8
+ }
runs/May20_23-34-46_3ef9e927b965/events.out.tfevents.1716248440.3ef9e927b965.34.1 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1d6bfc12703adba59cc1807aa4adef379d07e5bdaa0707bd1959da8fea0635d
3
- size 405
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bb8c1b6251f0a9031ea980c3d8d01d90990febc16a9338d778a424f3cf40b4e
3
+ size 722
test_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_accuracy": 0.38064516129032255,
4
+ "eval_loss": 1.8708232641220093,
5
+ "eval_runtime": 27.0879,
6
+ "eval_samples_per_second": 5.722,
7
+ "eval_steps_per_second": 1.44
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.34285714285714286,
3
+ "best_model_checkpoint": "videomae-base-finetuned-ucf101-subset/checkpoint-75",
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 75,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.13,
13
+ "grad_norm": 9.491246223449707,
14
+ "learning_rate": 4.850746268656717e-05,
15
+ "loss": 2.2907,
16
+ "step": 10
17
+ },
18
+ {
19
+ "epoch": 0.27,
20
+ "grad_norm": 10.176407814025879,
21
+ "learning_rate": 4.104477611940299e-05,
22
+ "loss": 2.2125,
23
+ "step": 20
24
+ },
25
+ {
26
+ "epoch": 0.4,
27
+ "grad_norm": 10.320022583007812,
28
+ "learning_rate": 3.358208955223881e-05,
29
+ "loss": 2.157,
30
+ "step": 30
31
+ },
32
+ {
33
+ "epoch": 0.53,
34
+ "grad_norm": 9.819656372070312,
35
+ "learning_rate": 2.6119402985074626e-05,
36
+ "loss": 2.013,
37
+ "step": 40
38
+ },
39
+ {
40
+ "epoch": 0.67,
41
+ "grad_norm": 12.585600852966309,
42
+ "learning_rate": 1.865671641791045e-05,
43
+ "loss": 2.0902,
44
+ "step": 50
45
+ },
46
+ {
47
+ "epoch": 0.8,
48
+ "grad_norm": 9.586479187011719,
49
+ "learning_rate": 1.119402985074627e-05,
50
+ "loss": 1.8302,
51
+ "step": 60
52
+ },
53
+ {
54
+ "epoch": 0.93,
55
+ "grad_norm": 18.528013229370117,
56
+ "learning_rate": 3.7313432835820893e-06,
57
+ "loss": 1.9035,
58
+ "step": 70
59
+ },
60
+ {
61
+ "epoch": 1.0,
62
+ "eval_accuracy": 0.34285714285714286,
63
+ "eval_loss": 1.724967360496521,
64
+ "eval_runtime": 13.021,
65
+ "eval_samples_per_second": 5.376,
66
+ "eval_steps_per_second": 1.382,
67
+ "step": 75
68
+ },
69
+ {
70
+ "epoch": 1.0,
71
+ "step": 75,
72
+ "total_flos": 3.73846047031296e+17,
73
+ "train_loss": 2.0445140584309898,
74
+ "train_runtime": 154.4764,
75
+ "train_samples_per_second": 1.942,
76
+ "train_steps_per_second": 0.486
77
+ },
78
+ {
79
+ "epoch": 1.0,
80
+ "eval_accuracy": 0.38064516129032255,
81
+ "eval_loss": 1.8708232641220093,
82
+ "eval_runtime": 29.3555,
83
+ "eval_samples_per_second": 5.28,
84
+ "eval_steps_per_second": 1.329,
85
+ "step": 75
86
+ },
87
+ {
88
+ "epoch": 1.0,
89
+ "eval_accuracy": 0.38064516129032255,
90
+ "eval_loss": 1.8708232641220093,
91
+ "eval_runtime": 27.0879,
92
+ "eval_samples_per_second": 5.722,
93
+ "eval_steps_per_second": 1.44,
94
+ "step": 75
95
+ }
96
+ ],
97
+ "logging_steps": 10,
98
+ "max_steps": 75,
99
+ "num_input_tokens_seen": 0,
100
+ "num_train_epochs": 9223372036854775807,
101
+ "save_steps": 500,
102
+ "total_flos": 3.73846047031296e+17,
103
+ "train_batch_size": 4,
104
+ "trial_name": null,
105
+ "trial_params": null
106
+ }