Shresthadev403 commited on
Commit
858e6ec
1 Parent(s): 0bc384d

End of training

Browse files
README.md CHANGED
@@ -17,13 +17,13 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset.
19
  It achieves the following results on the evaluation set:
20
- - eval_loss: 0.7651
21
- - eval_accuracy: 0.8156
22
- - eval_runtime: 154.1129
23
- - eval_samples_per_second: 98.305
24
- - eval_steps_per_second: 6.145
25
- - epoch: 26.4
26
- - step: 25000
27
 
28
  ## Model description
29
 
 
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset.
19
  It achieves the following results on the evaluation set:
20
+ - eval_loss: 0.7609
21
+ - eval_accuracy: 0.8149
22
+ - eval_runtime: 155.6779
23
+ - eval_samples_per_second: 97.316
24
+ - eval_steps_per_second: 6.083
25
+ - epoch: 27.46
26
+ - step: 26000
27
 
28
  ## Model description
29
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c07c9f219ab17dddb048edc81f52ed98de5117148c986cc434f835c984a9b252
3
  size 343528508
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57b51b89e39855e40356cecad14619ced2468fdb8d02c56cef592f87b8733b9d
3
  size 343528508
runs/Feb05_03-08-58_983d148b451b/events.out.tfevents.1707102539.983d148b451b.26.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d08cb3b0549f26d042533282b0adaec04901eddacd691dc2a26a4ad234f3355
3
- size 21382
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e3955422c6c1463171f2ade37cabbb1e2b34ea1916abd958b3a0ae68b6b8a0f
3
+ size 21871
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.8201980198019801,
3
  "best_model_checkpoint": "food-image-classification/checkpoint-22000",
4
- "epoch": 26.399155227032736,
5
  "eval_steps": 1000,
6
- "global_step": 25000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -382,6 +382,21 @@
382
  "eval_samples_per_second": 98.305,
383
  "eval_steps_per_second": 6.145,
384
  "step": 25000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
385
  }
386
  ],
387
  "logging_steps": 1000,
@@ -389,7 +404,7 @@
389
  "num_input_tokens_seen": 0,
390
  "num_train_epochs": 500,
391
  "save_steps": 1000,
392
- "total_flos": 1.240810652804309e+20,
393
  "train_batch_size": 16,
394
  "trial_name": null,
395
  "trial_params": null
 
1
  {
2
  "best_metric": 0.8201980198019801,
3
  "best_model_checkpoint": "food-image-classification/checkpoint-22000",
4
+ "epoch": 27.455121436114045,
5
  "eval_steps": 1000,
6
+ "global_step": 26000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
382
  "eval_samples_per_second": 98.305,
383
  "eval_steps_per_second": 6.145,
384
  "step": 25000
385
+ },
386
+ {
387
+ "epoch": 27.46,
388
+ "learning_rate": 2.7455121436114044e-05,
389
+ "loss": 0.3274,
390
+ "step": 26000
391
+ },
392
+ {
393
+ "epoch": 27.46,
394
+ "eval_accuracy": 0.8148514851485148,
395
+ "eval_loss": 0.760901927947998,
396
+ "eval_runtime": 155.6779,
397
+ "eval_samples_per_second": 97.316,
398
+ "eval_steps_per_second": 6.083,
399
+ "step": 26000
400
  }
401
  ],
402
  "logging_steps": 1000,
 
404
  "num_input_tokens_seen": 0,
405
  "num_train_epochs": 500,
406
  "save_steps": 1000,
407
+ "total_flos": 1.2904433271108772e+20,
408
  "train_batch_size": 16,
409
  "trial_name": null,
410
  "trial_params": null