bansilp commited on
Commit
a90d4a1
1 Parent(s): 5991f89

Model save

Browse files
README.md CHANGED
@@ -17,12 +17,12 @@ model-index:
17
  name: imagefolder
18
  type: imagefolder
19
  config: default
20
- split: test
21
  args: default
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
- value: 0.7716666666666666
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +32,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 1.1645
36
- - Accuracy: 0.7717
37
 
38
  ## Model description
39
 
@@ -52,25 +52,25 @@ More information needed
52
  ### Training hyperparameters
53
 
54
  The following hyperparameters were used during training:
55
- - learning_rate: 0.0003
56
  - train_batch_size: 16
57
  - eval_batch_size: 8
58
  - seed: 42
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
- - num_epochs: 1
62
  - mixed_precision_training: Native AMP
63
 
64
  ### Training results
65
 
66
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
67
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
68
- | 0.0072 | 0.91 | 500 | 1.1645 | 0.7717 |
69
 
70
 
71
  ### Framework versions
72
 
73
- - Transformers 4.35.2
74
- - Pytorch 2.1.0+cu121
75
- - Datasets 2.16.0
76
- - Tokenizers 0.15.0
 
17
  name: imagefolder
18
  type: imagefolder
19
  config: default
20
+ split: train
21
  args: default
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
+ value: 0.8907407407407407
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
32
 
33
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 0.4002
36
+ - Accuracy: 0.8907
37
 
38
  ## Model description
39
 
 
52
  ### Training hyperparameters
53
 
54
  The following hyperparameters were used during training:
55
+ - learning_rate: 0.0002
56
  - train_batch_size: 16
57
  - eval_batch_size: 8
58
  - seed: 42
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
+ - num_epochs: 5
62
  - mixed_precision_training: Native AMP
63
 
64
  ### Training results
65
 
66
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
67
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
68
+ | 0.0841 | 3.7 | 1000 | 0.4002 | 0.8907 |
69
 
70
 
71
  ### Framework versions
72
 
73
+ - Transformers 4.38.2
74
+ - Pytorch 2.2.1+cu121
75
+ - Datasets 2.18.0
76
+ - Tokenizers 0.15.2
config.json CHANGED
@@ -14,11 +14,10 @@
14
  "2": "country",
15
  "3": "disco",
16
  "4": "hiphop",
17
- "5": "jazz",
18
- "6": "metal",
19
- "7": "pop",
20
- "8": "reggae",
21
- "9": "rock"
22
  },
23
  "image_size": 224,
24
  "initializer_range": 0.02,
@@ -29,11 +28,10 @@
29
  "country": "2",
30
  "disco": "3",
31
  "hiphop": "4",
32
- "jazz": "5",
33
- "metal": "6",
34
- "pop": "7",
35
- "reggae": "8",
36
- "rock": "9"
37
  },
38
  "layer_norm_eps": 1e-12,
39
  "model_type": "vit",
@@ -44,5 +42,5 @@
44
  "problem_type": "single_label_classification",
45
  "qkv_bias": true,
46
  "torch_dtype": "float32",
47
- "transformers_version": "4.35.2"
48
  }
 
14
  "2": "country",
15
  "3": "disco",
16
  "4": "hiphop",
17
+ "5": "metal",
18
+ "6": "pop",
19
+ "7": "reggae",
20
+ "8": "rock"
 
21
  },
22
  "image_size": 224,
23
  "initializer_range": 0.02,
 
28
  "country": "2",
29
  "disco": "3",
30
  "hiphop": "4",
31
+ "metal": "5",
32
+ "pop": "6",
33
+ "reggae": "7",
34
+ "rock": "8"
 
35
  },
36
  "layer_norm_eps": 1e-12,
37
  "model_type": "vit",
 
42
  "problem_type": "single_label_classification",
43
  "qkv_bias": true,
44
  "torch_dtype": "float32",
45
+ "transformers_version": "4.38.2"
46
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:abbdc009182e7696031293c94c192c61c6f24445dadc81297e28c6d6d7df7961
3
- size 343248584
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e9b32304bec56840d3e04c9bdf0d5c99b1d89352687a840187992b9cad96e44
3
+ size 343245508
runs/Mar18_03-56-25_8e515caeb417/events.out.tfevents.1710734191.8e515caeb417.2485.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:562416510b325b9f777843b28fa5f387a5bc74bf5fc2b9a8a56aafb0d9406d5f
3
+ size 33893
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad6692415b070db68f4a37e66816edaae1e4a06514056f81c53142100cabedcc
3
- size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:076f18a557d228bc4658ceb819d314bb200c1ed02f65448a48d0daa4f673c0a9
3
+ size 4856