Fulwa commited on
Commit
d3cab50
1 Parent(s): a22d31b

End of training

Browse files
Files changed (5) hide show
  1. README.md +11 -11
  2. config.json +23 -21
  3. model.safetensors +2 -2
  4. preprocessor_config.json +16 -1
  5. training_args.bin +2 -2
README.md CHANGED
@@ -17,8 +17,8 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.9664
21
- - Accuracy: 0.6
22
 
23
  ## Model description
24
 
@@ -44,20 +44,20 @@ The following hyperparameters were used during training:
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_ratio: 0.1
47
- - training_steps: 100
48
 
49
  ### Training results
50
 
51
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
- |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
- | 2.0855 | 0.38 | 38 | 1.8140 | 0.3 |
54
- | 1.1008 | 1.38 | 76 | 1.1303 | 0.5571 |
55
- | 0.6874 | 2.24 | 100 | 0.9664 | 0.6 |
56
 
57
 
58
  ### Framework versions
59
 
60
- - Transformers 4.36.2
61
- - Pytorch 2.3.0+rocm6.0
62
  - Datasets 2.19.1
63
- - Tokenizers 0.15.2
 
17
 
18
  This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 2.3963
21
+ - Accuracy: 0.1154
22
 
23
  ## Model description
24
 
 
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_ratio: 0.1
47
+ - training_steps: 36
48
 
49
  ### Training results
50
 
51
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
+ |:-------------:|:------:|:----:|:---------------:|:--------:|
53
+ | 2.4449 | 0.3889 | 14 | 2.4051 | 0.0769 |
54
+ | 2.4615 | 1.3889 | 28 | 2.3911 | 0.1154 |
55
+ | 2.4631 | 2.2222 | 36 | 2.3963 | 0.1154 |
56
 
57
 
58
  ### Framework versions
59
 
60
+ - Transformers 4.40.1
61
+ - Pytorch 2.1.2+rocm6.0
62
  - Datasets 2.19.1
63
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -12,31 +12,33 @@
12
  "hidden_dropout_prob": 0.0,
13
  "hidden_size": 768,
14
  "id2label": {
15
- "0": "ApplyEyeMakeup",
16
- "1": "BasketballDunk",
17
- "2": "ApplyLipstick",
18
- "3": "BaseballPitch",
19
- "4": "BenchPress",
20
- "5": "Basketball",
21
- "6": "BalanceBeam",
22
- "7": "BandMarching",
23
- "8": "Archery",
24
- "9": "BabyCrawling"
 
25
  },
26
  "image_size": 224,
27
  "initializer_range": 0.02,
28
  "intermediate_size": 3072,
29
  "label2id": {
30
- "ApplyEyeMakeup": 0,
31
- "ApplyLipstick": 2,
32
- "Archery": 8,
33
- "BabyCrawling": 9,
34
- "BalanceBeam": 6,
35
- "BandMarching": 7,
36
- "BaseballPitch": 3,
37
- "Basketball": 5,
38
- "BasketballDunk": 1,
39
- "BenchPress": 4
 
40
  },
41
  "layer_norm_eps": 1e-12,
42
  "model_type": "videomae",
@@ -49,7 +51,7 @@
49
  "problem_type": "single_label_classification",
50
  "qkv_bias": true,
51
  "torch_dtype": "float32",
52
- "transformers_version": "4.36.2",
53
  "tubelet_size": 2,
54
  "use_mean_pooling": false
55
  }
 
12
  "hidden_dropout_prob": 0.0,
13
  "hidden_size": 768,
14
  "id2label": {
15
+ "0": " Book",
16
+ "1": " Ball",
17
+ "2": " Afternoon",
18
+ "3": "Animal",
19
+ "4": " Boat",
20
+ "5": " Bank",
21
+ "6": " Actor",
22
+ "7": " Blue",
23
+ "8": "child",
24
+ "9": "bathroom",
25
+ "10": " Chair"
26
  },
27
  "image_size": 224,
28
  "initializer_range": 0.02,
29
  "intermediate_size": 3072,
30
  "label2id": {
31
+ " Actor": 6,
32
+ " Afternoon": 2,
33
+ " Ball": 1,
34
+ " Bank": 5,
35
+ " Blue": 7,
36
+ " Boat": 4,
37
+ " Book": 0,
38
+ " Chair": 10,
39
+ "Animal": 3,
40
+ "bathroom": 9,
41
+ "child": 8
42
  },
43
  "layer_norm_eps": 1e-12,
44
  "model_type": "videomae",
 
51
  "problem_type": "single_label_classification",
52
  "qkv_bias": true,
53
  "torch_dtype": "float32",
54
+ "transformers_version": "4.40.1",
55
  "tubelet_size": 2,
56
  "use_mean_pooling": false
57
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9fbf010135175af726b8aeff637245fa77d3843172bd52171c0984808e32c14f
3
- size 344961984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faa8c3932321cd311c0c568725d9395b07e6e76a15514770eed364a3eb655f50
3
+ size 344965060
preprocessor_config.json CHANGED
@@ -1,4 +1,20 @@
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "crop_size": {
3
  "height": 224,
4
  "width": 224
@@ -7,7 +23,6 @@
7
  "do_normalize": true,
8
  "do_rescale": true,
9
  "do_resize": true,
10
- "feature_extractor_type": "VideoMAEFeatureExtractor",
11
  "image_mean": [
12
  0.485,
13
  0.456,
 
1
  {
2
+ "_valid_processor_keys": [
3
+ "videos",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_center_crop",
8
+ "crop_size",
9
+ "do_rescale",
10
+ "rescale_factor",
11
+ "do_normalize",
12
+ "image_mean",
13
+ "image_std",
14
+ "return_tensors",
15
+ "data_format",
16
+ "input_data_format"
17
+ ],
18
  "crop_size": {
19
  "height": 224,
20
  "width": 224
 
23
  "do_normalize": true,
24
  "do_rescale": true,
25
  "do_resize": true,
 
26
  "image_mean": [
27
  0.485,
28
  0.456,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:46a2e1e4777ae1c15e7d44624f1d1ea074a3e33fc19525a122ccf1b4669cb97e
3
- size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:395eb1ad1b9d276373380e3a1a22d522ce93275bb86ae6735c95352aea21b206
3
+ size 5048