Andyrasika commited on
Commit
d3686f8
1 Parent(s): 6e02f1b

Upload model

Browse files
Files changed (3) hide show
  1. README.md +6 -6
  2. adapter_config.json +8 -1
  3. adapter_model.safetensors +3 -0
README.md CHANGED
@@ -1,18 +1,20 @@
1
  ---
 
 
2
  license: apache-2.0
3
- base_model: google/vit-base-patch16-224-in21k
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
  - food101
8
  metrics:
9
  - accuracy
 
10
  model-index:
11
  - name: vit-base-patch16-224-in21k-finetuned-lora-food101
12
  results:
13
  - task:
14
- name: Image Classification
15
  type: image-classification
 
16
  dataset:
17
  name: food101
18
  type: food101
@@ -20,11 +22,9 @@ model-index:
20
  split: train
21
  args: default
22
  metrics:
23
- - name: Accuracy
24
- type: accuracy
25
  value: 0.855973597359736
26
- language:
27
- - en
28
  ---
29
 
30
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
1
  ---
2
+ language:
3
+ - en
4
  license: apache-2.0
 
5
  tags:
6
  - generated_from_trainer
7
  datasets:
8
  - food101
9
  metrics:
10
  - accuracy
11
+ base_model: google/vit-base-patch16-224-in21k
12
  model-index:
13
  - name: vit-base-patch16-224-in21k-finetuned-lora-food101
14
  results:
15
  - task:
 
16
  type: image-classification
17
+ name: Image Classification
18
  dataset:
19
  name: food101
20
  type: food101
 
22
  split: train
23
  args: default
24
  metrics:
25
+ - type: accuracy
 
26
  value: 0.855973597359736
27
+ name: Accuracy
 
28
  ---
29
 
30
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
adapter_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "auto_mapping": {
3
  "base_model_class": "ViTForImageClassification",
4
  "parent_library": "transformers.models.vit.modeling_vit"
@@ -10,17 +11,23 @@
10
  "init_lora_weights": true,
11
  "layers_pattern": null,
12
  "layers_to_transform": null,
 
13
  "lora_alpha": 16,
14
  "lora_dropout": 0.1,
 
 
15
  "modules_to_save": [
16
  "classifier"
17
  ],
18
  "peft_type": "LORA",
19
  "r": 16,
 
20
  "revision": null,
21
  "target_modules": [
22
  "query",
23
  "value"
24
  ],
25
- "task_type": null
 
 
26
  }
 
1
  {
2
+ "alpha_pattern": {},
3
  "auto_mapping": {
4
  "base_model_class": "ViTForImageClassification",
5
  "parent_library": "transformers.models.vit.modeling_vit"
 
11
  "init_lora_weights": true,
12
  "layers_pattern": null,
13
  "layers_to_transform": null,
14
+ "loftq_config": {},
15
  "lora_alpha": 16,
16
  "lora_dropout": 0.1,
17
+ "megatron_config": null,
18
+ "megatron_core": "megatron.core",
19
  "modules_to_save": [
20
  "classifier"
21
  ],
22
  "peft_type": "LORA",
23
  "r": 16,
24
+ "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
  "query",
28
  "value"
29
  ],
30
+ "task_type": null,
31
+ "use_dora": false,
32
+ "use_rslora": false
33
  }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e656924ca3ba76fb581e66102f8c9fe9b9d7f29bdc70f56c96758d539f884e5
3
+ size 2677140