AMfeta99 commited on
Commit
50259c4
1 Parent(s): 286264c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -5
README.md CHANGED
@@ -6,6 +6,7 @@ tags:
6
  - generated_from_trainer
7
  datasets:
8
  - imagefolder
 
9
  metrics:
10
  - accuracy
11
  - precision
@@ -36,6 +37,7 @@ model-index:
36
  - name: F1
37
  type: f1
38
  value: 0.9230769230769231
 
39
  ---
40
 
41
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -53,17 +55,25 @@ It achieves the following results on the evaluation set:
53
 
54
  ## Model description
55
 
56
- More information needed
 
 
 
 
57
 
58
  ## Intended uses & limitations
59
 
60
- More information needed
61
 
62
  ## Training and evaluation data
63
 
64
- More information needed
 
 
 
 
 
65
 
66
- ## Training procedure
67
 
68
  ### Training hyperparameters
69
 
@@ -91,4 +101,4 @@ The following hyperparameters were used during training:
91
  - Transformers 4.41.2
92
  - Pytorch 2.3.0+cu121
93
  - Datasets 2.20.0
94
- - Tokenizers 0.19.1
 
6
  - generated_from_trainer
7
  datasets:
8
  - imagefolder
9
+ - Mahadih534/brain-tumor-dataset
10
  metrics:
11
  - accuracy
12
  - precision
 
37
  - name: F1
38
  type: f1
39
  value: 0.9230769230769231
40
+ pipeline_tag: image-classification
41
  ---
42
 
43
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
55
 
56
  ## Model description
57
 
58
+ This model is a fine-tuned version of [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224), which is a Vision Transformer (ViT)
59
+
60
+ ViT model is originaly a transformer encoder model pre-trained and fine-tuned on ImageNet 2012.
61
+ It was introduced in the paper "An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale" by Dosovitskiy et al.
62
+ The model processes images as sequences of 16x16 patches, adding a [CLS] token for classification tasks, and uses absolute position embeddings. Pre-training enables the model to learn rich image representations, which can be leveraged for downstream tasks by adding a linear classifier on top of the [CLS] token. The weights were converted from the timm repository by Ross Wightman.
63
 
64
  ## Intended uses & limitations
65
 
66
+ This must be used for classification of x-ray images of the brain to diagnose of brain tumor.
67
 
68
  ## Training and evaluation data
69
 
70
+ The model was fine-tuned in the dataset [Mahadih534/brain-tumor-dataset](https://huggingface.co/datasets/Mahadih534/brain-tumor-dataset) that contains 253 brain images. This dataset was originally created by Yousef Ghanem.
71
+
72
+ The original dataset was splitted into training and evaluation subsets, 80% for training and 20% for evaluation.
73
+ For robust framework evaluation, the evaluation subset is further split into two equal parts for validation and testing.
74
+ This results in three distinct datasets: training, validation, and testing
75
+
76
 
 
77
 
78
  ### Training hyperparameters
79
 
 
101
  - Transformers 4.41.2
102
  - Pytorch 2.3.0+cu121
103
  - Datasets 2.20.0
104
+ - Tokenizers 0.19.1