mwz commited on
Commit
c03a7db
1 Parent(s): 274a7a3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -7
README.md CHANGED
@@ -18,19 +18,28 @@ This model is a fine-tuned version of [urduhack/roberta-urdu-small](https://hugg
18
  It achieves the following results on the evaluation set:
19
  - Loss: 0.4703
20
 
21
- ## Model description
22
 
23
- More information needed
 
 
 
 
24
 
25
- ## Intended uses & limitations
 
26
 
27
- More information needed
28
 
29
- ## Training and evaluation data
30
 
31
- More information needed
32
 
33
- ## Training procedure
 
 
 
 
34
 
35
  ### Training hyperparameters
36
 
 
18
  It achieves the following results on the evaluation set:
19
  - Loss: 0.4703
20
 
21
+ ## Model Details
22
 
23
+ - Model Name: Urdu Sentiment Classification
24
+ - Model Architecture: RobertaForSequenceClassification
25
+ - Base Model: urduhack/roberta-urdu-small
26
+ - Dataset: IMDB Urdu Reviews
27
+ - Task: Sentiment Classification (Positive/Negative)
28
 
29
+ ## Training Procedure
30
+ 1. The model was fine-tuned using the transformers library and the Trainer class from Hugging Face. The training process involved the following steps:
31
 
32
+ 2. Tokenization: The input Urdu text was tokenized using the RobertaTokenizerFast from the "urduhack/roberta-urdu-small" pre-trained model. The texts were padded and truncated to a maximum length of 256 tokens.
33
 
34
+ 3. Model Architecture: The "urduhack/roberta-urdu-small" pre-trained model was loaded as the base model for sequence classification using the RobertaForSequenceClassification class.
35
 
36
+ 4. Training Arguments: The training arguments were set, including the number of training epochs, batch size, learning rate, evaluation strategy, logging strategy, and more.
37
 
38
+ 5. Training: The model was trained on the training dataset using the Trainer class. The training process was performed with gradient-based optimization techniques to minimize the cross-entropy loss between predicted and actual sentiment labels.
39
+
40
+ 6. Evaluation: After each epoch, the model was evaluated on the validation dataset to monitor its performance. The evaluation results, including training loss and validation loss, were logged for analysis.
41
+
42
+ 7. Fine-Tuning: The model parameters were fine-tuned during the training process to optimize its performance on the IMDb Urdu movie reviews sentiment analysis task.
43
 
44
  ### Training hyperparameters
45