mariav commited on
Commit
baef4d3
1 Parent(s): 1725e9a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -5
README.md CHANGED
@@ -5,6 +5,12 @@ tags:
5
  model-index:
6
  - name: distilbert-base-german-cased-finetuned-amazon-reviews
7
  results: []
 
 
 
 
 
 
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -12,21 +18,23 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # distilbert-base-german-cased-finetuned-amazon-reviews
14
 
15
- This model is a fine-tuned version of [distilbert-base-german-cased](https://huggingface.co/distilbert-base-german-cased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
  - Loss: 3.8874
18
 
19
  ## Model description
20
 
21
- More information needed
22
 
23
  ## Intended uses & limitations
24
 
25
- More information needed
26
 
27
  ## Training and evaluation data
 
 
28
 
29
- More information needed
30
 
31
  ## Training procedure
32
 
@@ -61,4 +69,4 @@ The following hyperparameters were used during training:
61
 
62
  - Transformers 4.27.0
63
  - Pytorch 1.13.1+cu116
64
- - Tokenizers 0.13.2
5
  model-index:
6
  - name: distilbert-base-german-cased-finetuned-amazon-reviews
7
  results: []
8
+ datasets:
9
+ - amazon_reviews_multi
10
+ language:
11
+ - de
12
+ metrics:
13
+ - perplexity
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
 
19
  # distilbert-base-german-cased-finetuned-amazon-reviews
20
 
21
+ This model is a fine-tuned version of [distilbert-base-german-cased](https://huggingface.co/distilbert-base-german-cased) on the amazon_reviews_multi dataset.
22
  It achieves the following results on the evaluation set:
23
  - Loss: 3.8874
24
 
25
  ## Model description
26
 
27
+ This model is a fine-tuned version of distilbert-base-german-cased using the dataset from amazon_reviews_multi (available in Huggin Face). The purpose is to extend the model's domain, which, once fine-tuned, will be modified for the fill-in-the-gaps task. It's related to my other model (fine-tuned-spanish-bert) as a comparison of both performances.
28
 
29
  ## Intended uses & limitations
30
 
31
+ The use is limited to school use and the limitations have to do with the size of the dataset, since it does not allow for a large contribution, a larger dataset would have to be used to get a larger contribution.
32
 
33
  ## Training and evaluation data
34
+ I did a training that gives the training and validation set loss. (It takes a lot of time. If you're using colab, I recommend to use less Epochs because the result does not change too much, and even though the loss is quite high, the performance of the model based on the perplexity is not that bad)
35
+ Also, I checked the perplexity, which is one good measure for Languages Models. The value of the perplexity is considerabily good: 48'78.
36
 
37
+ - Evaluation: I checked the performance of my model in the notebook provided, just by generating examples.
38
 
39
  ## Training procedure
40
 
69
 
70
  - Transformers 4.27.0
71
  - Pytorch 1.13.1+cu116
72
+ - Tokenizers 0.13.2