jonas-luehrs commited on
Commit
7567fb7
1 Parent(s): 7377a0b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -11,6 +11,10 @@ metrics:
11
  model-index:
12
  - name: bert-base-uncased-MLP-scirepeval-chemistry-LARGE-textCLS-RHEOLOGY-20230913-3
13
  results: []
 
 
 
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -18,7 +22,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # bert-base-uncased-MLP-scirepeval-chemistry-LARGE-textCLS-RHEOLOGY-20230913-3
20
 
21
- This model is a fine-tuned version of [jonas-luehrs/bert-base-uncased-MLP-scirepeval-chemistry-LARGE](https://huggingface.co/jonas-luehrs/bert-base-uncased-MLP-scirepeval-chemistry-LARGE) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
  - Loss: 0.6836
24
  - F1: 0.7805
@@ -67,4 +71,4 @@ The following hyperparameters were used during training:
67
  - Transformers 4.33.1
68
  - Pytorch 2.0.1+cu118
69
  - Datasets 2.14.5
70
- - Tokenizers 0.13.3
 
11
  model-index:
12
  - name: bert-base-uncased-MLP-scirepeval-chemistry-LARGE-textCLS-RHEOLOGY-20230913-3
13
  results: []
14
+ datasets:
15
+ - bluesky333/chemical_language_understanding_benchmark
16
+ language:
17
+ - en
18
  ---
19
 
20
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
22
 
23
  # bert-base-uncased-MLP-scirepeval-chemistry-LARGE-textCLS-RHEOLOGY-20230913-3
24
 
25
+ This model is a fine-tuned version of [jonas-luehrs/bert-base-uncased-MLP-scirepeval-chemistry-LARGE](https://huggingface.co/jonas-luehrs/bert-base-uncased-MLP-scirepeval-chemistry-LARGE) on the RHEOLOGY dataset of the [blue333/chemical_language_understanding_benchmark](https://huggingface.co/datasets/bluesky333/chemical_language_understanding_benchmark).
26
  It achieves the following results on the evaluation set:
27
  - Loss: 0.6836
28
  - F1: 0.7805
 
71
  - Transformers 4.33.1
72
  - Pytorch 2.0.1+cu118
73
  - Datasets 2.14.5
74
+ - Tokenizers 0.13.3