kiddothe2b commited on
Commit
762a413
1 Parent(s): 45aff1f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -15,14 +15,14 @@ model-index:
15
 
16
  # Danish Legal LM
17
 
18
- This model is pre-training on a combination of the Danish part of the MultiEURLEX (Chalkidis et al., 2021) dataset comprising EU legislation and two subsets (`retsinformationdk`, `retspraksis`) of the Danish Gigaword Corpus (Derczynski et al., 2021)) comprising legal proceedings.
19
  It achieves the following results on the evaluation set:
20
  - Loss: 0.7302 (up to 128 tokens)
21
  - Loss: 0.7847 (up to 512 tokens)
22
 
23
  ## Model description
24
 
25
- This is a RoBERTa (Liu et al., 2019) model pre-training on Danish legal corpora. It follows a base configurations with 12 Transformer layers, each one with 768 hidden units and 12 attention heads.
26
 
27
  ## Intended uses & limitations
28
 
 
15
 
16
  # Danish Legal LM
17
 
18
+ This model is pre-training on a combination of the Danish part of the MultiEURLEX (Chalkidis et al., 2021) dataset comprising EU legislation and two subsets (`retsinformationdk`, `retspraksis`) of the Danish Gigaword Corpus (Derczynski et al., 2021) comprising legal proceedings.
19
  It achieves the following results on the evaluation set:
20
  - Loss: 0.7302 (up to 128 tokens)
21
  - Loss: 0.7847 (up to 512 tokens)
22
 
23
  ## Model description
24
 
25
+ This is a RoBERTa (Liu et al., 2019) model pre-trained on Danish legal corpora. It follows a base configuration with 12 Transformer layers, each one with 768 hidden units and 12 attention heads.
26
 
27
  ## Intended uses & limitations
28