SA-Yur-or commited on
Commit
27dc31d
1 Parent(s): 577a013

[doc]: minor fix of readme

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -37,7 +37,7 @@ Couple of articles about this problem: [*Problems with Synthetic Data*](https://
37
 
38
  - **Model type:** The custom architecture for binary sequence classification based on pre-trained RoBERTa, with a single output label.
39
  - **Language(s):** Primarily English.
40
- - **License:** Apache 2.0
41
  - **Finetuned from model:** [RoBERTa Large](https://huggingface.co/FacebookAI/roberta-large)
42
 
43
  ### Model Sources
@@ -116,8 +116,7 @@ A custom architecture was chosen for its ability to perform binary classificatio
116
 
117
  ## Performance
118
 
119
- The model was evaluated on a benchmark collected from the same datasets used for training, alongside a closed subset of SuperAnnotate. \
120
- However, there are no direct intersections of samples between the training data and the benchmark. \
121
  The benchmark comprises 1k samples, with 200 samples per category. \
122
  The model's performance is compared with open-source solutions and popular API detectors in the table below:
123
 
 
37
 
38
  - **Model type:** The custom architecture for binary sequence classification based on pre-trained RoBERTa, with a single output label.
39
  - **Language(s):** Primarily English.
40
+ - **License:** [SAIPL](https://huggingface.co/SuperAnnotate/roberta-large-llm-content-detector/blob/main/LICENSE)
41
  - **Finetuned from model:** [RoBERTa Large](https://huggingface.co/FacebookAI/roberta-large)
42
 
43
  ### Model Sources
 
116
 
117
  ## Performance
118
 
119
+ The model was evaluated on a benchmark consisting of a holdout subset of training data, alongside a closed subset of SuperAnnotate data. \
 
120
  The benchmark comprises 1k samples, with 200 samples per category. \
121
  The model's performance is compared with open-source solutions and popular API detectors in the table below:
122