pffaundez commited on
Commit
13be8c9
1 Parent(s): 948678c

Training in progress, step 500

Browse files
Files changed (3) hide show
  1. README.md +20 -60
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
README.md CHANGED
@@ -6,88 +6,48 @@ tags:
6
  model-index:
7
  - name: trueparagraph.ai-ELECTRA
8
  results: []
9
- language:
10
- - en
11
- metrics:
12
- - accuracy
13
- pipeline_tag: text-classification
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/659ee7cec0c53b7cb5c0afea/1LoHRRtIawlqdVameWeLu.png)
20
-
21
  # trueparagraph.ai-ELECTRA
22
 
23
- This model is a fine-tuned version of [google/electra-base-discriminator](https://huggingface.co/google/electra-base-discriminator) on the "16K-trueparagraph-STEM" dataset.
24
 
25
  ## Model description
26
 
27
- ELECTRA is a transformer-based model pre-trained using a novel approach called "Replaced Token Detection". The model is pre-trained to distinguish "real" input tokens from "fake" input tokens generated by another neural network. This fine-tuned version of ELECTRA is specifically trained on paragraphs from the STEM domain to detect AI-generated text.
28
-
29
- Key characteristics:
30
- - **Architecture**: Transformer-based model
31
- - **Pre-training objective**: Replaced Token Detection
32
- - **Fine-tuning objective**: Binary classification (Human-written vs AI-generated)
33
 
34
  ## Intended uses & limitations
35
 
36
- ### Intended uses
37
- - **AI Text Detection**: Identifying paragraphs in the STEM domain that are generated by AI versus those written by humans.
38
- - **Educational Tools**: Assisting educators in detecting AI-generated content in academic submissions.
39
- - **Research**: Analyzing the effectiveness of AI-generated content detection in STEM-related texts.
40
-
41
- ### Limitations
42
- - **Domain Specificity**: The model is fine-tuned specifically on STEM paragraphs and may not perform as well on texts from other domains.
43
- - **Generalization**: While the model is effective at detecting AI-generated text in STEM, it may not generalize well to other types of AI-generated content outside of its training data.
44
- - **Biases**: The model may inherit biases present in the training data, which could affect its performance and fairness.
45
 
46
  ## Training and evaluation data
47
 
48
- The model was fine-tuned on the "16K-trueparagraph-STEM" dataset, which consists of 16,000 paragraphs from various STEM domains. The dataset includes both human-written and AI-generated paragraphs to provide a balanced training set for the model.
49
-
50
- ### Dataset Details
51
- - **Size**: 16,000 paragraphs
52
- - **Sources**: Academic papers, research articles, and other STEM-related documents.
53
- - **Balance**: Approximately 50% human-written paragraphs and 50% AI-generated paragraphs.
54
 
55
  ## Training procedure
56
 
57
- ### Preprocessing
58
- - **Tokenization**: Texts were tokenized using the ELECTRA tokenizer.
59
- - **Truncation/Padding**: All inputs were truncated or padded to a maximum length of 512 tokens.
60
-
61
- ### Hyperparameters
62
- - **Optimizer**: AdamW
63
- - **Learning Rate**: 5e-5
64
- - **Batch Size**: 16
65
- - **Number of Epochs**: 3
66
-
67
- ### Training
68
- - **Loss Function**: Binary Cross-Entropy Loss
69
- - **Evaluation Metrics**: Accuracy, Precision, Recall, F1-Score, ROC-AUC
70
-
71
- ### Hardware
72
- - **Environment**: Training was conducted on a single NVIDIA Tesla V100 GPU.
73
- - **Training Time**: Approximately 4 hours.
74
-
75
- ### Evaluation
76
- - The model was evaluated on a hold-out validation set consisting of 10% of the total dataset.
77
- - **Validation Results**:
78
- - **Accuracy**: 0.93
79
- - **Precision**: 0.90
80
- - **Recall**: 0.98
81
- - **F1-Score**: 0.94
82
- - **ROC-AUC**: 0.93
83
-
84
- ### Post-processing
85
- - The final model weights were saved and uploaded to Hugging Face Model Hub.
86
- - A model card was created to document the training and evaluation processes, intended uses, and limitations of the model.
87
 
88
  ### Framework versions
89
 
90
  - Transformers 4.42.4
91
  - Pytorch 2.3.1+cu121
92
  - Datasets 2.20.0
93
- - Tokenizers 0.19.1
 
6
  model-index:
7
  - name: trueparagraph.ai-ELECTRA
8
  results: []
 
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
 
 
14
  # trueparagraph.ai-ELECTRA
15
 
16
+ This model is a fine-tuned version of [google/electra-base-discriminator](https://huggingface.co/google/electra-base-discriminator) on the None dataset.
17
 
18
  ## Model description
19
 
20
+ More information needed
 
 
 
 
 
21
 
22
  ## Intended uses & limitations
23
 
24
+ More information needed
 
 
 
 
 
 
 
 
25
 
26
  ## Training and evaluation data
27
 
28
+ More information needed
 
 
 
 
 
29
 
30
  ## Training procedure
31
 
32
+ ### Training hyperparameters
33
+
34
+ The following hyperparameters were used during training:
35
+ - learning_rate: 5e-05
36
+ - train_batch_size: 16
37
+ - eval_batch_size: 16
38
+ - seed: 42
39
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
+ - lr_scheduler_type: linear
41
+ - lr_scheduler_warmup_steps: 500
42
+ - num_epochs: 5
43
+
44
+ ### Training results
45
+
46
+
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
 
48
  ### Framework versions
49
 
50
  - Transformers 4.42.4
51
  - Pytorch 2.3.1+cu121
52
  - Datasets 2.20.0
53
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3efa0dabcded9064205ee39190f3b629f9e01e8ebd599549ba138d17de32fbc
3
  size 437959248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dd9339ff1a4653b727e224507a8e7f867ce6478ef2c5757c01b800dcc5cbd12
3
  size 437959248
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd6e8b19d712eb01a8d67dc134b5e8c8ce33fff792732007c903cd3a85a6cef4
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6704151b875d24de420af95d5bf962930aea9d7eb4133d1a6e754d31c7a7d92
3
  size 5112