Jed612 commited on
Commit
3e3cb7d
·
verified ·
1 Parent(s): 42336be

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +99 -0
README.md ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model Card for h10505jd-a63140nd-ED-Opt-B
2
+
3
+ <!-- Provide a quick summary of what the model is/does. -->
4
+
5
+ This is a sequence relation classification model that was trained to
6
+ detect whether a given piece of evidence is relevant to a given claim.
7
+
8
+
9
+ ## Model Details
10
+
11
+ ### Model Description
12
+
13
+ <!-- Provide a longer summary of what this model is. -->
14
+
15
+ This model addresses the Evidence Detection (ED) shared task: given a claim and a piece of evidence, determine if the evidence is relevant to that claim (binary classification). This model has a Bert preprocessor and encoder, that has not been fine-tuned, that feed into a multi layered BLSTM model with self-attention mechanism that was fine-tuned on 21K pairs of texts. The input sequences are concatenated to form a larger input sequence, with each sequence preceded by "CLAIM:" and "EVIDENCE:" respectively.
16
+
17
+ - **Developed by:** James Deslandes and Nikolaos Douranos
18
+ - **Language(s):** English
19
+ - **Model type:** Supervised
20
+ - **Model architecture:** BLSTM
21
+
22
+ ### Model Resources
23
+
24
+ <!-- Provide links where applicable. -->
25
+
26
+ - **Preprocessor:** "https://kaggle.com/models/tensorflow/bert/TensorFlow2/en-uncased-preprocess/3"
27
+ - **Encoder Model:** https://www.kaggle.com/models/tensorflow/bert/TensorFlow2/en-uncased-l-12-h-768-a-12/4
28
+ - **Repo:** https://huggingface.co/Jed612/encoder-BLSTM
29
+
30
+ ## Training Details
31
+
32
+ ### Training Data
33
+
34
+ This model was trained on 21K claim-evidence pairs.
35
+
36
+ ### Training Procedure
37
+
38
+ #### Training Hyperparameters
39
+
40
+ - batch_size: 32
41
+ - epochs: 4
42
+ - learning_rate: 1e-4
43
+
44
+ #### Speeds, Sizes, Times
45
+
46
+ - overall training time: 16 minutes
47
+ - duration per training epoch: 4 minutes
48
+ - model size: 500MB
49
+
50
+ ## Evaluation
51
+
52
+ ### Testing Data & Metrics
53
+
54
+ #### Testing Data
55
+
56
+ A seperate validation dataset of 6K claim-evidence pairs.
57
+
58
+ #### Metrics
59
+
60
+ - ROC AUC
61
+ - Specificity
62
+ - Precision
63
+ - Recall
64
+ - F1-score
65
+ - Accuracy
66
+ - average accuracy over 4 models
67
+
68
+ ### Results
69
+
70
+ The model obtained an ROC AUC of 0.91, a specificity of 92.8%, a precision of 78.1% a recall of 66.6%, an F1-score of 71.9% and an accuracy of 85.6%. Four different models with this structure were trained and their accuracies averaged to 85.4%. The error bars show twice the standard deviation, either side of the mean.
71
+
72
+ **Training and Validation Accuracy and Loss Mean:**
73
+
74
+ ![Graph of Training and Validation Accuracy and Loss Mean](https://external-content.duckduckgo.com/iu/?u=http%3A%2F%2Fdrive.google.com/uc?id=1gi_5a4mfwzQae6J1dX_IXf7p_C5knWeQ)
75
+
76
+ ## Technical Specifications
77
+
78
+ ### Hardware
79
+
80
+
81
+ - RAM: at least 4 GB
82
+ - Storage: at least 50 GB,
83
+ - GPU: T4
84
+
85
+ ### Software
86
+
87
+ - Tensorflow
88
+ - Tensorflow_hub
89
+ - Keras 2
90
+
91
+ ## Bias, Risks, and Limitations
92
+
93
+ Any inputs (concatenation of two sequences) longer than
94
+ 512 subwords will be truncated by the model.
95
+
96
+ ## Additional Information
97
+
98
+ The hyperparameters were determined by experimentation
99
+ with different values.