hbeadles commited on
Commit
e4b51ff
·
verified ·
1 Parent(s): 841b781

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -0
README.md CHANGED
@@ -1,3 +1,17 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ datasets:
4
+ - qiaojin/PubMedQA
5
+ language:
6
+ - en
7
+ tags:
8
+ - medical
9
  ---
10
+
11
+ *Author - Hayden Beadles*
12
+
13
+ This model is meant to evaluate the results of creating an Encoder / Decoder generative model using BERT. The model is then finetuned on $30000$ samples of the PubMedQA dataset. Instead of being finetuned on the columns question and final_answer, where final_answer is a set of yes / no answers, we instead fine tune on the more challenging long_answer column, which gives a short answer to the question.
14
+
15
+ The model was fine-tuned over 3 epochs, using the Adam learning rate scheduler, with a max length of 128 tokens.
16
+
17
+ The results are to help gauge BERT's abilities to answer (generate an answer) directly to a question, with no context provided. It is meant to evaluate the overall models training and attention towards a more focused topic, to see if BERTs base training gives it any advantages.