igorgavi commited on
Commit
a831713
1 Parent(s): a6781e6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -12
README.md CHANGED
@@ -161,9 +161,14 @@ if __name__ == "__main__":
161
  evaluator.join_all_results()
162
  ```
163
 
164
- ## Training data
 
 
 
 
165
 
166
- In order to train the model, it's transformers were trained with five datasets, which were:
 
167
 
168
  - **Scientific Papers (arXiv + PubMed)**: Cohan et al. (2018) found out that there were only
169
  datasets with short texts (with an average of 600 words) or datasets with longer texts with
@@ -199,18 +204,10 @@ kind of ATS that is aimed at answering the question “What is the document abou
199
  was obtained from BBC articles and each one of them is accompanied by a short gold-standard
200
  summary often written by its very author.
201
 
202
- Each of their documents was summarized through every summarization method applied in the code and evaluated
203
- in comparison with the gold-standard summaries.
204
-
205
- ## Training procedure
206
-
207
-
208
- ### Preprocessing
209
- [PERGUNTAR ARTHUR]
210
-
211
- Hey, look how easy it is to write LaTeX equations in here \\(Ax = b\\) or even $ Ax = b $
212
  ## Evaluation results
213
 
 
 
214
 
215
  Table 2: Results from Pre-trained Longformer + ML models.
216
 
 
161
  evaluator.join_all_results()
162
  ```
163
 
164
+ ### Preprocessing
165
+ [PERGUNTAR ARTHUR]
166
+
167
+ Hey, look how easy it is to write LaTeX equations in here \\(Ax = b\\) or even $ Ax = b $
168
+ ## Datasets
169
 
170
+ In order to evaluate the model, summaries were generated by each of its summarization methods, which
171
+ used as source texts documents achieved from existing datasets. The chosen datasets for evaluation were the following:
172
 
173
  - **Scientific Papers (arXiv + PubMed)**: Cohan et al. (2018) found out that there were only
174
  datasets with short texts (with an average of 600 words) or datasets with longer texts with
 
204
  was obtained from BBC articles and each one of them is accompanied by a short gold-standard
205
  summary often written by its very author.
206
 
 
 
 
 
 
 
 
 
 
 
207
  ## Evaluation results
208
 
209
+ Each of the datasets' documents was summarized through every summarization method applied in the code and evaluated
210
+ in comparison with the gold-standard summaries.
211
 
212
  Table 2: Results from Pre-trained Longformer + ML models.
213