ww commited on
Commit
a52378e
1 Parent(s): ab4ff29
Files changed (1) hide show
  1. README.md +13 -9
README.md CHANGED
@@ -31,17 +31,16 @@ model-index:
31
  verified: true
32
  verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMGFkMmI2ODM0NmY5NGNkNmUxYWViOWYxZDNkY2EzYWFmOWI4N2VhYzY5MGEzMTVhOTU4Zjc4YWViOGNjOWJjMCIsInZlcnNpb24iOjF9.fexrU1icJK5_MiifBtZWkeUvpmFISqBLDXSQJ8E6UnrRof-7cU0s4tX_dIsauHWtUpIHMPZCf5dlMWQKXZuAAA
33
  ---
 
34
 
35
- # roberta-base for QA
36
-
37
- This is the best model for QA [roberta-base](https://huggingface.co/roberta-base) model, fine-tuned using the [SQuAD2.0](https://huggingface.co/datasets/squad_v2) dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
38
-
39
 
40
  ## Overview
41
- **Language model:** roberta-base
42
- **Language:** English
43
- **Downstream-task:** Extractive QA
44
- **Training data:** SQuAD 2.0
 
45
  **Eval data:** SQuAD 2.0
46
  **Infrastructure**: 4x Tesla v100
47
 
@@ -57,20 +56,24 @@ lr_schedule = LinearWarmup
57
  warmup_proportion = 0.2
58
  doc_stride=128
59
  max_query_length=64
60
- ```
61
 
62
  ## Usage
63
 
64
  ### In Haystack
 
65
  Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in [Haystack](https://github.com/deepset-ai/haystack/):
 
66
  ```python
67
  reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2")
68
  # or
69
  reader = TransformersReader(model_name_or_path="deepset/roberta-base-squad2",tokenizer="deepset/roberta-base-squad2")
70
  ```
 
71
  For a complete example of ``roberta-base-squad2`` being used for Question Answering, check out the [Tutorials in Haystack Documentation](https://haystack.deepset.ai/tutorials/first-qa-system)
72
 
73
  ### In Transformers
 
74
  ```python
75
  from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
76
 
@@ -90,6 +93,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_name)
90
  ```
91
 
92
  ## Performance
 
93
  Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
94
 
95
  ```
 
31
  verified: true
32
  verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMGFkMmI2ODM0NmY5NGNkNmUxYWViOWYxZDNkY2EzYWFmOWI4N2VhYzY5MGEzMTVhOTU4Zjc4YWViOGNjOWJjMCIsInZlcnNpb24iOjF9.fexrU1icJK5_MiifBtZWkeUvpmFISqBLDXSQJ8E6UnrRof-7cU0s4tX_dIsauHWtUpIHMPZCf5dlMWQKXZuAAA
33
  ---
34
+ # algmon-base for QA
35
 
36
+ This is the base model for QA [roberta-base](https://huggingface.co/roberta-base) model, fine-tuned using the [SQuAD2.0](https://huggingface.co/datasets/squad_v2) dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
 
 
 
37
 
38
  ## Overview
39
+
40
+ **Language model:** roberta-base
41
+ **Language:** English
42
+ **Downstream-task:** Extractive QA
43
+ **Training data:** SQuAD 2.0
44
  **Eval data:** SQuAD 2.0
45
  **Infrastructure**: 4x Tesla v100
46
 
 
56
  warmup_proportion = 0.2
57
  doc_stride=128
58
  max_query_length=64
59
+ ```
60
 
61
  ## Usage
62
 
63
  ### In Haystack
64
+
65
  Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in [Haystack](https://github.com/deepset-ai/haystack/):
66
+
67
  ```python
68
  reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2")
69
  # or
70
  reader = TransformersReader(model_name_or_path="deepset/roberta-base-squad2",tokenizer="deepset/roberta-base-squad2")
71
  ```
72
+
73
  For a complete example of ``roberta-base-squad2`` being used for Question Answering, check out the [Tutorials in Haystack Documentation](https://haystack.deepset.ai/tutorials/first-qa-system)
74
 
75
  ### In Transformers
76
+
77
  ```python
78
  from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
79
 
 
93
  ```
94
 
95
  ## Performance
96
+
97
  Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
98
 
99
  ```