ydshieh HF staff commited on
Commit
79b26f6
1 Parent(s): 6326ccb

upload TF checkpoint

Browse files
Files changed (8) hide show
  1. .gitattributes +6 -24
  2. README.md +128 -0
  3. config.json +29 -0
  4. merges.txt +0 -0
  5. special_tokens_map.json +1 -0
  6. tf_model.h5 +3 -0
  7. tokenizer_config.json +1 -0
  8. vocab.json +0 -0
.gitattributes CHANGED
@@ -1,27 +1,9 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
  *.bin.* filter=lfs diff=lfs merge=lfs -text
5
- *.bz2 filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.onnx filter=lfs diff=lfs merge=lfs -text
14
- *.ot filter=lfs diff=lfs merge=lfs -text
15
- *.parquet filter=lfs diff=lfs merge=lfs -text
16
- *.pb filter=lfs diff=lfs merge=lfs -text
17
- *.pt filter=lfs diff=lfs merge=lfs -text
18
- *.pth filter=lfs diff=lfs merge=lfs -text
19
- *.rar filter=lfs diff=lfs merge=lfs -text
20
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
21
- *.tar.* filter=lfs diff=lfs merge=lfs -text
22
  *.tflite filter=lfs diff=lfs merge=lfs -text
23
- *.tgz filter=lfs diff=lfs merge=lfs -text
24
- *.xz filter=lfs diff=lfs merge=lfs -text
25
- *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
1
  *.bin.* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
2
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.h5 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
5
  *.tflite filter=lfs diff=lfs merge=lfs -text
6
+ *.tar.gz filter=lfs diff=lfs merge=lfs -text
7
+ *.ot filter=lfs diff=lfs merge=lfs -text
8
+ *.onnx filter=lfs diff=lfs merge=lfs -text
9
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
 
README.md ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ datasets:
4
+ - squad_v2
5
+ license: cc-by-4.0
6
+ ---
7
+
8
+ # roberta-base for QA
9
+
10
+ NOTE: This is version 2 of the model. See [this github issue](https://github.com/deepset-ai/FARM/issues/552) from the FARM repository for an explanation of why we updated. If you'd like to use version 1, specify `revision="v1.0"` when loading the model in Transformers 3.5. For exmaple:
11
+ ```
12
+ model_name = "deepset/roberta-base-squad2"
13
+ pipeline(model=model_name, tokenizer=model_name, revision="v1.0", task="question-answering")
14
+ ```
15
+
16
+ ## Overview
17
+ **Language model:** roberta-base
18
+ **Language:** English
19
+ **Downstream-task:** Extractive QA
20
+ **Training data:** SQuAD 2.0
21
+ **Eval data:** SQuAD 2.0
22
+ **Code:** See [example](https://github.com/deepset-ai/FARM/blob/master/examples/question_answering.py) in [FARM](https://github.com/deepset-ai/FARM/blob/master/examples/question_answering.py)
23
+ **Infrastructure**: 4x Tesla v100
24
+
25
+ ## Hyperparameters
26
+
27
+ ```
28
+ batch_size = 96
29
+ n_epochs = 2
30
+ base_LM_model = "roberta-base"
31
+ max_seq_len = 386
32
+ learning_rate = 3e-5
33
+ lr_schedule = LinearWarmup
34
+ warmup_proportion = 0.2
35
+ doc_stride=128
36
+ max_query_length=64
37
+ ```
38
+
39
+ ## Using a distilled model instead
40
+ Please note that we have also released a distilled version of this model called [deepset/tinyroberta-squad2](https://huggingface.co/deepset/tinyroberta-squad2). The distilled model has a comparable prediction quality and runs at twice the speed of the base model.
41
+
42
+ ## Performance
43
+ Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
44
+
45
+ ```
46
+ "exact": 79.87029394424324,
47
+ "f1": 82.91251169582613,
48
+
49
+ "total": 11873,
50
+ "HasAns_exact": 77.93522267206478,
51
+ "HasAns_f1": 84.02838248389763,
52
+ "HasAns_total": 5928,
53
+ "NoAns_exact": 81.79983179142137,
54
+ "NoAns_f1": 81.79983179142137,
55
+ "NoAns_total": 5945
56
+ ```
57
+
58
+ ## Usage
59
+
60
+ ### In Transformers
61
+ ```python
62
+ from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
63
+
64
+ model_name = "deepset/roberta-base-squad2"
65
+
66
+ # a) Get predictions
67
+ nlp = pipeline('question-answering', model=model_name, tokenizer=model_name)
68
+ QA_input = {
69
+ 'question': 'Why is model conversion important?',
70
+ 'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'
71
+ }
72
+ res = nlp(QA_input)
73
+
74
+ # b) Load model & tokenizer
75
+ model = AutoModelForQuestionAnswering.from_pretrained(model_name)
76
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
77
+ ```
78
+
79
+ ### In FARM
80
+
81
+ ```python
82
+ from farm.modeling.adaptive_model import AdaptiveModel
83
+ from farm.modeling.tokenization import Tokenizer
84
+ from farm.infer import Inferencer
85
+
86
+ model_name = "deepset/roberta-base-squad2"
87
+
88
+ # a) Get predictions
89
+ nlp = Inferencer.load(model_name, task_type="question_answering")
90
+ QA_input = [{"questions": ["Why is model conversion important?"],
91
+ "text": "The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks."}]
92
+ res = nlp.inference_from_dicts(dicts=QA_input, rest_api_schema=True)
93
+
94
+ # b) Load model & tokenizer
95
+ model = AdaptiveModel.convert_from_transformers(model_name, device="cpu", task_type="question_answering")
96
+ tokenizer = Tokenizer.load(model_name)
97
+ ```
98
+
99
+ ### In haystack
100
+ For doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in [haystack](https://github.com/deepset-ai/haystack/):
101
+ ```python
102
+ reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2")
103
+ # or
104
+ reader = TransformersReader(model_name_or_path="deepset/roberta-base-squad2",tokenizer="deepset/roberta-base-squad2")
105
+ ```
106
+
107
+
108
+ ## Authors
109
+ Branden Chan: `branden.chan [at] deepset.ai`
110
+ Timo Möller: `timo.moeller [at] deepset.ai`
111
+ Malte Pietsch: `malte.pietsch [at] deepset.ai`
112
+ Tanay Soni: `tanay.soni [at] deepset.ai`
113
+
114
+ ## About us
115
+ ![deepset logo](https://workablehr.s3.amazonaws.com/uploads/account/logo/476306/logo)
116
+ We bring NLP to the industry via open source!
117
+ Our focus: Industry specific language models & large scale QA systems.
118
+
119
+ Some of our work:
120
+ - [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
121
+ - [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
122
+ - [FARM](https://github.com/deepset-ai/FARM)
123
+ - [Haystack](https://github.com/deepset-ai/haystack/)
124
+
125
+ Get in touch:
126
+ [Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Slack](https://haystack.deepset.ai/community/join) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
127
+
128
+ By the way: [we're hiring!](http://www.deepset.ai/jobs)
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./roberta-base-squad2",
3
+ "architectures": [
4
+ "RobertaForQuestionAnswering"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "language": "english",
17
+ "layer_norm_eps": 1e-05,
18
+ "max_position_embeddings": 514,
19
+ "model_type": "roberta",
20
+ "name": "Roberta",
21
+ "num_attention_heads": 12,
22
+ "num_hidden_layers": 12,
23
+ "pad_token_id": 1,
24
+ "position_embedding_type": "absolute",
25
+ "transformers_version": "4.18.0.dev0",
26
+ "type_vocab_size": 1,
27
+ "use_cache": true,
28
+ "vocab_size": 50265
29
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
1
+ {"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true}}
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdd5e398eeb651eb7a1f19ec698a0ee68745dddd8c035b1187a509e81be9e684
3
+ size 496513360
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
1
+ {"do_lower_case": false, "model_max_length": 512, "full_tokenizer_file": null}
vocab.json ADDED
The diff for this file is too large to render. See raw diff