LysandreJik commited on
Commit
0abccbf
1 Parent(s): 3c565ef
config.json ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_reader": {
3
+ "type": "squad",
4
+ "token_indexers": {
5
+ "token_characters": {
6
+ "type": "characters",
7
+ "character_tokenizer": {
8
+ "byte_encoding": "utf-8",
9
+ "end_tokens": [
10
+ 260
11
+ ],
12
+ "start_tokens": [
13
+ 259
14
+ ]
15
+ },
16
+ "min_padding_length": 5
17
+ },
18
+ "tokens": {
19
+ "type": "single_id",
20
+ "lowercase_tokens": true
21
+ }
22
+ }
23
+ },
24
+ "model": {
25
+ "type": "bidaf",
26
+ "dropout": 0.2,
27
+ "matrix_attention": {
28
+ "type": "linear",
29
+ "combination": "x,y,x*y",
30
+ "tensor_1_dim": 200,
31
+ "tensor_2_dim": 200
32
+ },
33
+ "modeling_layer": {
34
+ "type": "lstm",
35
+ "bidirectional": true,
36
+ "dropout": 0.2,
37
+ "hidden_size": 100,
38
+ "input_size": 800,
39
+ "num_layers": 2
40
+ },
41
+ "num_highway_layers": 2,
42
+ "phrase_layer": {
43
+ "type": "lstm",
44
+ "bidirectional": true,
45
+ "hidden_size": 100,
46
+ "input_size": 200,
47
+ "num_layers": 1
48
+ },
49
+ "span_end_encoder": {
50
+ "type": "lstm",
51
+ "bidirectional": true,
52
+ "hidden_size": 100,
53
+ "input_size": 1400,
54
+ "num_layers": 1
55
+ },
56
+ "text_field_embedder": {
57
+ "token_embedders": {
58
+ "token_characters": {
59
+ "type": "character_encoding",
60
+ "dropout": 0.2,
61
+ "embedding": {
62
+ "embedding_dim": 16,
63
+ "num_embeddings": 262
64
+ },
65
+ "encoder": {
66
+ "type": "cnn",
67
+ "embedding_dim": 16,
68
+ "ngram_filter_sizes": [
69
+ 5
70
+ ],
71
+ "num_filters": 100
72
+ }
73
+ },
74
+ "tokens": {
75
+ "type": "embedding",
76
+ "embedding_dim": 100,
77
+ "pretrained_file": "https://allennlp.s3.amazonaws.com/datasets/glove/glove.6B.100d.txt.gz",
78
+ "trainable": false
79
+ }
80
+ }
81
+ }
82
+ },
83
+ "train_data_path": "https://allennlp.s3.amazonaws.com/datasets/squad/squad-train-v1.1.json",
84
+ "validation_data_path": "https://allennlp.s3.amazonaws.com/datasets/squad/squad-dev-v1.1.json",
85
+ "trainer": {
86
+ "cuda_device": 0,
87
+ "grad_norm": 5,
88
+ "learning_rate_scheduler": {
89
+ "type": "reduce_on_plateau",
90
+ "factor": 0.5,
91
+ "mode": "max",
92
+ "patience": 2
93
+ },
94
+ "num_epochs": 20,
95
+ "optimizer": {
96
+ "type": "adam",
97
+ "betas": [
98
+ 0.9,
99
+ 0.9
100
+ ]
101
+ },
102
+ "patience": 10,
103
+ "validation_metric": "+em"
104
+ },
105
+ "data_loader": {
106
+ "batch_sampler": {
107
+ "type": "bucket",
108
+ "batch_size": 40
109
+ }
110
+ }
111
+ }
vocabulary/non_padded_namespaces.txt ADDED
@@ -0,0 +1,2 @@
 
 
1
+ *labels
2
+ *tags
vocabulary/tokens.txt ADDED
The diff for this file is too large to render. See raw diff
weights.th ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd30ae4bacc172504f44f2b21d1fa50cf71855818316a40028025d36058bc7c7
3
+ size 49249265