Add adapter bert-base-uncased_sts_stackexchange_pfeiffer version cooking_stackexchange_com
Browse files- README.md +205 -0
- adapter_config.json +41 -0
- pytorch_adapter.bin +3 -0
- pytorch_head.bin +3 -0
README.md
ADDED
@@ -0,0 +1,205 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- bert
|
4 |
+
- adapterhub:sts/stackexchange
|
5 |
+
- adapter-transformers
|
6 |
+
license: "apache-2.0"
|
7 |
+
---
|
8 |
+
|
9 |
+
# Adapter `bert-base-uncased_sts_stackexchange_pfeiffer` for bert-base-uncased
|
10 |
+
|
11 |
+
Our adapters from the MultiCQA paper (https://arxiv.org/abs/2010.00980) trained on the different StackExchange forums (see "version") with self-supervised training signals of unlabeled questions.
|
12 |
+
|
13 |
+
|
14 |
+
**This adapter was created for usage with the [Adapters](https://github.com/Adapter-Hub/adapters) library.**
|
15 |
+
|
16 |
+
## Usage
|
17 |
+
|
18 |
+
First, install `adapters`:
|
19 |
+
|
20 |
+
```
|
21 |
+
pip install -U adapters
|
22 |
+
```
|
23 |
+
|
24 |
+
Now, the adapter can be loaded and activated like this:
|
25 |
+
|
26 |
+
```python
|
27 |
+
from adapters import AutoAdapterModel
|
28 |
+
|
29 |
+
model = AutoAdapterModel.from_pretrained("bert-base-uncased")
|
30 |
+
adapter_name = model.load_adapter("AdapterHub/bert-base-uncased_sts_stackexchange_pfeiffer")
|
31 |
+
model.set_active_adapters(adapter_name)
|
32 |
+
```
|
33 |
+
|
34 |
+
## Architecture & Training
|
35 |
+
|
36 |
+
- Adapter architecture: pfeiffer
|
37 |
+
- Prediction head: None
|
38 |
+
- Dataset: [StackExchange QA Similarity](https://arxiv.org/abs/2010.00980)
|
39 |
+
|
40 |
+
## Author Information
|
41 |
+
|
42 |
+
- Author name(s): Andreas Rücklé
|
43 |
+
- Author email: rueckle@ukp.informatik.tu-darmstadt.de
|
44 |
+
- Author links: [Website](http://rueckle.net), [GitHub](https://github.com/arueckle), [Twitter](https://twitter.com/arueckle)
|
45 |
+
|
46 |
+
## Versions
|
47 |
+
- `serverfault_com`
|
48 |
+
- `mechanics_stackexchange_com`
|
49 |
+
- `ux_stackexchange_com`
|
50 |
+
- `history_stackexchange_com`
|
51 |
+
- `judaism_stackexchange_com`
|
52 |
+
- `workplace_stackexchange_com`
|
53 |
+
- `freelancing_stackexchange_com`
|
54 |
+
- `matheducators_stackexchange_com`
|
55 |
+
- `literature_stackexchange_com`
|
56 |
+
- `softwareengineering_stackexchange_com`
|
57 |
+
- `patents_stackexchange_com`
|
58 |
+
- `pets_stackexchange_com`
|
59 |
+
- `softwarerecs_stackexchange_com`
|
60 |
+
- `writers_stackexchange_com`
|
61 |
+
- `gaming_stackexchange_com`
|
62 |
+
- `movies_stackexchange_com`
|
63 |
+
- `scifi_stackexchange_com`
|
64 |
+
- `parenting_stackexchange_com`
|
65 |
+
- `astronomy_stackexchange_com`
|
66 |
+
- `bitcoin_stackexchange_com`
|
67 |
+
- `rpg_stackexchange_com`
|
68 |
+
- `fitness_stackexchange_com`
|
69 |
+
- `cooking_stackexchange_com` **(main)**
|
70 |
+
- `devops_stackexchange_com`
|
71 |
+
- `avp_stackexchange_com`
|
72 |
+
- `economics_stackexchange_com`
|
73 |
+
- `health_stackexchange_com`
|
74 |
+
- `sitecore_stackexchange_com`
|
75 |
+
- `codegolf_stackexchange_com`
|
76 |
+
- `askubuntu_com`
|
77 |
+
- `sqa_stackexchange_com`
|
78 |
+
- `ai_stackexchange_com`
|
79 |
+
- `ell_stackexchange_com`
|
80 |
+
- `tor_stackexchange_com`
|
81 |
+
- `earthscience_stackexchange_com`
|
82 |
+
- `worldbuilding_stackexchange_com`
|
83 |
+
- `cstheory_stackexchange_com`
|
84 |
+
- `gis_stackexchange_com`
|
85 |
+
- `robotics_stackexchange_com`
|
86 |
+
- `sustainability_stackexchange_com`
|
87 |
+
- `mathematica_stackexchange_com`
|
88 |
+
- `hardwarerecs_stackexchange_com`
|
89 |
+
- `opensource_stackexchange_com`
|
90 |
+
- `homebrew_stackexchange_com`
|
91 |
+
- `diy_stackexchange_com`
|
92 |
+
- `lifehacks_stackexchange_com`
|
93 |
+
- `stats_stackexchange_com`
|
94 |
+
- `hsm_stackexchange_com`
|
95 |
+
- `christianity_stackexchange_com`
|
96 |
+
- `boardgames_stackexchange_com`
|
97 |
+
- `unix_stackexchange_com`
|
98 |
+
- `space_stackexchange_com`
|
99 |
+
- `android_stackexchange_com`
|
100 |
+
- `quantumcomputing_stackexchange_com`
|
101 |
+
- `windowsphone_stackexchange_com`
|
102 |
+
- `music_stackexchange_com`
|
103 |
+
- `cs_stackexchange_com`
|
104 |
+
- `networkengineering_stackexchange_com`
|
105 |
+
- `codereview_stackexchange_com`
|
106 |
+
- `crypto_stackexchange_com`
|
107 |
+
- `scicomp_stackexchange_com`
|
108 |
+
- `math_stackexchange_com`
|
109 |
+
- `anime_stackexchange_com`
|
110 |
+
- `superuser_com`
|
111 |
+
- `biology_stackexchange_com`
|
112 |
+
- `photo_stackexchange_com`
|
113 |
+
- `musicfans_stackexchange_com`
|
114 |
+
- `genealogy_stackexchange_com`
|
115 |
+
- `skeptics_stackexchange_com`
|
116 |
+
- `gamedev_stackexchange_com`
|
117 |
+
- `politics_stackexchange_com`
|
118 |
+
- `expatriates_stackexchange_com`
|
119 |
+
- `law_stackexchange_com`
|
120 |
+
- `chemistry_stackexchange_com`
|
121 |
+
- `blender_stackexchange_com`
|
122 |
+
- `apple_stackexchange_com`
|
123 |
+
- `dba_stackexchange_com`
|
124 |
+
- `quant_stackexchange_com`
|
125 |
+
- `elementaryos_stackexchange_com`
|
126 |
+
- `travel_stackexchange_com`
|
127 |
+
- `joomla_stackexchange_com`
|
128 |
+
- `physics_stackexchange_com`
|
129 |
+
- `interpersonal_stackexchange_com`
|
130 |
+
- `pm_stackexchange_com`
|
131 |
+
- `cogsci_stackexchange_com`
|
132 |
+
- `bricks_stackexchange_com`
|
133 |
+
- `ham_stackexchange_com`
|
134 |
+
- `opendata_stackexchange_com`
|
135 |
+
- `computergraphics_stackexchange_com`
|
136 |
+
- `dsp_stackexchange_com`
|
137 |
+
- `bicycles_stackexchange_com`
|
138 |
+
- `drupal_stackexchange_com`
|
139 |
+
- `philosophy_stackexchange_com`
|
140 |
+
- `arduino_stackexchange_com`
|
141 |
+
- `ethereum_stackexchange_com`
|
142 |
+
- `emacs_stackexchange_com`
|
143 |
+
- `woodworking_stackexchange_com`
|
144 |
+
- `vi_stackexchange_com`
|
145 |
+
- `sports_stackexchange_com`
|
146 |
+
- `gardening_stackexchange_com`
|
147 |
+
- `security_stackexchange_com`
|
148 |
+
- `webapps_stackexchange_com`
|
149 |
+
- `english_stackexchange_com`
|
150 |
+
- `datascience_stackexchange_com`
|
151 |
+
- `sharepoint_stackexchange_com`
|
152 |
+
- `expressionengine_stackexchange_com`
|
153 |
+
- `wordpress_stackexchange_com`
|
154 |
+
- `stackoverflow_com`
|
155 |
+
- `sound_stackexchange_com`
|
156 |
+
- `webmasters_stackexchange_com`
|
157 |
+
- `magento_stackexchange_com`
|
158 |
+
- `bioinformatics_stackexchange_com`
|
159 |
+
- `islam_stackexchange_com`
|
160 |
+
- `raspberrypi_stackexchange_com`
|
161 |
+
- `tex_stackexchange_com`
|
162 |
+
- `chess_stackexchange_com`
|
163 |
+
- `outdoors_stackexchange_com`
|
164 |
+
- `retrocomputing_stackexchange_com`
|
165 |
+
- `aviation_stackexchange_com`
|
166 |
+
- `tridion_stackexchange_com`
|
167 |
+
- `mythology_stackexchange_com`
|
168 |
+
- `eosio_stackexchange_com`
|
169 |
+
- `electronics_stackexchange_com`
|
170 |
+
- `money_stackexchange_com`
|
171 |
+
- `linguistics_stackexchange_com`
|
172 |
+
- `stackapps_com`
|
173 |
+
- `3dprinting_stackexchange_com`
|
174 |
+
- `engineering_stackexchange_com`
|
175 |
+
- `puzzling_stackexchange_com`
|
176 |
+
- `academia_stackexchange_com`
|
177 |
+
- `civicrm_stackexchange_com`
|
178 |
+
- `martialarts_stackexchange_com`
|
179 |
+
- `reverseengineering_stackexchange_com`
|
180 |
+
- `hinduism_stackexchange_com`
|
181 |
+
- `monero_stackexchange_com`
|
182 |
+
- `hermeneutics_stackexchange_com`
|
183 |
+
- `buddhism_stackexchange_com`
|
184 |
+
- `craftcms_stackexchange_com`
|
185 |
+
- `salesforce_stackexchange_com`
|
186 |
+
- `poker_stackexchange_com`
|
187 |
+
- `graphicdesign_stackexchange_com`
|
188 |
+
|
189 |
+
## Citation
|
190 |
+
|
191 |
+
```bibtex
|
192 |
+
@inproceedings{rueckle-etal-2020-multicqa,
|
193 |
+
title = "{MultiCQA}: Zero-Shot Transfer of Self-Supervised Text Matching Models on a Massive Scale",
|
194 |
+
author = {R{\"u}ckl{\'e}, Andreas and
|
195 |
+
Pfeiffer, Jonas and
|
196 |
+
Gurevych, Iryna},
|
197 |
+
booktitle = "Proceedings of The 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP-2020)",
|
198 |
+
year = "2020",
|
199 |
+
address = "Virtual Conference",
|
200 |
+
url = "https://arxiv.org/abs/2010.00980",
|
201 |
+
}
|
202 |
+
|
203 |
+
```
|
204 |
+
|
205 |
+
*This adapter has been auto-imported from https://github.com/Adapter-Hub/Hub/blob/master/adapters/ukp/bert-base-uncased_sts_stackexchange_pfeiffer.yaml*.
|
adapter_config.json
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"config": {
|
3 |
+
"adapter_residual_before_ln": false,
|
4 |
+
"cross_adapter": false,
|
5 |
+
"dropout": 0.0,
|
6 |
+
"factorized_phm_W": true,
|
7 |
+
"factorized_phm_rule": false,
|
8 |
+
"hypercomplex_nonlinearity": "glorot-uniform",
|
9 |
+
"init_weights": "bert",
|
10 |
+
"inv_adapter": null,
|
11 |
+
"inv_adapter_reduction_factor": null,
|
12 |
+
"is_parallel": false,
|
13 |
+
"learn_phm": true,
|
14 |
+
"leave_out": [],
|
15 |
+
"ln_after": false,
|
16 |
+
"ln_before": false,
|
17 |
+
"mh_adapter": false,
|
18 |
+
"non_linearity": "relu",
|
19 |
+
"original_ln_after": true,
|
20 |
+
"original_ln_before": true,
|
21 |
+
"output_adapter": true,
|
22 |
+
"phm_bias": true,
|
23 |
+
"phm_c_init": "normal",
|
24 |
+
"phm_dim": 4,
|
25 |
+
"phm_init_range": 0.0001,
|
26 |
+
"phm_layer": false,
|
27 |
+
"phm_rank": 1,
|
28 |
+
"reduction_factor": 12,
|
29 |
+
"residual_before_ln": true,
|
30 |
+
"scaling": 1.0,
|
31 |
+
"shared_W_phm": false,
|
32 |
+
"shared_phm_rule": true,
|
33 |
+
"use_gating": false
|
34 |
+
},
|
35 |
+
"hidden_size": 768,
|
36 |
+
"model_class": "BertAdapterModel",
|
37 |
+
"model_name": "bert-base-uncased",
|
38 |
+
"model_type": "bert",
|
39 |
+
"name": "cooking_stackexchange_com",
|
40 |
+
"version": "0.2.0"
|
41 |
+
}
|
pytorch_adapter.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9193ea08537c3cd33233e0fc3316b1a52800d8979cf90f5c201d48579f714ac0
|
3 |
+
size 4776166
|
pytorch_head.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:733b809d7effd1540811f0ef1c45b7c5b6d50e426ec4c0eeb7b6d7686dda5afe
|
3 |
+
size 3562
|