File size: 2,045 Bytes
e4ec04a
 
 
7496d1f
 
e4ec04a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
---
tags:
- xmod
- adapter-transformers
- adapterhub:th/cc100
language:
- th
license: "mit"
---

# Adapter `AdapterHub/xmod-base-th_TH` for AdapterHub/xmod-base

An [adapter](https://adapterhub.ml) for the `AdapterHub/xmod-base` model that was trained on the [th/cc100](https://adapterhub.ml/explore/th/cc100/) dataset.

This adapter was created for usage with the **[Adapters](https://github.com/Adapter-Hub/adapters)** library.

## Usage

First, install `adapters`:

```
pip install -U adapters
```

Now, the adapter can be loaded and activated like this:

```python
from adapters import AutoAdapterModel

model = AutoAdapterModel.from_pretrained("AdapterHub/xmod-base")
adapter_name = model.load_adapter("AdapterHub/xmod-base-th_TH", source="hf", set_active=True)
```

## Architecture & Training

This adapter was extracted from the original model checkpoint [facebook/xmod-base](https://huggingface.co/facebook/xmod-base) to allow loading it independently via the Adapters library.
For more information on architecture and training, please refer to the original model card.

## Evaluation results

<!-- Add some description here -->

## Citation

[Lifting the Curse of Multilinguality by Pre-training Modular Transformers (Pfeiffer et al., 2022)](http://dx.doi.org/10.18653/v1/2022.naacl-main.255)

```
@inproceedings{pfeiffer-etal-2022-lifting,
    title = "Lifting the Curse of Multilinguality by Pre-training Modular Transformers",
    author = "Pfeiffer, Jonas  and
      Goyal, Naman  and
      Lin, Xi  and
      Li, Xian  and
      Cross, James  and
      Riedel, Sebastian  and
      Artetxe, Mikel",
    booktitle = "Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
    month = jul,
    year = "2022",
    address = "Seattle, United States",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2022.naacl-main.255",
    doi = "10.18653/v1/2022.naacl-main.255",
    pages = "3479--3495"
}
```