Datasets:
user commited on
Commit ·
fb709ff
1
Parent(s): 29789ce
More metadata
Browse files
README.md
CHANGED
|
@@ -28,10 +28,119 @@ configs:
|
|
| 28 |
path: doc_type_v2_primary/validation.jsonl.zst
|
| 29 |
- split: test
|
| 30 |
path: doc_type_v2_primary/test.jsonl.zst
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
---
|
| 32 |
# Multilingual Document Classification Dataset
|
| 33 |
|
| 34 |
-
This dataset contains **100,000 text passages** across **100 non-English
|
| 35 |
|
| 36 |
Each original text passage is paired with its English translation and has been programmatically annotated with domain, writing genre, and educational classifications to facilitate cross-lingual classification and domain adaptation tasks.
|
| 37 |
|
|
@@ -46,7 +155,7 @@ Each original text passage is paired with its English translation and has been p
|
|
| 46 |
|
| 47 |
The dataset contains subset configurations tailored for specific training objectives.
|
| 48 |
* In the **`main` config**, each original text is stored alongside its English translation within a single row.
|
| 49 |
-
* In **subset configs** (
|
| 50 |
|
| 51 |
## Annotation & Classification Details
|
| 52 |
|
|
@@ -59,7 +168,7 @@ To generate granular metadata for domain, genre, and cognitive level, two primar
|
|
| 59 |
|
| 60 |
| Column Name | Source Model | Description / Purpose |
|
| 61 |
| :--- | :--- | :--- |
|
| 62 |
-
| `nvidia_domain` | NVIDIA Domain Classifier | General topical categorization (
|
| 63 |
| `doc_type_v1_primary` | EAI Distill 0.5B | High-level document genre classification (V1). |
|
| 64 |
| `doc_type_v2_primary` | EAI Distill 0.5B | Refined, granular document type classification (V2). |
|
| 65 |
|
|
@@ -115,9 +224,9 @@ The subset configurations are stripped down to the target `text`, its `language`
|
|
| 115 |
Users should keep the following limitations in mind when utilizing this dataset:
|
| 116 |
|
| 117 |
* **Source Translation Quality:** Since the source texts are derived from `HuggingFaceFW-finetranslations`, any artifacts, vocabulary choices, or grammatical inaccuracies in the underlying translations will carry over.
|
| 118 |
-
* **Language Distribution:** The dataset contains an uniform number of samples across languages. As a result, high-resource languages (
|
| 119 |
* **Class Imbalance:** Certain topical domains and document types are heavily over-represented compared to others. For instance, there are far more promotional news articles than niche categories like culinary recipes.
|
| 120 |
|
| 121 |
-
##
|
| 122 |
|
| 123 |
This dataset is released under the **Open Data Commons Attribution License (ODC-BY)**, matching the terms of the source datasets.
|
|
|
|
| 28 |
path: doc_type_v2_primary/validation.jsonl.zst
|
| 29 |
- split: test
|
| 30 |
path: doc_type_v2_primary/test.jsonl.zst
|
| 31 |
+
language_creators:
|
| 32 |
+
- machine-translated
|
| 33 |
+
- curated
|
| 34 |
+
task_categories:
|
| 35 |
+
- text-classification
|
| 36 |
+
tags:
|
| 37 |
+
- cross-lingual-classification
|
| 38 |
+
- domain-adaptation
|
| 39 |
+
- multilingual
|
| 40 |
+
language:
|
| 41 |
+
- afr
|
| 42 |
+
- als
|
| 43 |
+
- amh
|
| 44 |
+
- arb
|
| 45 |
+
- ars
|
| 46 |
+
- ary
|
| 47 |
+
- arz
|
| 48 |
+
- asm
|
| 49 |
+
- azj
|
| 50 |
+
- bel
|
| 51 |
+
- ben
|
| 52 |
+
- bew
|
| 53 |
+
- bos
|
| 54 |
+
- bul
|
| 55 |
+
- cat
|
| 56 |
+
- ces
|
| 57 |
+
- ckb
|
| 58 |
+
- cmn
|
| 59 |
+
- cym
|
| 60 |
+
- dan
|
| 61 |
+
- deu
|
| 62 |
+
- div
|
| 63 |
+
- ekk
|
| 64 |
+
- ell
|
| 65 |
+
- eng
|
| 66 |
+
- epo
|
| 67 |
+
- eus
|
| 68 |
+
- fao
|
| 69 |
+
- fas
|
| 70 |
+
- fil
|
| 71 |
+
- fin
|
| 72 |
+
- fra
|
| 73 |
+
- fry
|
| 74 |
+
- gle
|
| 75 |
+
- glg
|
| 76 |
+
- guj
|
| 77 |
+
- hau
|
| 78 |
+
- heb
|
| 79 |
+
- hin
|
| 80 |
+
- hrv
|
| 81 |
+
- hun
|
| 82 |
+
- hye
|
| 83 |
+
- ind
|
| 84 |
+
- isl
|
| 85 |
+
- ita
|
| 86 |
+
- jpn
|
| 87 |
+
- kan
|
| 88 |
+
- kat
|
| 89 |
+
- kaz
|
| 90 |
+
- khk
|
| 91 |
+
- khm
|
| 92 |
+
- kin
|
| 93 |
+
- kir
|
| 94 |
+
- kmr
|
| 95 |
+
- kor
|
| 96 |
+
- lao
|
| 97 |
+
- lat
|
| 98 |
+
- lit
|
| 99 |
+
- ltz
|
| 100 |
+
- lvs
|
| 101 |
+
- mal
|
| 102 |
+
- mar
|
| 103 |
+
- mkd
|
| 104 |
+
- mlt
|
| 105 |
+
- mya
|
| 106 |
+
- nld
|
| 107 |
+
- nno
|
| 108 |
+
- nob
|
| 109 |
+
- npi
|
| 110 |
+
- nrm
|
| 111 |
+
- ory
|
| 112 |
+
- pan
|
| 113 |
+
- pbt
|
| 114 |
+
- plt
|
| 115 |
+
- pol
|
| 116 |
+
- por
|
| 117 |
+
- ron
|
| 118 |
+
- rus
|
| 119 |
+
- sin
|
| 120 |
+
- slk
|
| 121 |
+
- slv
|
| 122 |
+
- snd
|
| 123 |
+
- som
|
| 124 |
+
- spa
|
| 125 |
+
- srp
|
| 126 |
+
- swe
|
| 127 |
+
- swh
|
| 128 |
+
- tam
|
| 129 |
+
- tel
|
| 130 |
+
- tgk
|
| 131 |
+
- tha
|
| 132 |
+
- tur
|
| 133 |
+
- ukr
|
| 134 |
+
- urd
|
| 135 |
+
- uzn
|
| 136 |
+
- vie
|
| 137 |
+
- xho
|
| 138 |
+
- yue
|
| 139 |
+
- zsm
|
| 140 |
---
|
| 141 |
# Multilingual Document Classification Dataset
|
| 142 |
|
| 143 |
+
This dataset contains **100,000 text passages** across **100 non-English language-script pairs** sourced from the [`agentlans/HuggingFaceFW-finetranslations-100-languages-sample`]([https://huggingface.co/datasets/agentlans/HuggingFaceFW-finetranslations-100-languages-sample](https://huggingface.co/datasets/agentlans/HuggingFaceFW-finetranslations-100-languages-sample)) collection.
|
| 144 |
|
| 145 |
Each original text passage is paired with its English translation and has been programmatically annotated with domain, writing genre, and educational classifications to facilitate cross-lingual classification and domain adaptation tasks.
|
| 146 |
|
|
|
|
| 155 |
|
| 156 |
The dataset contains subset configurations tailored for specific training objectives.
|
| 157 |
* In the **`main` config**, each original text is stored alongside its English translation within a single row.
|
| 158 |
+
* In **subset configs** (for example, configurations filtered or categorized by specific schema labels), the original texts and their English translations are stored as distinct, individual rows to allow direct training on the target language or translation.
|
| 159 |
|
| 160 |
## Annotation & Classification Details
|
| 161 |
|
|
|
|
| 168 |
|
| 169 |
| Column Name | Source Model | Description / Purpose |
|
| 170 |
| :--- | :--- | :--- |
|
| 171 |
+
| `nvidia_domain` | NVIDIA Domain Classifier | General topical categorization (for example, News, Food & Drink). |
|
| 172 |
| `doc_type_v1_primary` | EAI Distill 0.5B | High-level document genre classification (V1). |
|
| 173 |
| `doc_type_v2_primary` | EAI Distill 0.5B | Refined, granular document type classification (V2). |
|
| 174 |
|
|
|
|
| 224 |
Users should keep the following limitations in mind when utilizing this dataset:
|
| 225 |
|
| 226 |
* **Source Translation Quality:** Since the source texts are derived from `HuggingFaceFW-finetranslations`, any artifacts, vocabulary choices, or grammatical inaccuracies in the underlying translations will carry over.
|
| 227 |
+
* **Language Distribution:** The dataset contains an uniform number of samples across languages. As a result, high-resource languages (for example, Mandarin Chinese) have the same number of rows as lower-resource languages (for example, Assamese).
|
| 228 |
* **Class Imbalance:** Certain topical domains and document types are heavily over-represented compared to others. For instance, there are far more promotional news articles than niche categories like culinary recipes.
|
| 229 |
|
| 230 |
+
## Licence
|
| 231 |
|
| 232 |
This dataset is released under the **Open Data Commons Attribution License (ODC-BY)**, matching the terms of the source datasets.
|