holylovenia commited on
Commit
9c25c1d
1 Parent(s): 7a04972

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +12 -12
README.md CHANGED
@@ -19,9 +19,9 @@ tags:
19
  - machine-translation
20
  ---
21
 
22
- Democratizing access to natural language processing (NLP) technology is crucial, especially for underrepresented and extremely low-resource languages. Previous research has focused on developing labeled and unlabeled corpora for these languages through online scraping and document translation. While these methods have proven effective and cost-efficient, we have identified limitations in the resulting corpora, including a lack of lexical diversity and cultural relevance to local communities. To address this gap, we conduct a case study on Indonesian local languages. We compare the effectiveness of online scraping, human translation, and paragraph writing by native speakers in constructing datasets. Our findings demonstrate that datasets generated through paragraph writing by native speakers exhibit superior quality in terms of lexical diversity and cultural content. In addition, we present the NusaWrites benchmark, encompassing 12 underrepresented and extremely low-resource languages spoken by millions of individuals in Indonesia. Our empirical experiment results using existing multilingual large language models conclude the need to extend these models to more underrepresented languages.
23
- We introduce a novel high quality human curated corpora, i.e., NusaMenulis, which covers 12 languages spoken in Indonesia. The resource extend the coverage of languages to 5 new languages, i.e., Ambon (abs), Bima (bhp), Makassarese (mak), Palembang / Musi (mui), and Rejang (rej).
24
- For the rhetoric mode classification task, we cover 5 rhetoric modes, i.e., narrative, persuasive, argumentative, descriptive, and expository.
25
 
26
 
27
  ## Languages
@@ -31,25 +31,25 @@ ind, btk, bew, bug, jav, mad, mak, min, mui, rej, sun
31
  ## Supported Tasks
32
 
33
  Machine Translation
34
-
35
  ## Dataset Usage
36
  ### Using `datasets` library
37
  ```
38
- from datasets import load_dataset
39
- dset = datasets.load_dataset("SEACrowd/nusatranslation_mt", trust_remote_code=True)
40
  ```
41
  ### Using `seacrowd` library
42
  ```import seacrowd as sc
43
  # Load the dataset using the default config
44
- dset = sc.load_dataset("nusatranslation_mt", schema="seacrowd")
45
  # Check all available subsets (config names) of the dataset
46
- print(sc.available_config_names("nusatranslation_mt"))
47
  # Load the dataset using a specific config
48
- dset = sc.load_dataset_by_config_name(config_name="<config_name>")
49
  ```
50
-
51
- More details on how to load the `seacrowd` library can be found [here](https://github.com/SEACrowd/seacrowd-datahub?tab=readme-ov-file#how-to-use).
52
-
53
 
54
  ## Dataset Homepage
55
 
 
19
  - machine-translation
20
  ---
21
 
22
+ Democratizing access to natural language processing (NLP) technology is crucial, especially for underrepresented and extremely low-resource languages. Previous research has focused on developing labeled and unlabeled corpora for these languages through online scraping and document translation. While these methods have proven effective and cost-efficient, we have identified limitations in the resulting corpora, including a lack of lexical diversity and cultural relevance to local communities. To address this gap, we conduct a case study on Indonesian local languages. We compare the effectiveness of online scraping, human translation, and paragraph writing by native speakers in constructing datasets. Our findings demonstrate that datasets generated through paragraph writing by native speakers exhibit superior quality in terms of lexical diversity and cultural content. In addition, we present the NusaWrites benchmark, encompassing 12 underrepresented and extremely low-resource languages spoken by millions of individuals in Indonesia. Our empirical experiment results using existing multilingual large language models conclude the need to extend these models to more underrepresented languages.
23
+ We introduce a novel high quality human curated corpora, i.e., NusaMenulis, which covers 12 languages spoken in Indonesia. The resource extend the coverage of languages to 5 new languages, i.e., Ambon (abs), Bima (bhp), Makassarese (mak), Palembang / Musi (mui), and Rejang (rej).
24
+ For the rhetoric mode classification task, we cover 5 rhetoric modes, i.e., narrative, persuasive, argumentative, descriptive, and expository.
25
 
26
 
27
  ## Languages
 
31
  ## Supported Tasks
32
 
33
  Machine Translation
34
+
35
  ## Dataset Usage
36
  ### Using `datasets` library
37
  ```
38
+ from datasets import load_dataset
39
+ dset = datasets.load_dataset("SEACrowd/nusatranslation_mt", trust_remote_code=True)
40
  ```
41
  ### Using `seacrowd` library
42
  ```import seacrowd as sc
43
  # Load the dataset using the default config
44
+ dset = sc.load_dataset("nusatranslation_mt", schema="seacrowd")
45
  # Check all available subsets (config names) of the dataset
46
+ print(sc.available_config_names("nusatranslation_mt"))
47
  # Load the dataset using a specific config
48
+ dset = sc.load_dataset_by_config_name(config_name="<config_name>")
49
  ```
50
+
51
+ More details on how to load the `seacrowd` library can be found [here](https://github.com/SEACrowd/seacrowd-datahub?tab=readme-ov-file#how-to-use).
52
+
53
 
54
  ## Dataset Homepage
55