holylovenia commited on
Commit
8eaaf79
1 Parent(s): 7d44c42

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +12 -12
README.md CHANGED
@@ -18,9 +18,9 @@ tags:
18
  - sentiment-analysis
19
  ---
20
 
21
- Democratizing access to natural language processing (NLP) technology is crucial, especially for underrepresented and extremely low-resource languages. Previous research has focused on developing labeled and unlabeled corpora for these languages through online scraping and document translation. While these methods have proven effective and cost-efficient, we have identified limitations in the resulting corpora, including a lack of lexical diversity and cultural relevance to local communities. To address this gap, we conduct a case study on Indonesian local languages. We compare the effectiveness of online scraping, human translation, and paragraph writing by native speakers in constructing datasets. Our findings demonstrate that datasets generated through paragraph writing by native speakers exhibit superior quality in terms of lexical diversity and cultural content. In addition, we present the NusaWrites benchmark, encompassing 12 underrepresented and extremely low-resource languages spoken by millions of individuals in Indonesia. Our empirical experiment results using existing multilingual large language models conclude the need to extend these models to more underrepresented languages.
22
- We introduce a novel high quality human curated corpora, i.e., NusaMenulis, which covers 12 languages spoken in Indonesia. The resource extend the coverage of languages to 5 new languages, i.e., Ambon (abs), Bima (bhp), Makassarese (mak), Palembang / Musi (mui), and Rejang (rej).
23
- For the rhetoric mode classification task, we cover 5 rhetoric modes, i.e., narrative, persuasive, argumentative, descriptive, and expository.
24
 
25
 
26
  ## Languages
@@ -30,25 +30,25 @@ btk, bew, bug, jav, mad, mak, min, mui, rej, sun
30
  ## Supported Tasks
31
 
32
  Sentiment Analysis
33
-
34
  ## Dataset Usage
35
  ### Using `datasets` library
36
  ```
37
- from datasets import load_dataset
38
- dset = datasets.load_dataset("SEACrowd/nusatranslation_senti", trust_remote_code=True)
39
  ```
40
  ### Using `seacrowd` library
41
  ```import seacrowd as sc
42
  # Load the dataset using the default config
43
- dset = sc.load_dataset("nusatranslation_senti", schema="seacrowd")
44
  # Check all available subsets (config names) of the dataset
45
- print(sc.available_config_names("nusatranslation_senti"))
46
  # Load the dataset using a specific config
47
- dset = sc.load_dataset_by_config_name(config_name="<config_name>")
48
  ```
49
-
50
- More details on how to load the `seacrowd` library can be found [here](https://github.com/SEACrowd/seacrowd-datahub?tab=readme-ov-file#how-to-use).
51
-
52
 
53
  ## Dataset Homepage
54
 
 
18
  - sentiment-analysis
19
  ---
20
 
21
+ Democratizing access to natural language processing (NLP) technology is crucial, especially for underrepresented and extremely low-resource languages. Previous research has focused on developing labeled and unlabeled corpora for these languages through online scraping and document translation. While these methods have proven effective and cost-efficient, we have identified limitations in the resulting corpora, including a lack of lexical diversity and cultural relevance to local communities. To address this gap, we conduct a case study on Indonesian local languages. We compare the effectiveness of online scraping, human translation, and paragraph writing by native speakers in constructing datasets. Our findings demonstrate that datasets generated through paragraph writing by native speakers exhibit superior quality in terms of lexical diversity and cultural content. In addition, we present the NusaWrites benchmark, encompassing 12 underrepresented and extremely low-resource languages spoken by millions of individuals in Indonesia. Our empirical experiment results using existing multilingual large language models conclude the need to extend these models to more underrepresented languages.
22
+ We introduce a novel high quality human curated corpora, i.e., NusaMenulis, which covers 12 languages spoken in Indonesia. The resource extend the coverage of languages to 5 new languages, i.e., Ambon (abs), Bima (bhp), Makassarese (mak), Palembang / Musi (mui), and Rejang (rej).
23
+ For the rhetoric mode classification task, we cover 5 rhetoric modes, i.e., narrative, persuasive, argumentative, descriptive, and expository.
24
 
25
 
26
  ## Languages
 
30
  ## Supported Tasks
31
 
32
  Sentiment Analysis
33
+
34
  ## Dataset Usage
35
  ### Using `datasets` library
36
  ```
37
+ from datasets import load_dataset
38
+ dset = datasets.load_dataset("SEACrowd/nusatranslation_senti", trust_remote_code=True)
39
  ```
40
  ### Using `seacrowd` library
41
  ```import seacrowd as sc
42
  # Load the dataset using the default config
43
+ dset = sc.load_dataset("nusatranslation_senti", schema="seacrowd")
44
  # Check all available subsets (config names) of the dataset
45
+ print(sc.available_config_names("nusatranslation_senti"))
46
  # Load the dataset using a specific config
47
+ dset = sc.load_dataset_by_config_name(config_name="<config_name>")
48
  ```
49
+
50
+ More details on how to load the `seacrowd` library can be found [here](https://github.com/SEACrowd/seacrowd-datahub?tab=readme-ov-file#how-to-use).
51
+
52
 
53
  ## Dataset Homepage
54