versae commited on
Commit
3e9ac41
1 Parent(s): d0eabf1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -5
README.md CHANGED
@@ -1,6 +1,4 @@
1
  ---
2
- YAML tags:
3
-
4
  annotations_creators:
5
  - no-annotation
6
  language_creators:
@@ -22,7 +20,7 @@ task_ids:
22
  - speech-modeling
23
 
24
  ---
25
- # Dataset Card for NBAiLab/NPSC
26
 
27
 
28
  ## Table of Contents
@@ -54,13 +52,13 @@ The Norwegian Parliament Speech Corpus (NPSC) is a corpus for training a Norwegi
54
  ## How to Use
55
  ```python
56
  from datasets import load_dataset
57
- data = load_dataset("nb/NPSC", streaming=True)
58
  ```
59
  ## Download Data
60
  If you do not want to use the HuggingFace Dataset-library for training, or if you want to do additional pre-processing, it is also possible to download the files locally.
61
  ```bash
62
  # Clone the training set
63
- git clone https://huggingface.co/datasets/nb/NPSC
64
 
65
  # Create one large training file of all shards without unpacking
66
  cat NPSC/data/train*.gz > onefile.json.gz
 
1
  ---
 
 
2
  annotations_creators:
3
  - no-annotation
4
  language_creators:
 
20
  - speech-modeling
21
 
22
  ---
23
+ # Dataset Card for NbAiLab/NPSC
24
 
25
 
26
  ## Table of Contents
 
52
  ## How to Use
53
  ```python
54
  from datasets import load_dataset
55
+ data = load_dataset("NbAiLab/NPSC", streaming=True)
56
  ```
57
  ## Download Data
58
  If you do not want to use the HuggingFace Dataset-library for training, or if you want to do additional pre-processing, it is also possible to download the files locally.
59
  ```bash
60
  # Clone the training set
61
+ git clone https://huggingface.co/datasets/NbAiLab/NPSC
62
 
63
  # Create one large training file of all shards without unpacking
64
  cat NPSC/data/train*.gz > onefile.json.gz