Datasets:
The dataset preview is not available for this split.
Error code: StreamingRowsError Exception: ValueError Message: Cannot seek streaming HTTP file Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py", line 584, in compute_first_rows_response rows = get_rows( File "/src/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py", line 179, in decorator return func(*args, **kwargs) File "/src/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py", line 222, in get_rows ds = load_dataset( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1775, in load_dataset return builder_instance.as_streaming_dataset(split=split) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1245, in as_streaming_dataset splits_generators = {sg.name: sg for sg in self._split_generators(dl_manager)} File "/tmp/modules-cache/datasets_modules/datasets/code_x_glue_tt_text_to_text/eef07f49359a4a30ef8f3fe6802a25f7ef9603e35c54b63235e15bdbdcbfea87/code_x_glue_tt_text_to_text.py", line 103, in _split_generators return self.child._split_generators(dl_manager=dl_manager) File "/tmp/modules-cache/datasets_modules/datasets/code_x_glue_tt_text_to_text/eef07f49359a4a30ef8f3fe6802a25f7ef9603e35c54b63235e15bdbdcbfea87/common.py", line 50, in _split_generators downloaded_files[k] = dl_manager.download_and_extract(v) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py", line 1074, in download_and_extract return self.extract(self.download(url_or_urls)) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py", line 1026, in extract urlpaths = map_nested(self._extract, url_or_urls, map_tuple=True) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 443, in map_nested mapped = [ File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 444, in <listcomp> _single_map_nested((function, obj, types, None, True, None)) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 346, in _single_map_nested return function(data_struct) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py", line 1031, in _extract protocol = _get_extraction_protocol(urlpath, use_auth_token=self.download_config.use_auth_token) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py", line 434, in _get_extraction_protocol return _get_extraction_protocol_with_magic_number(f) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py", line 405, in _get_extraction_protocol_with_magic_number f.seek(0) File "/src/services/worker/.venv/lib/python3.9/site-packages/fsspec/implementations/http.py", line 747, in seek raise ValueError("Cannot seek streaming HTTP file") ValueError: Cannot seek streaming HTTP file
Need help to make the dataset viewer work? Open an discussion for direct support.
Dataset Card for "code_x_glue_tt_text_to_text"
Dataset Summary
CodeXGLUE text-to-text dataset, available at https://github.com/microsoft/CodeXGLUE/tree/main/Text-Text/text-to-text
The dataset we use is crawled and filtered from Microsoft Documentation, whose document located at https://github.com/MicrosoftDocs/.
Supported Tasks and Leaderboards
machine-translation
: The dataset can be used to train a model for translating Technical documentation between languages.
Languages
da_en, lv_en, no_en, zh_en
Dataset Structure
Data Instances
da_en
An example of 'test' looks as follows.
{
"id": 0,
"source": "4 . K\u00f8r modellen , og udgiv den som en webtjeneste .\n",
"target": "4 . Run the model , and publish it as a web service .\n"
}
lv_en
An example of 'train' looks as follows.
{
"id": 0,
"source": "title : Pakalpojumu objektu izveide\n",
"target": "title : Create service objects\n"
}
no_en
An example of 'validation' looks as follows.
{
"id": 0,
"source": "2 . \u00c5pne servicevaren du vil definere komponenter fra en stykkliste for .\n",
"target": "2 . Open the service item for which you want to set up components from a BOM .\n"
}
zh_en
An example of 'validation' looks as follows.
{
"id": 0,
"source": "& # 124 ; MCDUserNotificationReadStateFilterAny & # 124 ; 0 & # 124 ; \u5305\u62ec \u901a\u77e5 , \u800c \u4e0d \u8003\u8651 \u8bfb\u53d6 \u72b6\u6001 \u3002 & # 124 ;\n",
"target": "| MCDUserNotificationReadStateFilterAny | 0 | Include notifications regardless of read state . |\n"
}
Data Fields
In the following each data field in go is explained for each config. The data fields are the same among all splits.
da_en, lv_en, no_en, zh_en
field name | type | description |
---|---|---|
id | int32 | The index of the sample |
source | string | The source language version of the text |
target | string | The target language version of the text |
Data Splits
name | train | validation | test |
---|---|---|---|
da_en | 42701 | 1000 | 1000 |
lv_en | 18749 | 1000 | 1000 |
no_en | 44322 | 1000 | 1000 |
zh_en | 50154 | 1000 | 1000 |
Dataset Creation
Curation Rationale
[More Information Needed]
Source Data
Initial Data Collection and Normalization
[More Information Needed]
Who are the source language producers?
[More Information Needed]
Annotations
Annotation process
[More Information Needed]
Who are the annotators?
[More Information Needed]
Personal and Sensitive Information
[More Information Needed]
Considerations for Using the Data
Social Impact of Dataset
[More Information Needed]
Discussion of Biases
[More Information Needed]
Other Known Limitations
[More Information Needed]
Additional Information
Dataset Curators
https://github.com/microsoft, https://github.com/madlag
Licensing Information
Computational Use of Data Agreement (C-UDA) License.
Citation Information
@article{CodeXGLUE,
title={CodeXGLUE: A Benchmark Dataset and Open Challenge for Code Intelligence},
year={2020},}
Contributions
Thanks to @madlag (and partly also @ncoop57) for adding this dataset.
- Downloads last month
- 509