The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
[Errno 39] Directory not empty: '/storage/hf-datasets-cache/medium/datasets/68317226470012-config-parquet-and-info-hxue3-autotrain-data-code-9addaf52/hxue3___autotrain-data-code_summarization/default/0.0.0/29b3c541ba1e96bbaf2a38f0cec26b921f2d711d.incomplete'
Error code: UnexpectedError
Need help to make the dataset viewer work? Open a discussion for direct support.
_data_files
list | _fingerprint
string | _format_columns
sequence | _format_kwargs
dict | _format_type
null | _indexes
dict | _output_all_columns
bool | _split
null |
---|---|---|---|---|---|---|---|
[
{
"filename": "dataset.arrow"
}
] | a382f8e6102bede1 | [
"target",
"text"
] | {} | null | {} | false | null |
YAML Metadata
Warning:
The task_categories "conditional-text-generation" is not in the official list: text-classification, token-classification, table-question-answering, question-answering, zero-shot-classification, translation, summarization, feature-extraction, text-generation, text2text-generation, fill-mask, sentence-similarity, text-to-speech, text-to-audio, automatic-speech-recognition, audio-to-audio, audio-classification, voice-activity-detection, depth-estimation, image-classification, object-detection, image-segmentation, text-to-image, image-to-text, image-to-image, image-to-video, unconditional-image-generation, video-classification, reinforcement-learning, robotics, tabular-classification, tabular-regression, tabular-to-text, table-to-text, multiple-choice, text-retrieval, time-series-forecasting, text-to-video, image-text-to-text, visual-question-answering, document-question-answering, zero-shot-image-classification, graph-ml, mask-generation, zero-shot-object-detection, text-to-3d, image-to-3d, image-feature-extraction, other
AutoTrain Dataset for project: code_summarization
Dataset Descritpion
This dataset has been automatically processed by AutoTrain for project code_summarization.
Languages
The BCP-47 code for the dataset's language is en.
Dataset Structure
Data Instances
A sample from this dataset looks as follows:
[
{
"text": "def read(self, table, columns, keyset, index=\"\", limit=0, partition=None):\n \"\"\"Perform a ``St[...]",
"target": "Perform a ``StreamingRead`` API request for rows in a table.\n\n :type table: str\n :para[...]"
},
{
"text": "def maf_somatic_variant_stats(variant, variant_metadata):\n \"\"\"\n Parse out the variant calling [...]",
"target": "Parse out the variant calling statistics for a given variant from a MAF file\n\n Assumes the MAF fo[...]"
}
]
Dataset Fields
The dataset has the following fields (also called "features"):
{
"text": "Value(dtype='string', id=None)",
"target": "Value(dtype='string', id=None)"
}
Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
Split name | Num samples |
---|---|
train | 800 |
valid | 200 |
- Downloads last month
- 2