Dataset Viewer
View in Dataset Viewer
Viewer
The dataset viewer is not available for this dataset.
Cannot get the config names for the dataset.
Error code: ConfigNamesError Exception: FileNotFoundError Message: Couldn't find a dataset script at /src/services/worker/Vidyuth/marian-finetuned-kde4-en-to-fr/marian-finetuned-kde4-en-to-fr.py or any data file in the same directory. Couldn't find 'Vidyuth/marian-finetuned-kde4-en-to-fr' on the Hugging Face Hub either: FileNotFoundError: No (supported) data files or dataset script found in Vidyuth/marian-finetuned-kde4-en-to-fr. Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 55, in compute_config_names_response for config in sorted(get_dataset_config_names(path=dataset, token=hf_token)) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 351, in get_dataset_config_names dataset_module = dataset_module_factory( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1508, in dataset_module_factory raise FileNotFoundError( FileNotFoundError: Couldn't find a dataset script at /src/services/worker/Vidyuth/marian-finetuned-kde4-en-to-fr/marian-finetuned-kde4-en-to-fr.py or any data file in the same directory. Couldn't find 'Vidyuth/marian-finetuned-kde4-en-to-fr' on the Hugging Face Hub either: FileNotFoundError: No (supported) data files or dataset script found in Vidyuth/marian-finetuned-kde4-en-to-fr.
Need help to make the dataset viewer work? Open a discussion for direct support.
test-marian-finetuned-kde4-en-to-fr
This model is a fine-tuned version of Helsinki-NLP/opus-mt-en-fr on the kde4 dataset. It achieves the following results on the evaluation set:
- Loss: 0.8559
- Bleu: 52.9416
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
- mixed_precision_training: Native AMP
Training results
Framework versions
- Transformers 4.12.0.dev0
- Pytorch 1.8.1+cu111
- Datasets 1.12.2.dev0
- Tokenizers 0.10.3
- Downloads last month
- 0