language:
- el
tags:
- text
- language-modeling
- bert
- pretraining
- greek-media
- domain-adaptation
pipeline_tag: fill-mask
metrics:
- accuracy
model-index:
- name: greek-media-bert-base-uncased
results: []
Greek Media BERT (uncased)
This model is a domain-adapted version of nlpaueb/bert-base-greek-uncased-v1 on Greek media centric data.
Model description
Details will be updated soon.
Intended uses & limitations
Details will be updated soon.
Training and evaluation data
Details will be updated soon.
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
Training results
Details will be updated soon.
Framework versions
- Transformers 4.21.0.dev0
- Pytorch 1.12.0+cu116
- Tensorflow 2.11.0
- Datasets 2.3.2
- Tokenizers 0.12.1
Citation
The model has been officially released with the article "PIMA: Parameter-shared Intelligent Media Analytics Framework for Low Resource Language. Dimitrios Zaikis, Nikolaos Stylianou and Ioannis Vlahavas. In the Special Issue: New Techniques of Machine Learning and Deep Learning in Text Classification, Applied Sciences Journal. 2023" (https://www.mdpi.com/).
Article has been accepted for publication. Details will be updated soon.
If you use the model, please cite the following:
@Article{greek-media-bert-base-uncased,
AUTHOR = {Zaikis, Dimitrios and Stylianou, Nikolaos and Vlahavas, Ioannis},
TITLE = {PIMA: Parameter-shared Intelligent Media Analytics Framework for Low Resource Language},
JOURNAL = {Applied Sciences},
VOLUME = {},
YEAR = {2023},
NUMBER = {},
ARTICLE-NUMBER = {},
URL = {https://www.mdpi.com/},
ISSN = {},
DOI = {}
}