File size: 1,943 Bytes
4f4f462
f76cc6c
 
4f4f462
f76cc6c
 
2de3442
 
f76cc6c
 
84ff512
4f4f462
 
 
 
fe49961
4f4f462
 
 
2de3442
4f4f462
cc6af26
4f4f462
 
 
cc6af26
4f4f462
 
 
cc6af26
4f4f462
 
 
cc6af26
4f4f462
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc6af26
4f4f462
 
 
 
 
 
 
 
cc6af26
 
 
 
 
 
da55800
cc6af26
 
 
 
a3c3b1b
 
 
 
 
 
 
 
 
 
 
cc6af26
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
---
language:
- el
tags:
- text
- language-modeling
- bert
- pretraining
- greek-media
- domain-adaptation
pipeline_tag: fill-mask
metrics:
- accuracy
model-index:
- name: greek-media-bert-base-uncased
  results: []
---


# Greek Media BERT (uncased)

This model is a domain-adapted version of [nlpaueb/bert-base-greek-uncased-v1](https://huggingface.co/nlpaueb/bert-base-greek-uncased-v1) on Greek media centric data.

## Model description

Details will be updated soon.

## Intended uses & limitations

Details will be updated soon.

## Training and evaluation data

Details will be updated soon.

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0

### Training results
Details will be updated soon.

### Framework versions

- Transformers 4.21.0.dev0
- Pytorch 1.12.0+cu116
- Tensorflow 2.11.0
- Datasets 2.3.2
- Tokenizers 0.12.1

### Citation

The model has been officially released with the article "PIMA: Parameter-shared Intelligent Media Analytics Framework for Low Resource Language. 
Dimitrios Zaikis, Nikolaos Stylianou and Ioannis Vlahavas. 
In the Special Issue: New Techniques of Machine Learning and Deep Learning in Text Classification, Applied Sciences Journal. 
2023" (https://www.mdpi.com/2174928).

If you use the model, please cite the following:

```bibtex
@Article{app13053265,
  AUTHOR = {Zaikis, Dimitrios and Stylianou, Nikolaos and Vlahavas, Ioannis},
  TITLE = {PIMA: Parameter-Shared Intelligent Media Analytics Framework for Low Resource Languages},
  JOURNAL = {Applied Sciences},
  VOLUME = {13},
  YEAR = {2023},
  NUMBER = {5},
  ARTICLE-NUMBER = {3265},
  URL = {https://www.mdpi.com/2076-3417/13/5/3265},
  ISSN = {2076-3417},
  DOI = {10.3390/app13053265}
}
```