ValentinB commited on
Commit
bc70586
1 Parent(s): 52ce5e9

Add model files

Browse files
README.md CHANGED
@@ -1,3 +1,29 @@
1
  ---
2
  license: mit
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
  ---
4
+ The model generated in the Enrich4All project.<br>
5
+ Evaluated the perplexity of MLM Task fine-tuned for construction permits related corpus.<br>
6
+ Baseline model: https://huggingface.co/dumitrescustefan/bert-base-romanian-cased-v1 <br>
7
+ Scripts and corpus used for training: https://github.com/racai-ai/e4all-models
8
+
9
+ Corpus
10
+ ---------------
11
+
12
+ The construction authorization corpus is meant to ease the task of interested people to get informed on the legal framework related to activities like building, repairing, extending, and modifying their living environment, or setup of economic activities like establishing commercial or industrial centers. It is aimed as well to ease and reduce the activity of official representatives of regional administrative centers. The corpus is built to comply with the Romanian legislation in this domain and is structured in sets of labeled questions with a single answer each, covering various categories of issues:
13
+ * Construction activities and operations, including industrial structures, which require or do not require authorization,
14
+ * The necessary steps and documents to be acquired according to the Romanian regulations,
15
+ * validity terms,
16
+ * involved costs.
17
+ The data is acquired from two main sources:
18
+ * Internet: official sites, frequently asked questions
19
+ * Personal experiences of people: building permanent or provisory structures, replacing roofs, fences, installing photovoltaic panels, etc.
20
+
21
+ <br><br>
22
+ The construction permits corpus contains 500,351 words in 110 UTF-8 encoded files.
23
+
24
+ Results
25
+ -----------------
26
+ | MLM Task | Perplexity |
27
+ | ---------------------------------| ------------- |
28
+ | Baseline | 6.71 |
29
+ | Construction Permits Fine-tuning | 3.81 |
config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "dumitrescustefan/bert-base-romanian-cased-v1",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_position_embeddings": 512,
15
+ "model_type": "bert",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "pad_token_id": 0,
19
+ "position_embedding_type": "absolute",
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.15.0",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 50000
25
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61e4cb6ee420da7895385b71a5b74b57490785937d028c6a03c54cdb9e748402
3
+ size 498056363
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "special_tokens_map_file": null, "name_or_path": "dumitrescustefan/bert-base-romanian-cased-v1", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "BertTokenizer"}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0fc9e43636370fa25e2c07fb2b186d1f239410fc188d19b79d50aacd8a87d06
3
+ size 2927
vocab.txt ADDED
The diff for this file is too large to render. See raw diff