Vui Seng Chua commited on
Commit
e18b33e
1 Parent(s): 4e43528

Add content

Browse files
README.md ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ tags:
4
+ - generated_from_trainer
5
+ datasets:
6
+ - wikitext
7
+ metrics:
8
+ - accuracy
9
+ model-index:
10
+ - name: ov-opt-350m-8bit-85pc-sparse-kv-cache
11
+ results:
12
+ - task:
13
+ name: Causal Language Modeling
14
+ type: text-generation
15
+ dataset:
16
+ name: wikitext wikitext-2-raw-v1
17
+ type: wikitext
18
+ config: wikitext-2-raw-v1
19
+ split: validation
20
+ args: wikitext-2-raw-v1
21
+ metrics:
22
+ - name: Accuracy
23
+ type: accuracy
24
+ value: 0.02541544477028348
25
+ ---
26
+
27
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
28
+ should probably proofread and complete it, then remove this comment. -->
29
+
30
+ # ov-opt-350m-8bit-85pc-sparse-kv-cache
31
+
32
+ This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m) on the wikitext wikitext-2-raw-v1 dataset.
33
+ It achieves the following results on the evaluation set:
34
+ - Loss: 8.9750
35
+ - Accuracy: 0.0254
36
+
37
+ ## Model description
38
+
39
+ More information needed
40
+
41
+ ## Intended uses & limitations
42
+
43
+ More information needed
44
+
45
+ ## Training and evaluation data
46
+
47
+ More information needed
48
+
49
+ ## Training procedure
50
+
51
+ ### Training hyperparameters
52
+
53
+ The following hyperparameters were used during training:
54
+ - learning_rate: 5e-05
55
+ - train_batch_size: 8
56
+ - eval_batch_size: 1
57
+ - seed: 42
58
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
59
+ - lr_scheduler_type: linear
60
+ - training_steps: 1
61
+
62
+ ### Training results
63
+
64
+
65
+
66
+ ### Framework versions
67
+
68
+ - Transformers 4.30.2
69
+ - Pytorch 2.0.1+cu117
70
+ - Datasets 2.13.1
71
+ - Tokenizers 0.13.3
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.0,
3
+ "eval_accuracy": 0.02541544477028348,
4
+ "eval_loss": 8.974957466125488,
5
+ "eval_runtime": 6.8308,
6
+ "eval_samples": 3,
7
+ "eval_samples_per_second": 0.439,
8
+ "eval_steps_per_second": 0.439,
9
+ "perplexity": 7902.681933731476,
10
+ "train_loss": 10.797767639160156,
11
+ "train_runtime": 59.971,
12
+ "train_samples": 2355,
13
+ "train_samples_per_second": 0.133,
14
+ "train_steps_per_second": 0.017
15
+ }
compressed_graph.dot ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/opt-350m",
3
+ "_remove_final_layer_norm": false,
4
+ "activation_dropout": 0.0,
5
+ "activation_function": "relu",
6
+ "architectures": [
7
+ "OPTForCausalLM"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "bos_token_id": 2,
11
+ "do_layer_norm_before": false,
12
+ "dropout": 0.1,
13
+ "enable_bias": true,
14
+ "eos_token_id": 2,
15
+ "ffn_dim": 4096,
16
+ "hidden_size": 1024,
17
+ "init_std": 0.02,
18
+ "layer_norm_elementwise_affine": true,
19
+ "layerdrop": 0.0,
20
+ "max_position_embeddings": 2048,
21
+ "model_type": "opt",
22
+ "num_attention_heads": 16,
23
+ "num_hidden_layers": 24,
24
+ "pad_token_id": 1,
25
+ "prefix": "</s>",
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.30.2",
28
+ "use_cache": true,
29
+ "vocab_size": 50272,
30
+ "word_embed_proj_dim": 512
31
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.0,
3
+ "eval_accuracy": 0.02541544477028348,
4
+ "eval_loss": 8.974957466125488,
5
+ "eval_runtime": 6.8308,
6
+ "eval_samples": 3,
7
+ "eval_samples_per_second": 0.439,
8
+ "eval_steps_per_second": 0.439,
9
+ "perplexity": 7902.681933731476
10
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 2,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 1,
6
+ "transformers_version": "4.30.2"
7
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
nncf_output.log ADDED
The diff for this file is too large to render. See raw diff
 
openvino_config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "compression": [
3
+ {
4
+ "algorithm": "magnitude_sparsity",
5
+ "ignored_scopes": [
6
+ "{re}.*Embedding",
7
+ "{re}.*lm_head*"
8
+ ],
9
+ "params": {
10
+ "sparsity_level_setting_mode": "layer-wise"
11
+ },
12
+ "sparsity_init": 0.85
13
+ },
14
+ {
15
+ "algorithm": "quantization",
16
+ "export_to_onnx_standard_ops": false,
17
+ "ignored_scopes": [
18
+ "{re}.*Embedding*",
19
+ "{re}.*__add___*",
20
+ "{re}.*layer_norm_*"
21
+ ],
22
+ "initializer": {
23
+ "batchnorm_adaptation": {
24
+ "num_bn_adaptation_samples": 0
25
+ },
26
+ "range": {
27
+ "num_init_samples": 8,
28
+ "type": "mean_min_max"
29
+ }
30
+ },
31
+ "overflow_fix": "disable",
32
+ "scope_overrides": {
33
+ "activations": {
34
+ "activations": {
35
+ "mode": "symmetric"
36
+ },
37
+ "weights": {
38
+ "mode": "symmetric"
39
+ }
40
+ }
41
+ }
42
+ }
43
+ ],
44
+ "input_info": [
45
+ {
46
+ "keyword": "input_ids",
47
+ "sample_size": [
48
+ 8,
49
+ 1024
50
+ ],
51
+ "type": "long"
52
+ },
53
+ {
54
+ "keyword": "attention_mask",
55
+ "sample_size": [
56
+ 8,
57
+ 1024
58
+ ],
59
+ "type": "long"
60
+ }
61
+ ],
62
+ "log_dir": "/data1/vchua/temp/ov-opt-350m-8bit-85pc-sparse-kv-cache",
63
+ "optimum_version": "1.8.8",
64
+ "save_onnx_model": false,
65
+ "transformers_version": "4.30.2"
66
+ }
openvino_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78390eeb043f7b4dfd18a58cd1d5a0a5f25b83aeaf36c03e1591bac3a8cf0980
3
+ size 442503316
openvino_model.xml ADDED
The diff for this file is too large to render. See raw diff
 
original_graph.dot ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:770487bbb29d130751fa27906a76ea3ec4d835231872d2aa50f852e97c3a73ad
3
+ size 2538846347
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "</s>",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_prefix_space": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "</s>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": true,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "errors": "replace",
22
+ "model_max_length": 1000000000000000019884624838656,
23
+ "pad_token": {
24
+ "__type": "AddedToken",
25
+ "content": "<pad>",
26
+ "lstrip": false,
27
+ "normalized": true,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ "tokenizer_class": "GPT2Tokenizer",
32
+ "unk_token": {
33
+ "__type": "AddedToken",
34
+ "content": "</s>",
35
+ "lstrip": false,
36
+ "normalized": true,
37
+ "rstrip": false,
38
+ "single_word": false
39
+ }
40
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.0,
3
+ "train_loss": 10.797767639160156,
4
+ "train_runtime": 59.971,
5
+ "train_samples": 2355,
6
+ "train_samples_per_second": 0.133,
7
+ "train_steps_per_second": 0.017
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.003389830508474576,
5
+ "global_step": 1,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.0,
12
+ "step": 1,
13
+ "total_flos": 14924118441984.0,
14
+ "train_loss": 10.797767639160156,
15
+ "train_runtime": 59.971,
16
+ "train_samples_per_second": 0.133,
17
+ "train_steps_per_second": 0.017
18
+ }
19
+ ],
20
+ "max_steps": 1,
21
+ "num_train_epochs": 1,
22
+ "total_flos": 14924118441984.0,
23
+ "trial_name": null,
24
+ "trial_params": null
25
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99311528991158f97401aa49595d2b05000bf88fbcfd9aacb24d5fefbca07483
3
+ size 3963
vocab.json ADDED
The diff for this file is too large to render. See raw diff