ayjays132 commited on
Commit
7ca5101
1 Parent(s): 716657e

Delete tokenizer

Browse files
tokenizer/added_tokens.json DELETED
@@ -1,34 +0,0 @@
1
- {
2
- "<active_listening>": 50286,
3
- "<apology>": 50268,
4
- "<bos>": 50258,
5
- "<cls>": 50263,
6
- "<empathy>": 50285,
7
- "<eos>": 50259,
8
- "<farewell>": 50266,
9
- "<greeting>": 50265,
10
- "<mask>": 50264,
11
- "<open_ended_question>": 50287,
12
- "<pad>": 50260,
13
- "<reflective_summary>": 50288,
14
- "<sep>": 50262,
15
- "<thank>": 50267,
16
- "<unk>": 50261,
17
- "<|pad|>": 50257,
18
- "[PAD]": 50269,
19
- "active_listening": 50282,
20
- "amazeballs": 50271,
21
- "bromance": 50279,
22
- "crowdfunding": 50278,
23
- "cryptocurrency": 50272,
24
- "empathy": 50281,
25
- "facepalm": 50277,
26
- "hangry": 50280,
27
- "intelligent": 50270,
28
- "open_ended_question": 50283,
29
- "photobomb": 50276,
30
- "reflective_summary": 50284,
31
- "upcycle": 50275,
32
- "vlog": 50274,
33
- "webinar": 50273
34
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/config.json DELETED
@@ -1,85 +0,0 @@
1
- {
2
- "_name_or_path": "ayjays132/CustomGPT2Conversational",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "context_window": 20,
10
- "contextual_embedding_dim": 1024,
11
- "device": "cuda",
12
- "dropout_rate": 0.1,
13
- "embd_pdrop": 0.1,
14
- "embedding_dim": 1024,
15
- "eos_token_id": 50256,
16
- "hidden_dim": 1024,
17
- "initializer_range": 0.02,
18
- "innovative_growth_capacity": 50000,
19
- "integration_settings": {
20
- "config_name": "config.json",
21
- "load_from_transformers": true,
22
- "pytorch_dump_folder_path": "./model_save",
23
- "pytorch_model_bin_name": "pytorch_model.bin"
24
- },
25
- "layer_norm_epsilon": 1e-05,
26
- "max_memory_size": 100000,
27
- "max_neurons": 100,
28
- "meta_learning_rate": 0.001,
29
- "model_type": "gpt2",
30
- "n_ctx": 1024,
31
- "n_embd": 1024,
32
- "n_head": 16,
33
- "n_inner": null,
34
- "n_layer": 24,
35
- "n_positions": 1024,
36
- "num_embeddings": 50268,
37
- "num_heads": 64,
38
- "num_layers": 24,
39
- "output_attentions": true,
40
- "output_hidden_states": true,
41
- "pad_token_id": 0,
42
- "reorder_and_upcast_attn": false,
43
- "resid_pdrop": 0.1,
44
- "scale_attn_by_inverse_layer_idx": false,
45
- "scale_attn_weights": true,
46
- "sep_token_id": -1,
47
- "special_tokens": {
48
- "additional_special_tokens": [
49
- "<greeting>",
50
- "<farewell>",
51
- "<thank>",
52
- "<apology>"
53
- ],
54
- "bos_token": "<bos>",
55
- "cls_token": "<cls>",
56
- "eos_token": "<eos>",
57
- "mask_token": "<mask>",
58
- "pad_token": "<pad>",
59
- "sep_token": "<sep>",
60
- "unk_token": "<unk>"
61
- },
62
- "state_shape": null,
63
- "summary_activation": null,
64
- "summary_first_dropout": 0.1,
65
- "summary_proj_to_labels": true,
66
- "summary_type": "cls_index",
67
- "summary_use_proj": true,
68
- "target_q_model": null,
69
- "task_specific_params": {
70
- "text-generation": {
71
- "do_sample": true,
72
- "length_penalty": 1.0,
73
- "max_length": 50,
74
- "no_repeat_ngram_size": 2,
75
- "repetition_penalty": 1.2,
76
- "temperature": 0.9,
77
- "top_k": 50,
78
- "top_p": 0.95
79
- }
80
- },
81
- "torch_dtype": "float32",
82
- "transformers_version": "4.28.0.dev0",
83
- "use_cache": true,
84
- "vocab_size": 50257
85
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/generation_config.json DELETED
@@ -1,9 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "bos_token_id": 50256,
4
- "eos_token_id": 50256,
5
- "output_attentions": true,
6
- "output_hidden_states": true,
7
- "pad_token_id": 0,
8
- "transformers_version": "4.28.0.dev0"
9
- }
 
 
 
 
 
 
 
 
 
 
tokenizer/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer/model_index.json DELETED
@@ -1,38 +0,0 @@
1
- {
2
- "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.27.0",
4
- "_name_or_path": "CompVis/stable-diffusion-v1-4",
5
- "feature_extractor": [
6
- "transformers",
7
- "CLIPImageProcessor"
8
- ],
9
- "image_encoder": [
10
- null,
11
- null
12
- ],
13
- "requires_safety_checker": true,
14
- "safety_checker": [
15
- "stable_diffusion",
16
- "StableDiffusionSafetyChecker"
17
- ],
18
- "scheduler": [
19
- "diffusers",
20
- "PNDMScheduler"
21
- ],
22
- "text_encoder": [
23
- "transformers",
24
- "CLIPTextModel"
25
- ],
26
- "tokenizer": [
27
- "transformers",
28
- "CLIPTokenizer"
29
- ],
30
- "unet": [
31
- "diffusers",
32
- "UNet2DConditionModel"
33
- ],
34
- "vae": [
35
- "diffusers",
36
- "AutoencoderKL"
37
- ]
38
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0281d5c40c659ba4197119cab76775cf384955e2d8864889300b0a677dfb97e
3
- size 1444569818
 
 
 
 
tokenizer/special_tokens_map.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|startoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "<|endoftext|>",
17
- "unk_token": {
18
- "content": "<|endoftext|>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer/tokenizer_config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "bos_token": {
4
- "__type": "AddedToken",
5
- "content": "<|startoftext|>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false
10
- },
11
- "do_lower_case": true,
12
- "eos_token": {
13
- "__type": "AddedToken",
14
- "content": "<|endoftext|>",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
- "errors": "replace",
21
- "model_max_length": 77,
22
- "pad_token": "<|endoftext|>",
23
- "special_tokens_map_file": "./special_tokens_map.json",
24
- "tokenizer_class": "CLIPTokenizer",
25
- "unk_token": {
26
- "__type": "AddedToken",
27
- "content": "<|endoftext|>",
28
- "lstrip": false,
29
- "normalized": true,
30
- "rstrip": false,
31
- "single_word": false
32
- }
33
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/vocab.json DELETED
The diff for this file is too large to render. See raw diff