Unggi commited on
Commit
60452c9
โ€ข
1 Parent(s): 19b72b4

first commit

Browse files
Files changed (3) hide show
  1. app.py +63 -0
  2. gpt_model/config.json +51 -0
  3. gpt_model/pytorch_model.bin +3 -0
app.py ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ import numpy as np
4
+ #import itertools
5
+
6
+ import gradio as gr
7
+
8
+ import pandas as pd
9
+
10
+ # make function using import pip to install torch
11
+ import pip
12
+ #pip.main(['install', 'torch'])
13
+ #pip.main(['install', 'transformers'])
14
+
15
+ import torch
16
+ import transformers
17
+
18
+
19
+ # saved_model
20
+ def load_model():
21
+
22
+ pretrained_model_name = "skt/kogpt2-base-v2"
23
+
24
+ tokenizer = transformers.PreTrainedTokenizerFast.from_pretrained(
25
+ pretrained_model_name, # kogpt2
26
+ # kogpt๋Š” ์‚ฌ์ „์— ํ† ํฐ์„ ์ง€์ •ํ•ด์ฃผ์ง€ ์•Š์œผ๋ฉด, None ๊ฐ’์œผ๋กœ ๋ฐ˜์˜๋˜์–ด์žˆ์Œ
27
+ # ๋ฐ˜๋“œ์‹œ ์ง€์ •ํ•ด์ฃผ์–ด์•ผ ํ•จ
28
+ bos_token='</s>', eos_token='</s>', unk_token='<unk>',
29
+ pad_token='<pad>', mask_token='<mask>'
30
+ )
31
+ model = transformers.GPT2LMHeadModel.from_pretrained(
32
+ pretrained_model_name # kogpt2
33
+ )
34
+ model.resize_token_embeddings( len(tokenizer) )
35
+
36
+ return model, tokenizer
37
+
38
+
39
+ # main
40
+ def inference(prompt):
41
+
42
+ model, tokenizer = load_model()
43
+
44
+ input_ids = tokenizer.encode(prompt, return_tensors="pt")
45
+ gen_ids = model.generate(input_ids,
46
+ max_length=128,
47
+ repetition_penalty=2.0,
48
+ pad_token_id=tokenizer.pad_token_id,
49
+ eos_token_id=tokenizer.eos_token_id,
50
+ bos_token_id=tokenizer.bos_token_id,
51
+ use_cache=True)
52
+ outputs = tokenizer.decode(gen_ids[0])
53
+
54
+ return outputs
55
+
56
+
57
+ demo = gr.Interface(
58
+ fn=inference,
59
+ inputs="text",
60
+ outputs="text" #return ๊ฐ’
61
+ ).launch() # launch(share=True)๋ฅผ ์„ค์ •ํ•˜๋ฉด ์™ธ๋ถ€์—์„œ ์ ‘์† ๊ฐ€๋Šฅํ•œ ๋งํฌ๊ฐ€ ์ƒ์„ฑ๋จ
62
+
63
+ demo.launch()
gpt_model/config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "skt/kogpt2-base-v2",
3
+ "_num_labels": 1,
4
+ "activation_function": "gelu_new",
5
+ "architectures": [
6
+ "GPT2LMHeadModel"
7
+ ],
8
+ "attn_pdrop": 0.1,
9
+ "author": "Heewon Jeon(madjakarta@gmail.com)",
10
+ "bos_token_id": 0,
11
+ "created_date": "2021-04-28",
12
+ "embd_pdrop": 0.1,
13
+ "eos_token_id": 1,
14
+ "gradient_checkpointing": false,
15
+ "id2label": {
16
+ "0": "LABEL_0"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "label2id": {
20
+ "LABEL_0": 0
21
+ },
22
+ "layer_norm_epsilon": 1e-05,
23
+ "license": "CC-BY-NC-SA 4.0",
24
+ "model_type": "gpt2",
25
+ "n_ctx": 1024,
26
+ "n_embd": 768,
27
+ "n_head": 12,
28
+ "n_inner": null,
29
+ "n_layer": 12,
30
+ "n_positions": 1024,
31
+ "pad_token_id": 3,
32
+ "reorder_and_upcast_attn": false,
33
+ "resid_pdrop": 0.1,
34
+ "scale_attn_by_inverse_layer_idx": false,
35
+ "scale_attn_weights": true,
36
+ "summary_activation": null,
37
+ "summary_first_dropout": 0.1,
38
+ "summary_proj_to_labels": true,
39
+ "summary_type": "cls_index",
40
+ "summary_use_proj": true,
41
+ "task_specific_params": {
42
+ "text-generation": {
43
+ "do_sample": true,
44
+ "max_length": 50
45
+ }
46
+ },
47
+ "torch_dtype": "float32",
48
+ "transformers_version": "4.12.5",
49
+ "use_cache": true,
50
+ "vocab_size": 51200
51
+ }
gpt_model/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1347fc29e6015fe4da9fbcf6770c1a962f201eeeea15391bf943fe62f46ee4ae
3
+ size 513298281