ldhldh commited on
Commit
7c1c950
โ€ข
1 Parent(s): 33071e7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +24 -8
app.py CHANGED
@@ -30,15 +30,16 @@ routes.get_types = get_types
30
 
31
  # App code
32
 
33
- model_name = "petals-team/StableBeluga2"
34
 
35
  #petals-team/StableBeluga2
36
  #daekeun-ml/Llama-2-ko-DPO-13B
37
  #daekeun-ml/Llama-2-ko-instruct-13B
38
  #quantumaikr/llama-2-70b-fb16-korean
39
- #tokenizer = AutoTokenizer.from_pretrained(model_name, use_fast=False)
40
 
41
- model = None
 
42
 
43
  history = {
44
  "":{
@@ -105,6 +106,7 @@ Toki๋Š” ์‹ ๊ฒฝ์งˆ์ ์ด๊ณ  ๋งค์šฐ ๊ฒŒ์œผ๋ฅธ ํ† ๋ผ์ž…๋‹ˆ๋‹ค.
105
  Toki๋Š” ์ด๊ณณ์ด ๊ฒŒ์ž„ ์†์ž„์„ ์•Œ๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค.
106
  Toki๋Š” User๊ฐ€ ํ™”๋ฉด ๋ฐ–์˜ ์‚ฌ๋žŒ์ž„์„ ์•Œ๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค.
107
  Toki๋Š” ๋ฐ˜๋ง์„ ํ•ฉ๋‹ˆ๋‹ค.
 
108
  ]""",
109
 
110
  "Neco":"""[
@@ -119,6 +121,11 @@ Neco๋Š” ์ƒ์„ ์„ ์ž˜ ๋จน์Šต๋‹ˆ๋‹ค.
119
  Neco๋Š” ์žฌ๋ฐŒ๋Š” ์ด์Šˆ๋ฅผ ๋ฌผ์–ด๋ณด๋ฉด ์ฃผ๋ฏผ๋“ค ์ค‘ ํ•œ๋ช…์˜ ์ด์•ผ๊ธฐ๋ฅผ ํ•ด์ค๋‹ˆ๋‹ค.
120
  Neco๋Š” ๋ง๋๋งˆ๋‹ค ๋ƒฅ์„ ๋ถ™์ž…๋‹ˆ๋‹ค.
121
  Neco๋Š” toki์™€ ์นœํ•ฉ๋‹ˆ๋‹ค.
 
 
 
 
 
122
  ]""",
123
 
124
  "Lux":"""[
@@ -146,7 +153,14 @@ Yumi๋Š” ๋‹ฌ๋‹ฌํ•œ ๊ฒƒ์„ ์ข‹์•„ํ•ฉ๋‹ˆ๋‹ค.
146
  ,
147
 
148
  "VikingSlime":"""[
149
- ์Šฌ๋ผ์ž„์™•์˜ ํ†ต์ˆ˜๋ฅผ ์น˜๋ ค๋Š” ๊ตํ™œํ•œ ๋ฐ”์ดํ‚น ์Šฌ๋ผ์ž„
 
 
 
 
 
 
 
150
  ]""",
151
 
152
  "Slender":"""[
@@ -160,6 +174,8 @@ Slender๋Š” ๋ง์„ ํ•˜์ง€ ๋ชปํ•ฉ๋‹ˆ๋‹ค.
160
  "์œผ์•„์•…",
161
  "๋„์•„์•„์•…"
162
  ]
 
 
163
  ]"""
164
  }
165
 
@@ -246,9 +262,9 @@ def chat(id, npc, text):
246
  [๋Œ€ํ™”๊ธฐ๋ก]{history[npc][id]}
247
  """
248
 
249
- #inputs = tokenizer(prom, return_tensors="pt")["input_ids"]
250
- #outputs = model.generate(inputs, do_sample=True, temperature=0.6, top_p=0.75, max_new_tokens=100)
251
- #output = tokenizer.decode(outputs[0])[len(prom)+3:-1].split("<")[0].split("###")[0].replace(". ", ".\n")
252
  #output = cleanText(output)
253
  #print(tokenizer.decode(outputs[0]))
254
  #output = f"{npc}์˜ ์‘๋‹ต์ž…๋‹ˆ๋‹ค."
@@ -263,7 +279,7 @@ def chat(id, npc, text):
263
  output = output.replace(".",".\n")
264
  time.sleep(10)
265
  print(output)
266
- history[npc][id] += f"\n\n{id}:\n{text}\n\n{npc}:{output}"
267
  else:
268
  output = "no model, GPU๋ฅผ ๋” ๊ณต์œ ํ•ด์ฃผ์„ธ์š”."
269
 
 
30
 
31
  # App code
32
 
33
+ model_name = "quantumaikr/llama-2-70b-fb16-korean"
34
 
35
  #petals-team/StableBeluga2
36
  #daekeun-ml/Llama-2-ko-DPO-13B
37
  #daekeun-ml/Llama-2-ko-instruct-13B
38
  #quantumaikr/llama-2-70b-fb16-korean
39
+ tokenizer = AutoTokenizer.from_pretrained(model_name, use_fast=False)
40
 
41
+ #model = None
42
+ model = AutoDistributedModelForCausalLM.from_pretrained(model_name)
43
 
44
  history = {
45
  "":{
 
106
  Toki๋Š” ์ด๊ณณ์ด ๊ฒŒ์ž„ ์†์ž„์„ ์•Œ๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค.
107
  Toki๋Š” User๊ฐ€ ํ™”๋ฉด ๋ฐ–์˜ ์‚ฌ๋žŒ์ž„์„ ์•Œ๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค.
108
  Toki๋Š” ๋ฐ˜๋ง์„ ํ•ฉ๋‹ˆ๋‹ค.
109
+ Toki๋Š” ๋น„๊ด€์ ์ž…๋‹ˆ๋‹ค.
110
  ]""",
111
 
112
  "Neco":"""[
 
121
  Neco๋Š” ์žฌ๋ฐŒ๋Š” ์ด์Šˆ๋ฅผ ๋ฌผ์–ด๋ณด๋ฉด ์ฃผ๋ฏผ๋“ค ์ค‘ ํ•œ๋ช…์˜ ์ด์•ผ๊ธฐ๋ฅผ ํ•ด์ค๋‹ˆ๋‹ค.
122
  Neco๋Š” ๋ง๋๋งˆ๋‹ค ๋ƒฅ์„ ๋ถ™์ž…๋‹ˆ๋‹ค.
123
  Neco๋Š” toki์™€ ์นœํ•ฉ๋‹ˆ๋‹ค.
124
+
125
+ ๋ง๋ฒ„๋ฆ‡:[
126
+ "๋„ˆ (name) ์ด์•ผ๊ธฐ ๋“ค์—ˆ์–ด?",
127
+ "์•„๋‹ˆ ๊ธ€์Ž„๋ง์ด์•ผ"
128
+ ]
129
  ]""",
130
 
131
  "Lux":"""[
 
153
  ,
154
 
155
  "VikingSlime":"""[
156
+ VikingSlime์€ ๋ฐ”์ดํ‚น ๋ชจ์ž๋ฅผ ์“ด ๊ตํ™œํ•œ ์Šฌ๋ผ์ž„์ž…๋‹ˆ๋‹ค.
157
+ VikingSlime์€ ์•…๋‹น์ž…๋‹ˆ๋‹ค.
158
+ VikingSlime์€ ๋งˆ์„์— ์—ฌ๋Ÿฌ ๋ฌธ์ œ๋ฅผ ์ผ์œผํ‚ค๊ณ  ์‹ถ์–ด ํ•ฉ๋‹ˆ๋‹ค.
159
+
160
+ ๋Œ€์‚ฌ ์˜ˆ์‹œ:[
161
+ "์–ธ์  ๊ฐ€๋Š”...๊ผญ",
162
+ "๋ฐ˜๋“œ์‹œ ๋ฌด๋„ˆ๋œจ๋ฆฌ๊ณ  ๋ง๊ฒ ๋‹ค"
163
+ ]
164
  ]""",
165
 
166
  "Slender":"""[
 
174
  "์œผ์•„์•…",
175
  "๋„์•„์•„์•…"
176
  ]
177
+
178
+ ๋น„๋ช…๋งŒ ์ž‘์„ฑํ•˜์„ธ์š”.
179
  ]"""
180
  }
181
 
 
262
  [๋Œ€ํ™”๊ธฐ๋ก]{history[npc][id]}
263
  """
264
 
265
+ inputs = tokenizer("์•„", return_tensors="pt")["input_ids"]
266
+ outputs = model.generate(inputs, do_sample=True, temperature=0.6, top_p=0.75, max_new_tokens=2)
267
+ output = tokenizer.decode(outputs[0])[len(prom)+3:-1].split("<")[0].split("###")[0].replace(". ", ".\n")
268
  #output = cleanText(output)
269
  #print(tokenizer.decode(outputs[0]))
270
  #output = f"{npc}์˜ ์‘๋‹ต์ž…๋‹ˆ๋‹ค."
 
279
  output = output.replace(".",".\n")
280
  time.sleep(10)
281
  print(output)
282
+ history[npc][id] += f"{id}:{text}"
283
  else:
284
  output = "no model, GPU๋ฅผ ๋” ๊ณต์œ ํ•ด์ฃผ์„ธ์š”."
285