ldhldh commited on
Commit
5568b10
β€’
1 Parent(s): 1baf1cd

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -21
app.py CHANGED
@@ -25,22 +25,13 @@ torch_device = "cuda" if torch.cuda.is_available() else "cpu"
25
  print("Running on device:", torch_device)
26
  print("CPU threads:", torch.get_num_threads())
27
 
28
- loader = PyPDFLoader("total.pdf")
29
- pages = loader.load()
30
-
31
- # 데이터λ₯Ό λΆˆλŸ¬μ™€μ„œ ν…μŠ€νŠΈλ₯Ό μΌμ •ν•œ 수둜 λ‚˜λˆ„κ³  κ΅¬λΆ„μžλ‘œ μ—°κ²°ν•˜λŠ” μž‘μ—…
32
- text_splitter = RecursiveCharacterTextSplitter(chunk_size=1000, chunk_overlap=0)
33
- texts = text_splitter.split_documents(pages)
34
-
35
- print(f"λ¬Έμ„œμ— {len(texts)}개의 λ¬Έμ„œλ₯Ό 가지고 μžˆμŠ΅λ‹ˆλ‹€.")
36
-
37
  hf_hub_download(repo_id='StarFox7/Llama-2-ko-7B-chat-gguf', filename='Llama-2-ko-7B-chat-gguf-q4_0.bin', local_dir='./')
38
 
39
  llm = LlamaCpp(
40
  model_path='Llama-2-ko-7B-chat-gguf-q4_0.bin',
41
  temperature=0.5,
42
  top_p=0.9,
43
- max_tokens=128,
44
  verbose=True,
45
  n_ctx=2048,
46
  n_gpu_layers=-1,
@@ -50,16 +41,8 @@ llm = LlamaCpp(
50
  # μž„λ² λ”© λͺ¨λΈ λ‘œλ“œ
51
  embeddings = HuggingFaceEmbeddings(model_name="intfloat/multilingual-e5-large")
52
 
53
- # λ¬Έμ„œμ— μžˆλŠ” ν…μŠ€νŠΈλ₯Ό μž„λ² λ”©ν•˜κ³  FAISS 에 인덱슀λ₯Ό ꡬ좕함
54
- index = FAISS.from_documents(
55
- documents=texts,
56
- embedding=embeddings,
57
- )
58
-
59
- # faiss_db 둜 λ‘œμ»¬μ— μ €μž₯ν•˜κΈ°
60
- index.save_local("faiss_db")
61
  # faiss_db 둜 λ‘œμ»¬μ— λ‘œλ“œν•˜κΈ°
62
- docsearch = FAISS.load_local("faiss_db", embeddings)
63
 
64
  embeddings_filter = EmbeddingsFilter(
65
  embeddings=embeddings,
@@ -118,13 +101,11 @@ def gen(x, id, customer_data):
118
 
119
  ### λͺ…λ Ήμ–΄:
120
  λ‹€μŒ 지침을 μ°Έκ³ ν•˜μ—¬ μƒλ‹΄μ›μœΌλ‘œμ„œ κ³ κ°μ—κ²Œ ν•„μš”ν•œ 응닡을 μ œκ³΅ν•˜μ„Έμš”.
121
-
122
  [지침]
123
  1.고객의 κ°€μž… 정보λ₯Ό κΌ­ ν™•μΈν•˜μ—¬ 고객이 κ°€μž…ν•œ λ³΄ν—˜μ— λŒ€ν•œ λ‚΄μš©λ§Œ μ œκ³΅ν•˜μ„Έμš”.
124
  2.고객이 κ°€μž…ν•œ λ³΄ν—˜μ΄λΌλ©΄ 고객의 μ§ˆλ¬Έμ— λŒ€ν•΄ 적절히 λ‹΅λ³€ν•˜μ„Έμš”.
125
  3.고객이 κ°€μž…ν•˜μ§€ μ•Šμ€ λ³΄ν—˜μ˜ 보상에 κ΄€ν•œ μ§ˆλ¬Έμ€ κ΄€λ ¨ λ³΄ν—˜μ„ μ†Œκ°œν•˜λ©° 보상이 λΆˆκ°€λŠ₯ν•˜λ‹€λŠ” 점을 μ•ˆλ‚΄ν•˜μ„Έμš”.
126
  4.고객이 κ°€μž…ν•˜μ§€ μ•Šμ€ λ³΄ν—˜μ€ κ°€μž…μ΄ ν•„μš”ν•˜λ‹€κ³  λ³΄ν—˜λͺ…을 ν™•μ‹€ν•˜κ²Œ μ–ΈκΈ‰ν•˜μ„Έμš”.
127
-
128
  λ‹€μŒ μž…λ ₯에 μ£Όμ–΄μ§€λŠ” 고객의 λ³΄ν—˜ κ°€μž… 정보와 상담 기둝을 보고 κ³ κ°μ—κ²Œ λ„μ›€λ˜λŠ” 정보λ₯Ό μ œκ³΅ν•˜μ„Έμš”. μ°¨κ·Όμ°¨κ·Ό μƒκ°ν•˜μ—¬ λ‹΅λ³€ν•˜μ„Έμš”. 당신은 잘 ν•  수 μžˆμŠ΅λ‹ˆλ‹€.
129
 
130
  ### μž…λ ₯:
 
25
  print("Running on device:", torch_device)
26
  print("CPU threads:", torch.get_num_threads())
27
 
 
 
 
 
 
 
 
 
 
28
  hf_hub_download(repo_id='StarFox7/Llama-2-ko-7B-chat-gguf', filename='Llama-2-ko-7B-chat-gguf-q4_0.bin', local_dir='./')
29
 
30
  llm = LlamaCpp(
31
  model_path='Llama-2-ko-7B-chat-gguf-q4_0.bin',
32
  temperature=0.5,
33
  top_p=0.9,
34
+ max_tokens=80,
35
  verbose=True,
36
  n_ctx=2048,
37
  n_gpu_layers=-1,
 
41
  # μž„λ² λ”© λͺ¨λΈ λ‘œλ“œ
42
  embeddings = HuggingFaceEmbeddings(model_name="intfloat/multilingual-e5-large")
43
 
 
 
 
 
 
 
 
 
44
  # faiss_db 둜 λ‘œμ»¬μ— λ‘œλ“œν•˜κΈ°
45
+ docsearch = FAISS.load_local("", embeddings)
46
 
47
  embeddings_filter = EmbeddingsFilter(
48
  embeddings=embeddings,
 
101
 
102
  ### λͺ…λ Ήμ–΄:
103
  λ‹€μŒ 지침을 μ°Έκ³ ν•˜μ—¬ μƒλ‹΄μ›μœΌλ‘œμ„œ κ³ κ°μ—κ²Œ ν•„μš”ν•œ 응닡을 μ œκ³΅ν•˜μ„Έμš”.
 
104
  [지침]
105
  1.고객의 κ°€μž… 정보λ₯Ό κΌ­ ν™•μΈν•˜μ—¬ 고객이 κ°€μž…ν•œ λ³΄ν—˜μ— λŒ€ν•œ λ‚΄μš©λ§Œ μ œκ³΅ν•˜μ„Έμš”.
106
  2.고객이 κ°€μž…ν•œ λ³΄ν—˜μ΄λΌλ©΄ 고객의 μ§ˆλ¬Έμ— λŒ€ν•΄ 적절히 λ‹΅λ³€ν•˜μ„Έμš”.
107
  3.고객이 κ°€μž…ν•˜μ§€ μ•Šμ€ λ³΄ν—˜μ˜ 보상에 κ΄€ν•œ μ§ˆλ¬Έμ€ κ΄€λ ¨ λ³΄ν—˜μ„ μ†Œκ°œν•˜λ©° 보상이 λΆˆκ°€λŠ₯ν•˜λ‹€λŠ” 점을 μ•ˆλ‚΄ν•˜μ„Έμš”.
108
  4.고객이 κ°€μž…ν•˜μ§€ μ•Šμ€ λ³΄ν—˜μ€ κ°€μž…μ΄ ν•„μš”ν•˜λ‹€κ³  λ³΄ν—˜λͺ…을 ν™•μ‹€ν•˜κ²Œ μ–ΈκΈ‰ν•˜μ„Έμš”.
 
109
  λ‹€μŒ μž…λ ₯에 μ£Όμ–΄μ§€λŠ” 고객의 λ³΄ν—˜ κ°€μž… 정보와 상담 기둝을 보고 κ³ κ°μ—κ²Œ λ„μ›€λ˜λŠ” 정보λ₯Ό μ œκ³΅ν•˜μ„Έμš”. μ°¨κ·Όμ°¨κ·Ό μƒκ°ν•˜μ—¬ λ‹΅λ³€ν•˜μ„Έμš”. 당신은 잘 ν•  수 μžˆμŠ΅λ‹ˆλ‹€.
110
 
111
  ### μž…λ ₯: