HFHAB commited on
Commit
c760d77
1 Parent(s): cb4fa58

Update main.py

Browse files
Files changed (1) hide show
  1. main.py +5 -11
main.py CHANGED
@@ -10,8 +10,8 @@ client = InferenceClient("FacebookAI/roberta-large-mnli")
10
 
11
  class Item(BaseModel):
12
  prompt: str
13
- history: list
14
- system_prompt: str
15
  #temperature: float = 0.0
16
  #max_new_tokens: int = 1048
17
  #top_p: float = 0.15
@@ -26,13 +26,6 @@ class Item(BaseModel):
26
  # prompt += f"[INST] {message} [/INST]"
27
  # return prompt
28
 
29
- def format_prompt(message, history):
30
- prompt = "<s>"
31
- for user_prompt, bot_response in history:
32
- prompt += f"[INST] {user_prompt} [/INST]"
33
- prompt += f" {bot_response}</s> "
34
- prompt += f"[INST] {message} [/INST]"
35
- return prompt
36
 
37
  def generate(item: Item):
38
  #temperature = float(item.temperature)
@@ -50,8 +43,9 @@ def generate(item: Item):
50
  # )
51
 
52
  #formatted_prompt = format_prompt(f"{item.system_prompt}, {item.prompt}", item.history)
53
- text = item.prompt
54
- print(item.prompt)
 
55
  labels = ["Requirement", "Information"]
56
  print(labels)
57
  stream = client.zero_shot_classification(text, labels)
 
10
 
11
  class Item(BaseModel):
12
  prompt: str
13
+ #history: list
14
+ #system_prompt: str
15
  #temperature: float = 0.0
16
  #max_new_tokens: int = 1048
17
  #top_p: float = 0.15
 
26
  # prompt += f"[INST] {message} [/INST]"
27
  # return prompt
28
 
 
 
 
 
 
 
 
29
 
30
  def generate(item: Item):
31
  #temperature = float(item.temperature)
 
43
  # )
44
 
45
  #formatted_prompt = format_prompt(f"{item.system_prompt}, {item.prompt}", item.history)
46
+ #text = format_prompt(f"{item.system_prompt}, {item.prompt}", item.history)
47
+ text = f"{item.prompt}
48
+ print(text)
49
  labels = ["Requirement", "Information"]
50
  print(labels)
51
  stream = client.zero_shot_classification(text, labels)