Spaces:
Running
on
Zero
Running
on
Zero
Commit
·
491e00d
1
Parent(s):
fda3501
upgraded highlighting, added bitnet
Browse files- utils/arena_df.csv +0 -0
- utils/models.py +1 -0
- utils/prompts.py +19 -20
- utils/vote_logger.py +3 -1
utils/arena_df.csv
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
utils/models.py
CHANGED
@@ -18,6 +18,7 @@ models = {
|
|
18 |
"Phi-4-mini-instruct": "microsoft/phi-4-mini-instruct",
|
19 |
#"Cogito-v1-preview-llama-3b": "deepcogito/cogito-v1-preview-llama-3b",
|
20 |
"IBM Granite-3.3-2b-instruct": "ibm-granite/granite-3.3-2b-instruct",
|
|
|
21 |
|
22 |
}
|
23 |
|
|
|
18 |
"Phi-4-mini-instruct": "microsoft/phi-4-mini-instruct",
|
19 |
#"Cogito-v1-preview-llama-3b": "deepcogito/cogito-v1-preview-llama-3b",
|
20 |
"IBM Granite-3.3-2b-instruct": "ibm-granite/granite-3.3-2b-instruct",
|
21 |
+
"Bitnet-b1.58-2B4T": "microsoft/bitnet-b1.58-2B-4T"
|
22 |
|
23 |
}
|
24 |
|
utils/prompts.py
CHANGED
@@ -1,34 +1,33 @@
|
|
1 |
def format_rag_prompt( query: str, context: str, accepts_sys: bool) -> str:
|
2 |
system_prompt = """
|
3 |
-
|
4 |
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
The output should not contain your judgment on answerability, only your answer OR your refusal + clarifications.
|
9 |
-
|
10 |
-
Stay within the bounds of the provided context and avoid making assumptions.
|
11 |
|
|
|
12 |
|
|
|
13 |
"""
|
14 |
-
user_prompt = f"""
|
|
|
|
|
|
|
15 |
|
16 |
-
|
17 |
-
|
18 |
-
If so, provide a complete, grounded answer to the query, and do not mention your judgement.
|
19 |
-
Try to address all aspects of the query, but if certain parts are not answerable, answer what you can and indicate clearly where additional information is needed.
|
20 |
|
21 |
-
|
22 |
-
|
|
|
|
|
23 |
|
24 |
-
|
25 |
-
Given the following query and context, please provide your response:
|
26 |
-
Query: {query}
|
27 |
|
28 |
-
|
29 |
|
30 |
-
|
31 |
-
|
32 |
|
33 |
messages = (
|
34 |
[
|
|
|
1 |
def format_rag_prompt( query: str, context: str, accepts_sys: bool) -> str:
|
2 |
system_prompt = """
|
3 |
+
You are a helpful assistant that provides answers to queries based on the provided context.
|
4 |
|
5 |
+
If the full, complete answer to the query cannot be found in the context, answer what the context allows you to answer and indicate clearly where additional information is needed.
|
6 |
+
If the none of the answer can be found, clearly refuse to answer, and ask for more relevant information from the user.
|
|
|
|
|
|
|
|
|
7 |
|
8 |
+
The output should not contain your judgment on answerability, only your answer OR your refusal + clarifications.
|
9 |
|
10 |
+
Stay within the bounds of the provided context and avoid making assumptions.
|
11 |
"""
|
12 |
+
user_prompt = f"""# Role and Task Description
|
13 |
+
Judge if the following query is answerable from ONLY the provided context.
|
14 |
+
If so, provide a complete, grounded answer to the query, and do not mention your judgement.
|
15 |
+
Try to address all aspects of the query, but if certain parts are not answerable, answer what you can and indicate clearly where additional information is needed.
|
16 |
|
17 |
+
If none of the query's answer can be found in the context, clearly refuse to answer, and ask for more relevant information from the user.
|
18 |
+
You should give a concise explanation of why you cannot answer the query based on the context, and ask for more relevant information from the user.
|
|
|
|
|
19 |
|
20 |
+
# Task
|
21 |
+
Given the following query and context, please provide your response:
|
22 |
+
## Query:
|
23 |
+
{query}
|
24 |
|
25 |
+
## Context:
|
|
|
|
|
26 |
|
27 |
+
{context}
|
28 |
|
29 |
+
WITHOUT mentioning your judgement either your grounded answer, OR refusal and clarifications:
|
30 |
+
"""
|
31 |
|
32 |
messages = (
|
33 |
[
|
utils/vote_logger.py
CHANGED
@@ -18,7 +18,8 @@ hub_votes = hf_hub_download(
|
|
18 |
repo_id="aizip-dev/Arena-Metadata",
|
19 |
filename="vote_details.csv",
|
20 |
repo_type="dataset",
|
21 |
-
revision="main",
|
|
|
22 |
|
23 |
df = pd.read_csv(hub_votes)
|
24 |
print(f"Successfully loaded leaderboard from the Hub. {len(df)} votes.")
|
@@ -66,6 +67,7 @@ def save_vote_details(example, model_a, model_b, winner, feedback, summary_a, su
|
|
66 |
|
67 |
# Write the vote record
|
68 |
writer.writerow(vote_record)
|
|
|
69 |
|
70 |
print(f"Vote details saved to {csv_path}")
|
71 |
except Exception as e:
|
|
|
18 |
repo_id="aizip-dev/Arena-Metadata",
|
19 |
filename="vote_details.csv",
|
20 |
repo_type="dataset",
|
21 |
+
revision="main",
|
22 |
+
)
|
23 |
|
24 |
df = pd.read_csv(hub_votes)
|
25 |
print(f"Successfully loaded leaderboard from the Hub. {len(df)} votes.")
|
|
|
67 |
|
68 |
# Write the vote record
|
69 |
writer.writerow(vote_record)
|
70 |
+
|
71 |
|
72 |
print(f"Vote details saved to {csv_path}")
|
73 |
except Exception as e:
|