Upload prompt template judgebench-vanilla-prompt.yaml
Browse files
judgebench-vanilla-prompt.yaml
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
prompt:
|
2 |
-
template:
|
3 |
- role: user
|
4 |
content: |-
|
5 |
You are a helpful assistant in evaluating the quality of the outputs for a given instruction.
|
@@ -15,15 +15,16 @@ prompt:
|
|
15 |
# Output (b):
|
16 |
{{response_b}}
|
17 |
# Which is better, Output (a) or Output (b)? Your response should be either “Output (a)” or “Output (b)”:
|
18 |
-
|
19 |
template_variables:
|
20 |
- question
|
21 |
- response_a
|
22 |
- response_b
|
23 |
|
24 |
-
metadata:
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
client_parameters:
|
29 |
-
|
|
|
|
1 |
prompt:
|
2 |
+
template:
|
3 |
- role: user
|
4 |
content: |-
|
5 |
You are a helpful assistant in evaluating the quality of the outputs for a given instruction.
|
|
|
15 |
# Output (b):
|
16 |
{{response_b}}
|
17 |
# Which is better, Output (a) or Output (b)? Your response should be either “Output (a)” or “Output (b)”:
|
18 |
+
|
19 |
template_variables:
|
20 |
- question
|
21 |
- response_a
|
22 |
- response_b
|
23 |
|
24 |
+
metadata:
|
25 |
+
source: Appendix A.5, https://arxiv.org/pdf/2410.12784
|
26 |
+
description: This prompt is used in the JudgeBench paper to evaluate the quality of the outputs for a given instruction.
|
27 |
+
|
28 |
client_parameters:
|
29 |
+
temperature: 0
|
30 |
+
custom_data: {}
|