|
from abc import abstractmethod |
|
from typing import Any, Dict, List, Literal, Optional |
|
|
|
from .api import infer |
|
from .artifact import fetch_artifact |
|
from .dataclass import Field |
|
from .formats import Format, SystemFormat |
|
from .inference import InferenceEngine, LogProbInferenceEngine, OpenAiInferenceEngine |
|
from .metrics import BulkInstanceMetric |
|
from .operator import SequentialOperator |
|
from .settings_utils import get_settings |
|
from .system_prompts import EmptySystemPrompt, SystemPrompt |
|
from .templates import Template |
|
|
|
settings = get_settings() |
|
|
|
|
|
def get_task_data_dict(task_data): |
|
import json |
|
|
|
|
|
|
|
return json.loads(task_data) if isinstance(task_data, str) else task_data |
|
|
|
|
|
class LLMAsJudgeBase(BulkInstanceMetric): |
|
"""LLM-as-judge-base metric class for evaluating correctness of generated predictions. |
|
|
|
Attributes: |
|
main_score (str): The main score label used for evaluation. |
|
task (str): The type of task the llm as judge runs. This defines the output and input |
|
format of the judge model. |
|
template (Template): The template used when generating inputs for the judge llm. |
|
format (Format): The format used when generating inputs for judge llm. |
|
system_prompt (SystemPrompt): The system prompt used when generating inputs for judge llm. |
|
inference_model (InferenceEngine): The module that creates the inference of the judge llm. |
|
reduction_map (dict): A dictionary specifying the reduction method for the metric. |
|
batch_size (int): The size of the bulk. |
|
""" |
|
|
|
main_score: str = "llm_as_judge" |
|
task: str |
|
template: Template |
|
system_prompt: SystemPrompt = Field(default_factory=EmptySystemPrompt) |
|
format: Format = Field(default_factory=SystemFormat) |
|
inference_model: InferenceEngine |
|
reduction_map: Optional[Dict[str, List[str]]] = None |
|
batch_size: int = 32 |
|
prediction_type = Any |
|
|
|
def verify(self): |
|
if not isinstance(self.template, Template): |
|
raise ValueError( |
|
f"Provided template argument to 'LLMAsJudge' metric is not of type Template, but {type(self.template)}" |
|
) |
|
if self.format and not isinstance(self.format, Format): |
|
raise ValueError( |
|
f"Provided format argument to 'LLMAsJudge' metric is not of type Format, but {type(self.format)}" |
|
) |
|
|
|
if self.system_prompt and not isinstance(self.system_prompt, SystemPrompt): |
|
raise ValueError( |
|
f"Provided system_prompt argument to 'LLMAsJudge' metric is not of type SystemPrompt, but {type(self.system_prompt)}" |
|
) |
|
|
|
if isinstance(self.inference_model, OpenAiInferenceEngine): |
|
if self.format and type(self.format) is not SystemFormat: |
|
raise ValueError( |
|
"Error in 'LLMAsJudge' metric. Inference model 'OpenAiInferenceEngine' does " |
|
"not support formatting. Please remove the format definition from the recipe" |
|
" (OpenAi Chat API take care of the formatting automatically)." |
|
) |
|
if self.system_prompt and type(self.system_prompt) is not EmptySystemPrompt: |
|
raise ValueError( |
|
"Error in 'LLMAsJudge' metric. Inference model 'OpenAiInferenceEngine' does " |
|
"not support system prompt. Please remove the system_prompt definition from the recipe" |
|
" (Current implementation of Unitxt does not support this." |
|
" Support will be added in future updates)." |
|
) |
|
|
|
@abstractmethod |
|
def get_full_task_name(self): |
|
pass |
|
|
|
def compute( |
|
self, |
|
references: List[List[Any]], |
|
predictions: List[Any], |
|
task_data: List[Dict], |
|
) -> List[Dict[str, Any]]: |
|
instances = self.prepare_instances(references, predictions, task_data) |
|
outputs = self.infer_instances(instances) |
|
return self.get_metric_results_from_prediction_outputs(outputs) |
|
|
|
@abstractmethod |
|
def prepare_instances( |
|
self, references, predictions, task_data |
|
) -> List[Dict[str, Any]]: |
|
"""Generate a list of instances for inference. |
|
|
|
Each generated instance should include all the fields required by the metrics' task and template, to |
|
create the source prompt for the judge. |
|
""" |
|
pass |
|
|
|
@abstractmethod |
|
def infer_instances(self, instances: List[Dict[str, Any]]) -> List[Dict[str, Any]]: |
|
"""Generate the dataset and call the inference engine to generate the judges' predictions. |
|
|
|
Return the list of the produced instances with their generated judge predictions. |
|
""" |
|
pass |
|
|
|
@abstractmethod |
|
def get_metric_results_from_prediction_outputs( |
|
self, outputs: List[Dict[str, Any]] |
|
) -> List[Dict[str, Any]]: |
|
"""Generate a scores' dictionary for each instance. |
|
|
|
Return the list of scores dictionaries for the input instances. |
|
""" |
|
pass |
|
|
|
|
|
class LLMAsJudge(LLMAsJudgeBase): |
|
"""LLM-as-judge-based metric class for evaluating correctness of generated predictions. |
|
|
|
This class uses the source prompt given to the generator and the generator's predictions to evaluate |
|
correctness using one of three supported tasks (rating.single_turn, rating.single_turn_with_reference, |
|
pairwise_comparative_rating.single_turn). |
|
|
|
Attributes: |
|
main_score (str): The main score label used for evaluation. |
|
task (Literal["rating.single_turn","rating.single_turn_with_reference", |
|
"pairwise_comparative_rating.single_turn"]): The type of task the llm as judge runs. |
|
This defines the output and input format of the judge model. |
|
template (Template): The template used when generating inputs for the judge llm. |
|
format (Format): The format used when generating inputs for judge llm. |
|
system_prompt (SystemPrompt): The system prompt used when generating inputs for judge llm. |
|
strip_system_prompt_and_format_from_inputs (bool): Whether to strip the system prompt and formatting from the |
|
inputs that the models that is being judges received, when they are inserted to the llm-as-judge prompt. |
|
inference_model (InferenceEngine): The module that creates the inference of the judge llm. |
|
reduction_map (dict): A dictionary specifying the reduction method for the metric. |
|
batch_size (int): The size of the bulk. |
|
""" |
|
|
|
task: Literal[ |
|
"rating.single_turn", |
|
"rating.single_turn_with_reference", |
|
"pairwise_comparative_rating.single_turn", |
|
] |
|
strip_system_prompt_and_format_from_inputs: bool = True |
|
|
|
def _get_input_instances(self, task_data: List[Dict]) -> List: |
|
if self.strip_system_prompt_and_format_from_inputs: |
|
instances = [] |
|
for task_data_instance in task_data: |
|
template = task_data_instance["metadata"]["template"] |
|
template, _ = fetch_artifact(template) |
|
instance = SequentialOperator( |
|
steps=[template, "formats.empty"] |
|
).process_instance( |
|
{ |
|
"input_fields": task_data_instance, |
|
"reference_fields": task_data_instance, |
|
} |
|
) |
|
instances.append(instance["source"]) |
|
""" |
|
We also have access to: instance["target"] |
|
instance["references"] |
|
""" |
|
return instances |
|
return [t["source"] for t in task_data] |
|
|
|
def _get_instance_for_judge_model( |
|
self, input_instances: List[str], predictions: List, references: List |
|
) -> List[Dict]: |
|
if self.task == "rating.single_turn": |
|
instances = [ |
|
{ |
|
"question": input_instance, |
|
"answer": prediction, |
|
} |
|
for input_instance, prediction, reference in zip( |
|
input_instances, predictions, references |
|
) |
|
] |
|
elif self.task == "rating.single_turn_with_reference": |
|
instances = [ |
|
{ |
|
"question": input_instance, |
|
"answer": prediction, |
|
"reference_answer": reference[0], |
|
} |
|
for input_instance, prediction, reference in zip( |
|
input_instances, predictions, references |
|
) |
|
] |
|
elif self.task == "pairwise_comparative_rating.single_turn": |
|
instances = [ |
|
{ |
|
"question": input_instance, |
|
"answer_a": prediction, |
|
"answer_b": reference[0], |
|
"model_a": "input_model", |
|
"model_b": "baseline_model", |
|
} |
|
for input_instance, prediction, reference in zip( |
|
input_instances, predictions, references |
|
) |
|
] |
|
else: |
|
raise NotImplementedError( |
|
f"Error in 'LLMAsJudge' metric. {self.task} is not a supported task type." |
|
) |
|
return instances |
|
|
|
def prepare(self): |
|
super().prepare() |
|
if self.task == "pairwise_comparative_rating.single_turn": |
|
self.reduction_map = {"weighted_win_rate": [self.main_score]} |
|
if self.reduction_map is None: |
|
self.reduction_map = {"mean": [self.main_score]} |
|
|
|
def verify(self): |
|
super().verify() |
|
supported_tasks = [ |
|
"rating.single_turn", |
|
"rating.single_turn_with_reference", |
|
"pairwise_comparative_rating.single_turn", |
|
] |
|
assert self.task in supported_tasks, ( |
|
f"Error in 'LLMAsJudge' metric. {self.task} is not a supported task type." |
|
f"The supported tasks types are: {', '.join(supported_tasks)}." |
|
) |
|
|
|
def get_full_task_name(self): |
|
return f"tasks.response_assessment.{self.task}" |
|
|
|
def infer_instances(self, instances): |
|
return infer( |
|
instances, |
|
engine=self.inference_model, |
|
task=self.get_full_task_name(), |
|
template=self.template, |
|
system_prompt=self.system_prompt, |
|
format=self.format, |
|
return_data=True, |
|
) |
|
|
|
def get_metric_results_from_prediction_outputs(self, outputs): |
|
results = [] |
|
for instance in outputs: |
|
if self.task == "pairwise_comparative_rating.single_turn": |
|
task_data = get_task_data_dict(instance["task_data"]) |
|
is_model_b_the_baseline = task_data["model_b"] == "baseline_model" |
|
if is_model_b_the_baseline: |
|
model_a_preference_score = instance["prediction"] |
|
else: |
|
model_a_preference_score = instance["prediction"] * -1 |
|
|
|
result = { |
|
self.main_score: model_a_preference_score, |
|
"judge_raw_output": instance["raw_prediction"], |
|
"judge_raw_input": instance["source"], |
|
} |
|
else: |
|
result = { |
|
self.main_score: instance["prediction"], |
|
"judge_raw_output": instance["raw_prediction"], |
|
"judge_raw_input": instance["source"], |
|
} |
|
results.append(result) |
|
return results |
|
|
|
def prepare_instances(self, references, predictions, task_data): |
|
input_instances = self._get_input_instances(task_data) |
|
return self._get_instance_for_judge_model( |
|
input_instances, predictions, references |
|
) |
|
|
|
|
|
class TaskBasedLLMasJudge(LLMAsJudgeBase): |
|
"""LLM-as-judge-based metric class for evaluating correctness of generated predictions. |
|
|
|
This class can use any task and matching template to evaluate the predictions. All |
|
task/templates field are taken from the instance's task_data. |
|
The instances sent to the judge can either be: 1.a unitxt dataset, in which case the predictions are |
|
copied to a specified field of the task. 2. dictionaries with the fields required by the task and template. |
|
|
|
Attributes: |
|
main_score (str): The main score label used for evaluation. |
|
task (str): The type of task the llm as judge runs. |
|
This defines the output and input format of the judge model. |
|
template (Template): The template used when generating inputs for the judge llm. |
|
format (Format): The format used when generating inputs for judge llm. |
|
system_prompt (SystemPrompt): The system prompt used when generating inputs for judge llm. |
|
strip_system_prompt_and_format_from_inputs (bool): Whether to strip the system prompt and formatting from the |
|
inputs that the models that is being judges received, when they are inserted to the llm-as-judge prompt. |
|
inference_model (InferenceEngine): The module that creates the inference of the judge llm. |
|
reduction_map (dict): A dictionary specifying the reduction method for the metric. |
|
batch_size (int): The size of the bulk. |
|
infer_log_probs(bool): whether to perform the inference using logprobs. If true, the template's |
|
post-processing must support the logprobs output. |
|
judge_to_generator_fields_mapping (Dict[str, str]): optional mapping between the names of the fields in the generator task and the |
|
judge task. For example, if the generator task uses "reference_answers" and the judge task expect "ground_truth", |
|
include {"ground_truth": "reference_answers"} in this dictionary. |
|
prediction_field: if indicated, and prediction exist, copy prediction to this field name in task_data. |
|
include_meta_data (bool): whether to include the inference per-instance metadata in the returned results. |
|
|
|
""" |
|
|
|
infer_log_probs: bool = False |
|
judge_to_generator_fields_mapping: Dict[str, str] = {} |
|
prediction_field: Optional[str] = None |
|
include_meta_data: bool = True |
|
|
|
|
|
|
|
|
|
def preprocess_instance(self, instance): |
|
if "task_data" not in instance: |
|
instance["task_data"] = instance.copy() |
|
if "prediction" not in instance: |
|
instance["prediction"] = None |
|
if "references" not in instance: |
|
instance["references"] = [""] |
|
return instance |
|
|
|
def verify(self): |
|
super().verify() |
|
if self.infer_log_probs and not isinstance( |
|
self.inference_model, LogProbInferenceEngine |
|
): |
|
raise NotImplementedError( |
|
f"Error in TaskBasedLLMasJudge: return_log_probs set to True but supplied engine " |
|
f"{self.inference_model.__class__.__name__} does not support logprobs." |
|
) |
|
if self.include_meta_data and not hasattr( |
|
self.inference_model, "get_return_object" |
|
): |
|
Warning( |
|
f"Supplied inference engine {self.inference_model.__class__.__name__} does not support " |
|
"return_meta_data. Setting return_meta_data to False. Metadata scores will not appear " |
|
"in returned instances scores." |
|
) |
|
self.include_meta_data = False |
|
|
|
def prepare(self): |
|
super().prepare() |
|
self.reduction_map = {"mean": [self.main_score]} |
|
self.score_prefix = f"{self.inference_model.get_engine_id()}_" |
|
|
|
def get_full_task_name(self): |
|
return self.task |
|
|
|
def get_metric_results_from_prediction_outputs(self, outputs): |
|
results = [] |
|
for instance in outputs: |
|
result = { |
|
self.main_score: instance["prediction"], |
|
f"{self.main_score}_judge_raw_output": instance["raw_prediction"], |
|
f"{self.main_score}_judge_raw_input": instance["source"], |
|
} |
|
if self.include_meta_data: |
|
meta_data = { |
|
f"{self.main_score}_{k}": v |
|
for k, v in instance["infer_meta_data"].items() |
|
} |
|
result.update(meta_data) |
|
results.append(result) |
|
return results |
|
|
|
def prepare_instances(self, references, predictions, task_data): |
|
from . import get_from_catalog |
|
|
|
instances = [] |
|
judge_task = get_from_catalog(self.get_full_task_name()) |
|
judge_task_input_fields = judge_task.input_fields |
|
|
|
for input_instance, prediction, _ in zip(task_data, predictions, references): |
|
input_instance = get_task_data_dict(input_instance) |
|
|
|
instance_task_data = {} |
|
for judge_task_input_field in judge_task_input_fields: |
|
orig_task_field_name = self.judge_to_generator_fields_mapping.get( |
|
judge_task_input_field, judge_task_input_field |
|
) |
|
new_val = input_instance.get(orig_task_field_name) |
|
if new_val: |
|
instance_task_data[judge_task_input_field] = new_val |
|
|
|
if self.prediction_field and prediction: |
|
instance_task_data[self.prediction_field] = str(prediction) |
|
instance_task_data = judge_task.process(instance_task_data)["input_fields"] |
|
instances.append(instance_task_data) |
|
|
|
return instances |
|
|
|
def infer_instances(self, instances): |
|
return infer( |
|
instances, |
|
engine=self.inference_model, |
|
task=self.get_full_task_name(), |
|
template=self.template, |
|
system_prompt=self.system_prompt, |
|
format=self.format, |
|
return_data=True, |
|
return_log_probs=self.infer_log_probs, |
|
return_meta_data=self.include_meta_data, |
|
) |
|
|