| |
| |
| |
|
|
| from copy import deepcopy |
| from typing import Any, Dict, List, Optional |
| from warnings import warn |
|
|
| from pandas import DataFrame |
| from pandas import concat as pd_concat |
|
|
| from .base import BaseEvaluationRunResult |
|
|
|
|
| class EvaluationRunResult(BaseEvaluationRunResult): |
| """ |
| Contains the inputs and the outputs of an evaluation pipeline and provides methods to inspect them. |
| """ |
|
|
| def __init__(self, run_name: str, inputs: Dict[str, List[Any]], results: Dict[str, Dict[str, Any]]): |
| """ |
| Initialize a new evaluation run result. |
| |
| :param run_name: |
| Name of the evaluation run. |
| :param inputs: |
| Dictionary containing the inputs used for the run. |
| Each key is the name of the input and its value is |
| a list of input values. The length of the lists should |
| be the same. |
| :param results: |
| Dictionary containing the results of the evaluators |
| used in the evaluation pipeline. Each key is the name |
| of the metric and its value is dictionary with the following |
| keys: |
| - 'score': The aggregated score for the metric. |
| - 'individual_scores': A list of scores for each input sample. |
| """ |
| self.run_name = run_name |
| self.inputs = deepcopy(inputs) |
| self.results = deepcopy(results) |
|
|
| if len(inputs) == 0: |
| raise ValueError("No inputs provided.") |
| if len({len(l) for l in inputs.values()}) != 1: |
| raise ValueError("Lengths of the inputs should be the same.") |
|
|
| expected_len = len(next(iter(inputs.values()))) |
|
|
| for metric, outputs in results.items(): |
| if "score" not in outputs: |
| raise ValueError(f"Aggregate score missing for {metric}.") |
| if "individual_scores" not in outputs: |
| raise ValueError(f"Individual scores missing for {metric}.") |
|
|
| if len(outputs["individual_scores"]) != expected_len: |
| raise ValueError( |
| f"Length of individual scores for '{metric}' should be the same as the inputs. " |
| f"Got {len(outputs['individual_scores'])} but expected {expected_len}." |
| ) |
|
|
| def score_report(self) -> DataFrame: |
| """ |
| Transforms the results into a Pandas DataFrame with the aggregated scores for each metric. |
| |
| :returns: |
| Pandas DataFrame with the aggregated scores. |
| """ |
| results = {k: v["score"] for k, v in self.results.items()} |
| df = DataFrame.from_dict(results, orient="index", columns=["score"]).reset_index() |
| df.columns = ["metrics", "score"] |
| return df |
|
|
| def to_pandas(self) -> DataFrame: |
| """ |
| Creates a Pandas DataFrame containing the scores of each metric for every input sample. |
| |
| :returns: |
| Pandas DataFrame with the scores. |
| """ |
| inputs_columns = list(self.inputs.keys()) |
| inputs_values = list(self.inputs.values()) |
| inputs_values = list(map(list, zip(*inputs_values))) |
| df_inputs = DataFrame(inputs_values, columns=inputs_columns) |
|
|
| scores_columns = list(self.results.keys()) |
| scores_values = [v["individual_scores"] for v in self.results.values()] |
| scores_values = list(map(list, zip(*scores_values))) |
| df_scores = DataFrame(scores_values, columns=scores_columns) |
|
|
| return df_inputs.join(df_scores) |
|
|
| def comparative_individual_scores_report( |
| self, other: "BaseEvaluationRunResult", keep_columns: Optional[List[str]] = None |
| ) -> DataFrame: |
| """ |
| Creates a Pandas DataFrame with the scores for each metric in the results of two different evaluation runs. |
| |
| The inputs to both evaluation runs is assumed to be the same. |
| |
| :param other: |
| Results of another evaluation run to compare with. |
| :param keep_columns: |
| List of common column names to keep from the inputs of the evaluation runs to compare. |
| :returns: |
| Pandas DataFrame with the score comparison. |
| """ |
| if not isinstance(other, EvaluationRunResult): |
| raise ValueError("Comparative scores can only be computed between EvaluationRunResults.") |
|
|
| this_name = self.run_name |
| other_name = other.run_name |
| if this_name == other_name: |
| warn(f"The run names of the two evaluation results are the same ('{this_name}')") |
| this_name = f"{this_name}_first" |
| other_name = f"{other_name}_second" |
|
|
| if self.inputs.keys() != other.inputs.keys(): |
| warn(f"The input columns differ between the results; using the input columns of '{this_name}'.") |
|
|
| pipe_a_df = self.to_pandas() |
| pipe_b_df = other.to_pandas() |
|
|
| if keep_columns is None: |
| ignore = list(self.inputs.keys()) |
| else: |
| ignore = [col for col in list(self.inputs.keys()) if col not in keep_columns] |
|
|
| pipe_b_df.drop(columns=ignore, inplace=True, errors="ignore") |
| pipe_b_df.columns = [f"{other_name}_{column}" for column in pipe_b_df.columns] |
| pipe_a_df.columns = [f"{this_name}_{col}" if col not in ignore else col for col in pipe_a_df.columns] |
|
|
| results_df = pd_concat([pipe_a_df, pipe_b_df], axis=1) |
|
|
| return results_df |
|
|