--- title: MASE emoji: 🤗 colorFrom: blue colorTo: red sdk: gradio sdk_version: 3.19.1 app_file: app.py pinned: false tags: - evaluate - metric description: >- Mean Absolute Scaled Error (MASE) is the mean absolute error of the forecast values, divided by the mean absolute error of the in-sample one-step naive forecast on the training set. --- # Metric Card for MASE ## Metric Description Mean Absolute Scaled Error (MASE) is the mean absolute error of the forecast values, divided by the mean absolute error of the in-sample one-step naive forecast. For prediction $x_i$ and corresponding ground truth $y_i$ as well as training data $z_t$ with seasonality $p$ the metric is given by: ![image](https://user-images.githubusercontent.com/8100/200009284-7ce4ccaa-373c-42f0-acbb-f81d52a97512.png) This metric is: * independent of the scale of the data; * has predictable behavior when predicted/ground-truth data is near zero; * symmetric; * interpretable, as values greater than one indicate that in-sample one-step forecasts from the naïve method perform better than the forecast values under consideration. ## How to Use At minimum, this metric requires predictions, references and training data as inputs. ```python >>> mase_metric = evaluate.load("mase") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> training = [5, 0.5, 4, 6, 3, 5, 2] >>> results = mase_metric.compute(predictions=predictions, references=references, training=training) ``` ### Inputs Mandatory inputs: - `predictions`: numeric array-like of shape (`n_samples,`) or (`n_samples`, `n_outputs`), representing the estimated target values. - `references`: numeric array-like of shape (`n_samples,`) or (`n_samples`, `n_outputs`), representing the ground truth (correct) target values. - `training`: numeric array-like of shape (`n_train_samples,`) or (`n_train_samples`, `n_outputs`), representing the in sample training data. Optional arguments: - `periodicity`: the seasonal periodicity of training data. The default is 1. - `sample_weight`: numeric array-like of shape (`n_samples,`) representing sample weights. The default is `None`. - `multioutput`: `raw_values`, `uniform_average` or numeric array-like of shape (`n_outputs,`), which defines the aggregation of multiple output values. The default value is `uniform_average`. - `raw_values` returns a full set of errors in case of multioutput input. - `uniform_average` means that the errors of all outputs are averaged with uniform weight. - the array-like value defines weights used to average errors. ### Output Values This metric outputs a dictionary, containing the mean absolute error score, which is of type: - `float`: if multioutput is `uniform_average` or an ndarray of weights, then the weighted average of all output errors is returned. - numeric array-like of shape (`n_outputs,`): if multioutput is `raw_values`, then the score is returned for each output separately. Each MASE `float` value ranges from `0.0` to `1.0`, with the best value being 0.0. Output Example(s): ```python {'mase': 0.5} ``` If `multioutput="raw_values"`: ```python {'mase': array([0.5, 1. ])} ``` #### Values from Popular Papers ### Examples Example with the `uniform_average` config: ```python >>> mase_metric = evaluate.load("mase") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> training = [5, 0.5, 4, 6, 3, 5, 2] >>> results = mase_metric.compute(predictions=predictions, references=references, training=training) >>> print(results) {'mase': 0.1833...} ``` Example with multi-dimensional lists, and the `raw_values` config: ```python >>> mase_metric = evaluate.load("mase", "multilist") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0.1, 2], [-1, 2], [8, -5]] >>> training = [[0.5, 1], [-1, 1], [7, -6]] >>> results = mase_metric.compute(predictions=predictions, references=references, training=training) >>> print(results) {'mase': 0.1818...} >>> results = mase_metric.compute(predictions=predictions, references=references, training=training, multioutput='raw_values') >>> print(results) {'mase': array([0.1052..., 0.2857...])} ``` ## Limitations and Bias ## Citation(s) ```bibtex @article{HYNDMAN2006679, title = {Another look at measures of forecast accuracy}, journal = {International Journal of Forecasting}, volume = {22}, number = {4}, pages = {679--688}, year = {2006}, issn = {0169-2070}, doi = {https://doi.org/10.1016/j.ijforecast.2006.03.001}, url = {https://www.sciencedirect.com/science/article/pii/S0169207006000239}, author = {Rob J. Hyndman and Anne B. Koehler}, } ``` ## Further References - [Mean absolute scaled error - Wikipedia](https://en.wikipedia.org/wiki/Mean_absolute_scaled_errorr)