Spaces:
Sleeping
Sleeping
import numpy as np | |
import matplotlib.pyplot as plt | |
import gradio as gr | |
def modified_huber_loss(y_true, y_pred): | |
z = y_pred * y_true | |
loss = -4 * z | |
loss[z >= -1] = (1 - z[z >= -1]) ** 2 | |
loss[z >= 1.0] = 0 | |
return loss | |
def plot_loss_func(): | |
xmin, xmax = -4, 4 | |
xx = np.linspace(xmin, xmax, 100) | |
lw = 2 | |
plt.clf() | |
fig = plt.figure(figsize=(10, 10), dpi=100) | |
plt.plot([xmin, 0, 0, xmax], [1, 1, 0, 0], color="gold", lw=lw, label="Zero-one loss") | |
plt.plot(xx, np.where(xx < 1, 1 - xx, 0), color="teal", lw=lw, label="Hinge loss") | |
plt.plot(xx, -np.minimum(xx, 0), color="yellowgreen", lw=lw, label="Perceptron loss") | |
plt.plot(xx, np.log2(1 + np.exp(-xx)), color="cornflowerblue", lw=lw, label="Log loss") | |
plt.plot( | |
xx, | |
np.where(xx < 1, 1 - xx, 0) ** 2, | |
color="orange", | |
lw=lw, | |
label="Squared hinge loss", | |
) | |
plt.plot( | |
xx, | |
modified_huber_loss(xx, 1), | |
color="darkorchid", | |
lw=lw, | |
linestyle="--", | |
label="Modified Huber loss", | |
) | |
plt.ylim((0, 8)) | |
plt.legend(loc="upper right") | |
plt.xlabel(r"Decision function $f(x)$") | |
plt.ylabel("$L(y=1, f(x))$") | |
return fig | |
title = "SGD convex loss functions" | |
detail = "This plot shows the convex loss functions supported by SGDClassifiers(Linear classifiers (SVM, logistic regression, etc.) with SGD training)." | |
def explain(name): | |
# print("name=",name) | |
if name == "0-1 loss": | |
docstr = "Explanation for " + name + ": " +\ | |
" This is the simplest loss function used in classification problems. It counts how many mistakes a hypothesis function makes on a training set. " +\ | |
" A loss of 1 is accounted if its mispredicted and a loss of 0 for the correct prediction. " +\ | |
" This function is non differentiable and hence not used in Optimization problems. " | |
elif name == "Hinge loss": | |
docstr = "Explanation for " + name + ": " +\ | |
" This is the loss function used in maximum-margin classification in SVMs. "+\ | |
" Z_i = y_i*(w.T * x_i + b), if Z_i > 0 the point x_i is correctly classified and Z_i < 0 , x_i is incorrectly classified "+\ | |
" Z_i >= 1, hinge loss =0 , Z_i < 1 , hinge loss = 1- Z_i " | |
elif name == "Perceptron loss": | |
docstr = "Explanation for " + name + ": " +\ | |
" This is the linear loss function used in perceptron algorithm. "+\ | |
" The binary classifier function which decides whether the input represented by vector of numbers belongs to a class or not. " | |
elif name == "Squared Hinge loss": | |
docstr = "Explanation for " + name + ":" +\ | |
" This represents the square verison of Hinge loss and used in classification algorithms where Performance is important. "+\ | |
" If we want a more fine decision boundary where we want to punish larger errors more significantly than the smaller errors. " | |
elif name == "Modified Huber loss": | |
docstr = "Explanation for " + name + ":" +\ | |
" The Huber loss function balances the best of both Mean Squared Error and Mean Absolute Error. "+\ | |
" Its a piecewise function and hyper parameter delta is to be found first and then loss optimization step." | |
else: | |
docstr = " Logistic Loss is a loss function used for Logistic Regression. Please refer wikipedia for the Log loss equation." +\ | |
" L2 regularization is most important for logistic regression models. " | |
return docstr | |
with gr.Blocks(title=title) as demo: | |
gr.Markdown(f"# {title}") | |
gr.Markdown(f"# {detail}") | |
gr.Markdown(" **[Demo is based on sklearn docs](https://scikit-learn.org/stable/auto_examples/linear_model/plot_sgd_loss_functions.html#sphx-glr-auto-examples-linear-model-plot-sgd-loss-functions-py)**") | |
with gr.Column(variant="panel"): | |
btn = gr.Button(value="SGD convex loss functions") | |
btn.click(plot_loss_func, outputs= gr.Plot() ) # | |
dd = gr.Dropdown(["0-1 loss", "Hinge loss", "Perceptron loss", "Squared Hinge loss", "Modified Huber loss", "Log Loss"], label="loss", info="Select a Loss from the dropdown for a detailed explanation") | |
# inp = gr.Textbox(placeholder="Select a Loss from the dropdown for a detailed explanation") | |
out = gr.Textbox(label="explanation of the loss function") | |
dd.change(explain, dd, out) | |
demo.launch() |