6 Loss Functions
6.1 Misc
- Also see
- Make sure your Loss Function and Performance metrics correlate
- Loss function are typically minimized and metrics are typically maximized. So, when choosing a loss function, you should choose one that negatively correlates with your metric.
- Simulate some data, fit a model that using various loss functions, and score them with your performance metric. Then, choose the loss function that correlates with your metric the best.
- Example: Cross-Entropy vs. Accuracy
- Image is from a video, disregard the solid gridlines (actuallly video screens)
- X-Axis is Accuracy.
- Shows how the low values of the loss function correlate well with the performance metric, Accuracy.
- Mean Absolute Error (MAE) isn’t typically used (instead of MSE) because it is not differentiable, thus gradient descent can’t be used to find its minimum. Stochastic Gradient Descent is used to overcome this problem. (See Algorithms, ML >> SVM >> Regression)
- When you optimize a model using the MAE or MSE, your model will end up predicting the median or mean of the distribution, respectively
- Absolute error is more robust to outliers
- Minimizes the median value whereas MSE minimizes the mean value
- Stochastic Gradient Descent: the gradient of the loss is estimated each sample at a time and the model is updated along the way with a decreasing strength schedule (aka learning rate)
- Also see Model building, sklearn >> Algorithms>> Stochaistic Gradient Descent (SGD)
6.2 0/1 Loss (aka binary loss)
- It’s basically the misclassification rate (additive inverse of accuracy) but as a loss function
- Sum the 1s for misclassified labels and 0 for correct classifications. Then, take the average
- 1 (or maybe 0 too) doesn’t have to be the incorrect classification value. You could weight the incorrect classifications differently. If it’s 2, then you’d double the weight against an incorrect classifcation. Although it wouldn’t be 0/1 loss anymore.
- Often used to evaluate classifiers, but is not useful in guiding optimization since it is non-differentiable and non-continuous.
6.3 Asymmetric Huber Loss
- θ is the δ in Huber loss. There are two “delta” parameters now to adjust the amount of asymmetry on either the left/lower half (θL) or right/upper half (θR)
- x is “y-f(x)” in Huber loss
- The bottom condition, θR(2x - θR) is equivalent to the bottom condition in Huber loss
- The top condition is just the conjugate(?) of the bottom condition
- The “1/2” isn’t present in the middle condition for some reason as it is for the top condition in Huber loss.
- Checked Wiki and 1/2 is present in Huber loss there. Guess you could add it.
- Potentially a more programmer-friendly version
6.4 Brier Loss
- How far your predictions lie from the true values
- a mean square error in the probability space
6.5 Focal Loss
- Where γ is the focusing parameter and α is a class balancing weight
- Both are hyperparameters that can be tuned during CV
- γ = 0 results in the cross-entropy formula
- As γ gets larger, the more importance will be given to misclassified examples and less importance towards easy examples.
- γ = 2 and α =.25 is a recommended starting points (paper)
- α can be set as the inverse class frequency
- γ = 2 and α =.25 is a recommended starting points (paper)
- Gist with Focal loss function for keras
- Focuses on the examples that the model gets wrong rather than the ones that it can confidently predict, ensuring that predictions on hard examples improve over time rather than becoming overly confident with easy ones
6.6 Geometric Trace Mean Squared Error (GTMSE)
\[ GTMSE = \sum_{j=1}^h \log \frac{1}{T-j}\sum_{t=1}^{T-j} e^2_{t+j|t} \]
- Multi-step Estimators and Shrinkage Effect in Time Series Models
- Forecasting loss function that imposes shrinkage on parameters similar to other estimators but does it more mildly because of the logarithms in the formula
- What the logarithms do is make variances of all forecast errors similar to each other. As a result, when used, GTMSE does not focus on the larger variances as other methods do but minimizes all of them simultaneously similarly.
- Making models less stochastic and more inert. Therefore, performance depends on each specific situation and the available data and shouldn’t be used universally.
- Packages: {smooth}, {greybox}
6.7 Huber Loss
- Huber loss (green, delta =1) and squared error loss (blue) as a function of y-f(x)
- A combination of squared loss and absolute loss
- For residuals > δ, the absolute value of the residual is used which is based on the unbiased estimator of the median
- Therefore more robust to outliers
- The loss is the square of the usual residual (y — f(x)) only when the absolute value of the residual is smaller than a fixed parameter, δ.
- For residuals > δ, the absolute value of the residual is used which is based on the unbiased estimator of the median
- Delta is a tuneable parameter
- The function is quadratic for small values of |y - f(x)|, and linear for large values
- Commonly used in deep learning where it helps to avoid the exploding gradient problem due to its insensitivity to large errors.
- {yardstick::huber_loss}
6.8 Mean Absolute Log Error (MALE)
- article
- f is the forecast and y is the observed
- Similar to RMSLE
- Requirement: data are strictly positive: they do not take zero or negative values
- Predicts the median of your distribution
- Underestimates and overestimates are punished equally harshly
- Transform it back onto a relative scale by taking the exponential (exp(MALE)) which is the (geometric) mean relative error
- Interpetation:
- Example: exp(MALE) of 1.2 means you expect to be wrong by a factor of 1.2 in either direction, on average. (Explain this to your boss as a 20% average percentage error.)
- Comparison with MAPE, sMAPE
- Provided your errors are sufficiently small (e.g. less than 10%), it probably doesn’t make much difference which of the MALE, sMAPE or MAPE you use.
- MALE is is between the MAPE and sMAPE in sensitivity to outliers. For
- Example: compared to being wrong by +10x, being wrong by +100x gives you 1.2 times your original sMAPE, 2 times your original MALE, and 9 times your original MAPE.
6.9 Mean Absolute Scaled Error (MASE)
- Alternative to using percentage errors (e.g. MAPE) when comparing forecast accuracy across series with different units since it is independent of the scale of the data
- Greater than one: worse than the average naive forecast
- Less than one: better than the average naive forecast
- Non-seasonal ts scaled error
- Seasonal ts scaled error
- MASE = mean(|qj|)
6.10 Mean Absolute Scaled Log Error (MASLE)
- Then, MASLE = Mean(SLE)
- The area cut-off in the numerator should read “f T+h” for hth-step ahead forecast
- f is the forecast and y is the observed
- Scaled version of MALE
- It’s been normalized by the error of a benchmark method (e.g. the naive or seasonal naive methods) to place the errors on the same scale
6.11 Mean Balanced Log Loss (aka Balanced Cross-Entropy)
- Adds a weight, β, to the log loss formula to handle class imbalance
- Can tune during CV
- Where p is the observed 0/1 and p^ is the predicted probability
- For just balanced log-loss, remove the summation and 1/N
6.12 Mean log loss (aka cross-entropy)
Where p(yi) is the model output and yi is the observed 0/1
For just log-loss, remove the summation and 1/N
If the model predicts that an observation should be labeled 1 and assigns a high probability to that prediction, a high penalty will be incurred when the true label is 0. If the model had assigned a lower probability to that prediction, a lower penalty would have been incurred
- (might) be equivalent to K-L Divergence
Above equation is for a binary model but can be extended multiclassification models
- You’d replace the above expression and have a term for each category — each being yi * log(p(yi)) where each p(yi) is the probability for that category
{yardstick::mn_log_loss}
Notes
- Handles class imbalance poorly
- In sliced season 2 w/Silge on bird strike dataset, mean log loss scored a balanced outcome variable better than the unbalanced one.
- Fails to distinguish between hard and easy examples. Hard examples are those in which the model repeatedly makes huge errors, whereas easy examples are those which are easily classified. As a result, Cross-Entropy loss fails to pay more attention to hard examples.
- Handles class imbalance poorly
6.13 Mean Squared Error (MSE)
- A quadratic loss function that’s calculation is based on the squared errors, so it is more sensitive to the larger deviations associated with tail events
- A custom asymmetric MSE loss function can better account for large deviations in errors
- With α∈(0,1) being the parameter we can adjust to change the degree of asymmetry
6.14 Quantile Loss
\[ \frac{1}{n} \sum_{i=1}^n \left | \gamma - \mathbb{1}(f(x_i) - \hat f(x_i) \lt 0) \right | \cdot \left | f(x_i) - \hat f(x_i) \right | \]
- AKA Pinball Loss
- \(\gamma \in (0, 1)\) is the quantile. Residuals less than 0 use \(\gamma - 1\) as a penalty and residuals greater than 0 use \(\gamma\)
- Allows you to predict a certain percentile of the outcome variable
- Not effective at predicting tail events
- Using the median assigns equal weights to underpredictions and overpredictions. This is also equivalent to using the MAE loss function.
- For quantiles greater than 50, underpredictions are penalyzed more than overpredictions
- The asymmetrical penalty pushes the model to prefer overprediction to underprediction for high quantiles which places importance of accurately capturing higher values in the distribution
- Example: 90th quantile (source)
- X-Axis is mislabelled — it’s just the prediction value.
- 100 is the vertex which is the actual value and a prediction of 105 is a loss value of 0.5
- Underpredictions (when the actual value is greater than the predicted value): The penalty is 0.9 * (actual - predicted)
- Overpredictions (when the actual value is less than or equal to the predicted value): The penalty is -0.1 * (actual - predicted)
- The intuition behind this is that for the 90th quantile, we expect the actual value to be below our prediction 90% of the time. When it’s not (i.e., when we underpredict), we incur a much larger penalty.
- For quantiles less than 50, overpredictions are penalyzed more than underpredictions
- Similar reasoning.
6.15 Relative Absolute Error
- ( |y_pred — y_true | / y_true)
6.16 Root Mean Squared Logarithmic Error (RMSLE)
Also see Understanding the metric: RMSLE
mainly used when predictions have large deviations. Values range from 0 up to millions and we don’t want to punish deviations in prediction as much as with MSE
Predicts the geometric mean of your distribution
- The geometric mean is most useful when numbers in the series are not independent of each other or if numbers tend to make large fluctuations.
Similar to MALE except the exp(RMSLE) is still not directly intepretable
You can just add 1 to your outcome then log it then used RMSE for your loss
Drob did log(outcome) + 1 but that doesn’t match the formula so I dunno if he’s right.
mutate(outcome = log(outcome + 1)
OR extend {yardstick} by creating a custom metric to calculate RMSLE
library(rlang) <- function(truth, estimate, na_rm = TRUE, ...) { rmsle_vec <- function(truth, estimate) { rmsle_impl sqrt(mean((log(truth + 1) - log(estimate + 1))^2)) }metric_vec_template( metric_impl = rmsle_impl, truth = truth, estimate = estimate, na_rm = na_rm, cls = "numeric", ... ) }<- function(data, ...) { rmsle UseMethod("rmsle") }<- new_numeric_metric(rmsle, direction = "minimize") rmsle <- function(data, truth, estimate, na_rm = TRUE, ...) { rmsle.data.frame metric_summarizer( metric_nm = "rmsle", metric_fn = rmsle_vec, data = data, truth = !!enquo(truth), estimate = !!enquo(estimate), na_rm = na_rm, ... ) }
6.17 Symmetric Mean Absolute Percentage Error (SMAPE)
SMAPE = (1/n) * Σ(|forecast – actual| / ((|actual| + |forecast|)/2) * 100
<- function(a, f) {return (1/length(a) * sum(2*abs(f-a) / (abs(a)+abs(f))*100))} smape
packages: yardstick, Metrics
Check notebook — think this metric has issues with values around 0
Handles data where the scale varies over time; it is relatively comparable across time series; deals reasonably well with outliers
Underestimates and overestimates are punished equally harshly
- MAPE punished overestimates more severely than underestimates
6.18 Weighted Root Mean Square Scaled Error (WRMSSE)
- guessing the 42840 is nrow of the dataset in the example and there was no explanation of how wi is calculated