Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

What is a loss function in simple words?

Can anyone please explain in simple words and possibly with some examples what is a loss function in the field of machine learning/neural networks?

This came out while I was following a Tensorflow tutorial: https://www.tensorflow.org/get_started/get_started

like image 752
Federico Avatar asked Mar 18 '17 18:03

Federico


People also ask

What is a loss function in simple terms?

In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cost" associated with the event.

What is a loss function in statistics?

A loss function specifies a penalty for an incorrect estimate from a statistical model. Typical loss functions might specify the penalty as a function of the difference between the estimate and the true value, or simply as a binary value depending on whether the estimate is accurate within a certain range.

Why is loss function used?

At its core, a loss function is a measure of how good your prediction model does in terms of being able to predict the expected outcome(or value). We convert the learning problem into an optimization problem, define a loss function and then optimize the algorithm to minimize the loss function.

Which of the following defines a loss function?

1. Mean Square Error / Quadratic Loss / L2 Loss. We define MSE loss function as the average of squared differences between the actual and the predicted value. It's the most commonly used regression loss function. The corresponding cost function is the mean of these squared errors (MSE).


2 Answers

It describes how far off the result your network produced is from the expected result - it indicates the magnitude of error your model made on its prediciton.

You can then take that error and 'backpropagate' it through your model, adjusting its weights and making it get closer to the truth the next time around.

like image 127
Piotr Trochim Avatar answered Oct 22 '22 00:10

Piotr Trochim


The loss function is how you're penalizing your output.

The following example is for a supervised setting i.e. when you know the correct result should be. Although loss functions can be applied even in unsupervised settings.

Suppose you have a model that always predicts 1. Just the scalar value 1.

You can have many loss functions applied to this model. L2 is the euclidean distance.

If I pass in some value say 2 and I want my model to learn the x**2 function then the result should be 4 (because 2*2 = 4). If we apply the L2 loss then its computed as ||4 - 1||^2 = 9.

We can also make up our own loss function. We can say the loss function is always 10. So no matter what our model outputs the loss will be constant.

Why do we care about loss functions? Well they determine how poorly the model did and in the context of backpropagation and neural networks. They also determine the gradients from the final layer to be propagated so the model can learn.

As other comments have suggested I think you should start with basic material. Here's a good link to start off with http://neuralnetworksanddeeplearning.com/

like image 44
Steven Avatar answered Oct 21 '22 22:10

Steven