Loss function in GradientBoostingRegressor

Scikit Learn GradientBoostingRegressor:

I was looking at the scikit-Learn documentation for GradientBoostingRegressor.

Here it says that we can use 'ls' as a loss function which is least squares regression. But I am confused since least squares regression is a method to minimize the SSE loss function.

So shouldn't they mention SSE here?

Topic loss-function terminology gbm scikit-learn machine-learning

Category Data Science


Note that the algorithm is called Gradient Boosting Regressor.

The idea is that you boost decision trees minimizing the gradient. This gradient is a loss function that can take more forms.

The algorithm aggregates each decision tree in the error of the previously fitted and predicted decision tree. There you have your desired loss function.

This parameter is regarding that.


It would seem that you are over-interpreting what is essentially just convenience shorthand names for the model arguments, and not formal terminology; here, "‘ls’ refers to least squares regression" should be interpreted as "'ls' is the loss function used in least-squares regression".

Formally you do have a point of course - sse would be a more appropriate naming convention here; discussions about such naming conventions are not uncommon among the community, see for example the thread loss function name consistency in gradient boosting (which BTW was resolved here). And you would be most welcome opening a relevant issue for the convention used here.

About

Geeks Mental is a community that publishes articles and tutorials about Web, Android, Data Science, new techniques and Linux security.