(12) What is alternating least square method in recommendation system? - Quora
In SGD you are repeatedly picking some subset of the loss function to minimize -- one or more cells in the rating matrix -- and setting the parameters to better make just those 0.In ALS you're minimizing the entire loss function at once, but, only twiddling half the parameters. That's because the optimization has an easy algebraic solution -- if half your parameters are fixed. So you fix half, recompute the other half, and repeat. There is no gradient in the optimization step since each optimization problem is convex and doesn't need an approximate approach. But, each problem you're solving is not the "real" optimization problem -- you fixed half the parameters.
You initialize M with random unit vectors usually. It's in feature space so wouldn't quite make sense to have averages over ratings.
Read full article from (12) What is alternating least square method in recommendation system? - Quora
No comments:
Post a Comment