As a beginner in ml, I am watching a video on YouTube about designing a model of song recommendations on Spotify. For each user, there to be predictions about which songs to be recommended.
One of the features was number of historical engagements, such as views/likes.
The point I don't understand is when the interviewer asks what happens if some user is just clicking more than others in general because they are just more active in general they are clickers who click a lot. and explains that this is why we would need to normalize the clicks (for example, by dividing by some total of a user's activity).
If a model is predicting per group as here, then the prediction is for songs per user. So do you really have to normalize? What is the difference if User A has 900 clicks for Song A and 100 clicks for Song B, while User B has 20 clicks for Song A, 5 clicks for Song B, and 5 clicks for Song C. Would it matter? At the end of the day, we are deciding between songs per user we don't care about different users.
And even if you were to normalize, you could simulate and get User A say 0.9 for Song A and 0.1 for Song B. But User B watched 3 songs, so maybe his distribution is 0.6, 0.2, and 0.2.
How does the normalization help you? You still can't compare 0.9 and 0.6. They are the best scores per group, but still are incomparable across groups. How would raw click numbers differ from ratios?
But per user it would work to get the greatest percentage just as in raw numbers.
So my point is would you need to normalize when predictions are per groups. Do you have to care about a comparison across the groups?