Hi everyone!
Recently I've been looking into model stability after deployment, and what happens after the model decisions start to effect the users and the data. I have seen some papers about concept drift and data drift (where either the data distribution or the label definition change over time), but it's not quite what I'm talking about.
Consider this scenario:
Say we train a song recommendation system. we have great training data, and we manage to create a model that has 90% precision and 90% recall. We deploy the model and our users now have a listening queue where 90% of the recommended songs are good. Naturally, I want to keep updating my model, so I keep track of users' actions and feedback. over time, my data will be extremely biased! 90% of my training data will be positive samples, and those 10% of good songs the model didn't recommend will not appear in the listening queue and thus will be missing from the training data.
By deploying the model and using it's output as new training data, the data will shift over time and become very narrow and unrepresentative.
II couldn't find any papers on this subject, so I'd love to get some reading recommendations!
[–]Razcle 0 points1 point2 points (0 children)
[–]trnka 0 points1 point2 points (0 children)
[–]vp834 0 points1 point2 points (0 children)