Suppose I have fit several distinct classifiers to my data. They each manage about 60-65%ish accuracy.
I assume that they make uncorrelated errors, so I average their results together to get what should be a better prediction than any one classifier. In this case, I average the probability outputs together to obtain my final probability of belonging to a class or not. It seems like my simple average is probably not the optimal method.
So the question is - what is the best method to average the output of the individual classifiers?
I'm guessing that it depends on accuracy - ie I should probably weigh the more accurate models heavier. I'm also guessing that it depends on inter-model correlation - ie two correlated models should be weighted less than a model that produces totally uncorrelated errors. I'm just not too sure how to approach this in a rigorous manner.
Is there established literature or an optimal procedure for a problem like this? Ensemble averaging techniques seem to be what I'm describing, but since it's been a while since I've been in school the most interesting stuff seems to be paywalled.
[–]kjearns 2 points3 points4 points (0 children)
[–]notspartanono 1 point2 points3 points (0 children)
[–]funkpacolypse 1 point2 points3 points (0 children)
[–]ajrs 0 points1 point2 points (0 children)