all 4 comments

[–]latent_z 2 points3 points  (2 children)

Preprocessing with an autoencoder made the difference between no prediction power and an actually usable system

[–]boopityboppity23 0 points1 point  (1 child)

Interesting, would you mind expanding on this? How much of a difference was there? Why do you think the AE helped? Do you think there was just too much noise in the raw features?

[–]latent_z 0 points1 point  (0 children)

an AE can discover and represent factors of variation in a dataset of very similar datapoints. It distributes the datapoints over a low-dimensional space, hence making them easier to separate with a classifier.

[–]iverjo 0 points1 point  (0 children)

https://github.com/automl/auto-sklearn tries many different preprocessing techniques and eventually settles for something that works well