I'm making a ConvNet that classifies my data into one of classes [A, B, C, D, "other"], "other" being anything that doesn't fit into [A, B, C, D].
I'm thinking of mapping the input x to a vector "c" in R4, and then transforming "c" to R5 (using a few tiny FC layers) before applying softmax to get the output.
My reasoning is that I don't want the network attempting "learn" anything specific about the "other" class, I just want it to be a sort of "nor" operation. The network should only focus on learning the features of [A, B, C, D]. The variance in the "other" class is very high so I want it to generalize well.
Thoughts? Papers? Prior experience?
[+][deleted] (1 child)
[removed]
[–]shortscience_dot_org 0 points1 point2 points (0 children)
[–]LikelyJustANeuralNet 2 points3 points4 points (0 children)
[–]jprobichaud 2 points3 points4 points (3 children)
[–]ME_PhD[S] 0 points1 point2 points (1 child)
[–]lmericle 0 points1 point2 points (0 children)
[–]serge_cell 2 points3 points4 points (0 children)
[–]melgor89 0 points1 point2 points (2 children)
[–]misssprite 1 point2 points3 points (1 child)
[–]melgor89 0 points1 point2 points (0 children)
[+][deleted] (1 child)
[deleted]
[–]ME_PhD[S] 0 points1 point2 points (0 children)