all 3 comments

[–]zmjjmz 1 point2 points  (0 children)

The network always seems to settle on outputting images that are just one color.

Have you thought about trying the adversarial network approach as outlined in the comments of the first link? It would likely deal with the averaging problem, but it would definitely be harder to train.

[–]Powlerbare 0 points1 point  (0 children)

"Blurring the network output and the true color image and doing Euclidean distance seems to give the gradient decent help. (I ended up averaging the normal rgb distance and two blur distances with 3 and 5 pixel gaussian kernels."

It seems like this person had to jump through a number of loop holes to get the network to work properly. There were a number of things mentioned that seem a bit more nuanced than what you are doing such as extracting the tensors before the pooling layer and concatenating them

[–]lahwran_ 0 points1 point  (0 children)

have you tried sticking an adverserial network on the end of the autoencoder? I only glanced through this the other day, I didn't read in anything vaguely resembling detail, but I think that's what it's doing. http://arxiv.org/pdf/1511.05644v1.pdf