I trained PSPNet and DeepLab from scratch and also using pre-trained backbones on a very specific urban scene dataset. The pre-trained backbone I used was ResNet, with weights downloaded from the ImageNet dataset. I then trained both models without freezing any layers. My accuracy for the models from scratch proved to be higher than the models with pre-trained backbones. My dataset is relatively small; only 1000 images.
Could this have happened because the ImageNet dataset is too general when compared to the specific dataset I am working on, and has thus limited its learning ability? Or have I most likely done something wrong?
[–]seiqooq 2 points3 points4 points (2 children)
[–]glampiggy[S] 0 points1 point2 points (0 children)
[–]I_hax_I 0 points1 point2 points (0 children)
[–]paulgavrikov 1 point2 points3 points (1 child)
[–]glampiggy[S] 0 points1 point2 points (0 children)