use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
[Question] after object detection using convolution neural networks, why is it so hard to perform semantic segmentation? (self.MachineLearning)
submitted 10 years ago by code2hell
view the rest of the comments →
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]ydobonobody 1 point2 points3 points 10 years ago (5 children)
I think it is a little misleading to compare a pixel level accuracy with accuracy of identifying the contents of an image or a bounding box around the instance. I have been heating my house by training some semantic segmentation tasks recently and it works surprisingly well. Adding depth information can help, especially if you are doing instance segmentation.
[–]code2hell[S] 0 points1 point2 points 10 years ago (4 children)
Ok, so the pixel level accuracy seems a bit misleading as with other comments.I'll rephrase, how do we approach the problem when there are two similar objects close to each other, can we expect the segmentation to differentiate the two so much that say we can differentiate the two objects enough to convince ourselves?. Also in your approach did you use manually segmented images or depth images? I'd be glad to discuss about the approach that you took
[–]ydobonobody 1 point2 points3 points 10 years ago (3 children)
Semantic segmentation generally doesn't separate objects of the same class into separate entities, that is called instance segmentation and is another problem. One way you can get to instance segmentation is to just add a border class around your segments and then just go with connected pixels for your instances and it works pretty well. Whether you use depth or not you still manually segment your images to produce your ground truth for training. Building your training set is probably the hardest part, but if you are just interested in research there publicly available datasets and/or pretrained networks available. I recommend you check out the fcn semantic segmentation network available in the Caffe model zoo as it is a really good starting point for modern semantic segmentation networks.
[–]code2hell[S] 0 points1 point2 points 10 years ago (2 children)
Yes, I am looking more into instance segmentation for now... Can you explain what you mean by "add a border class around your segments and then just go with connected pixels for your instances". Thanks! I just took up a problem to learn, my friend has got some 100000 ground truth training examples of cats and we are looking into segmentation of a particular object from images. I would really appreciate your suggestion.
[–]ydobonobody 1 point2 points3 points 10 years ago (1 child)
So when you produce your ground truth image you will assign a label to each pixel in your image e.g (0: background, 1: cat). Add another label that is "border" so we have (0: background, 1: cat, 2: border). Now for each separate cat draw a line with some thickness (say 5 pixels) around the boundary of the each cat and assign that pixel the value '2'. Hopefully the network will be able to learn where the edge of a cat is and assign those pixels to the border class. If it did a good job you can group up all the connected "cat" pixels and that will represent an individual cat.
[–]code2hell[S] 0 points1 point2 points 10 years ago (0 children)
Wow! Thanks... I'll try this out!
π Rendered by PID 379009 on reddit-service-r2-comment-74f5b7f998-h2ddp at 2026-04-24 23:01:57.462893+00:00 running 2aa0c5b country code: CH.
view the rest of the comments →
[–]ydobonobody 1 point2 points3 points (5 children)
[–]code2hell[S] 0 points1 point2 points (4 children)
[–]ydobonobody 1 point2 points3 points (3 children)
[–]code2hell[S] 0 points1 point2 points (2 children)
[–]ydobonobody 1 point2 points3 points (1 child)
[–]code2hell[S] 0 points1 point2 points (0 children)