Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

question about the format of the top3 scores in the precomputed dataset. #14

Open
WxWstranger opened this issue Mar 27, 2020 · 2 comments
Assignees
Labels
question Further information is requested

Comments

@WxWstranger
Copy link

WxWstranger commented Mar 27, 2020

First of all, thanks a lot for sharing this great project. I want to use this model to predict the scenes on COCO dataset, however, I'm confused about the format of the top3 scores in the precomputed dataset. Why is the highest score of a pixel about 0.4 but 1. what function do you use to get these scores?

In https://github.com/CSAILVision/semantic-segmentation-pytorch the scores are product by softmax.

Thanks a lot!

@alexlopezcifuentes
Copy link
Member

Hi! This question is further discussed in the paper, in case you want to take a look.

In summary, if you have an RGB image as input for the Semantic Segmentation Network you will end up having a probability distribution of semantic scores for each pixel. This means that if you have 120 semantic classes, for each pixel you will have a distribution of 120 scores. The sum of all that scores must be 1 (probability distribution) but this does not mean that the top score must be 1. You can have 120 scores summing up to 1 with the highest one being 0.4.

@alexlopezcifuentes alexlopezcifuentes self-assigned this Mar 27, 2020
@alexlopezcifuentes alexlopezcifuentes added the question Further information is requested label Mar 27, 2020
@WxWstranger
Copy link
Author

Hi! This question is further discussed in the paper, in case you want to take a look.

In summary, if you have an RGB image as input for the Semantic Segmentation Network you will end up having a probability distribution of semantic scores for each pixel. This means that if you have 120 semantic classes, for each pixel you will have a distribution of 120 scores. The sum of all that scores must be 1 (probability distribution) but this does not mean that the top score must be 1. You can have 120 scores summing up to 1 with the highest one being 0.4.

Thanks for your quick reply!

So I think the scores is simply got by Softmax like the implement https://github.com/CSAILVision/semantic-segmentation-pytorch.

However, when I run the evaluate.py on MITIndoor dataset, I checked the top3 scores of a pixel of the first sample, I got the following data:
2020-03-27_183844

The largest score is about 0.4, the 2nd and 3rd large scores are 0. Is that means the sum of all the scores is not 1? or I misunderstand somewhere?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants