You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
At this stage, pool5 has size [batch_size, 7, 7, 512]. Now, as far as I understand, you are using a filter of size 7 by 7, in order to make your feature map of size [batch_size, 1, 1, 4096]. However, if you look at the code of conv2d_basic(...), the code is:
The problem here is that the output of conv6 actually remains [batch_size, 7, 7, 4096] because you're using padding, and I am not sure that this is what we want. If we look at the official code, we'll see the sixth convolutional layer coded as:
They aren't using padding in this layer, which means that conv6 is actually of size [batch_size, 1, 1, 4096]. Pretty sure this is the entire point of conv6.
Am I missing something, or that part of the code was a mistake from your part?
Anyway, cheers for the code. The cleanest TF implementation of F-CNN I have seen so far.
The text was updated successfully, but these errors were encountered:
Hi @TheRevanchist
It's been a while since I looked into the code, but if it is implemented with padding I think you are right in that this was a mistake on my part (Nice catch!). Having said that, I believe since we are using convolutions all the way it might not be affecting the segmentation.
Anyways thanks for pointing it out. If you have already fixed it you could do a PR and I will review and merge the changes.
I kind of played with it, but the results seem to be significantly worse than yours (visually at least, the validation error rate seem to be around the same). Which made me doubt myself if there is something I didn't understand and so the padding in that layer was needed, or I made some mistake in the code.
I will give another look to it soon, and see if I can fix it.
Hi,
On the sixth convolutional layer, the code is:
At this stage, pool5 has size [batch_size, 7, 7, 512]. Now, as far as I understand, you are using a filter of size 7 by 7, in order to make your feature map of size [batch_size, 1, 1, 4096]. However, if you look at the code of conv2d_basic(...), the code is:
The problem here is that the output of conv6 actually remains [batch_size, 7, 7, 4096] because you're using padding, and I am not sure that this is what we want. If we look at the official code, we'll see the sixth convolutional layer coded as:
They aren't using padding in this layer, which means that conv6 is actually of size [batch_size, 1, 1, 4096]. Pretty sure this is the entire point of conv6.
Am I missing something, or that part of the code was a mistake from your part?
Anyway, cheers for the code. The cleanest TF implementation of F-CNN I have seen so far.
The text was updated successfully, but these errors were encountered: