Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

receptive field is not composed of directly adacent pixels in the input image #27

Open
barcavia opened this issue Feb 16, 2024 · 0 comments

Comments

@barcavia
Copy link

barcavia commented Feb 16, 2024

Hi, first of all, your work is so inspiring!
I want to consider how the receptive field is established in the network:

In BagNet-9:

The first convolutional layer uses a kernel size of 1x1. Since the stride is 1 by default, each output neuron in this layer sees a receptive field of size 1x1, centered around a single pixel in the input image.

The second convolutional layer uses a kernel size of 3x3 with a stride of 2. This means that each output neuron in this layer sees a receptive field of size 3x3, but with a stride of 2, the receptive fields are not directly adjacent. The receptive fields overlap but are not contiguous.

The third and fourth convolutional layers also use 3x3 kernels with a stride of 2. Similar to the second layer, the receptive fields of neurons in these layers overlap but are not directly adjacent.
Given the strides used in the convolutional layers, the 9x9 receptive field of BagNet-9 is not composed of directly adjacent pixels in the input image.

in your heatmaps visualization, you are dividing the input image into all 9x9 patches, assuming the 9x9 pixels are direct neighbors in the input image.
what am I missing?

Thanks in advance,
Bar

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant