Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Separate running mean and variance calculation #8

Open
Minju-Jung opened this issue Jul 6, 2016 · 1 comment
Open

Separate running mean and variance calculation #8

Minju-Jung opened this issue Jul 6, 2016 · 1 comment

Comments

@Minju-Jung
Copy link

Minju-Jung commented Jul 6, 2016

Thank you for sharing your nice torch source code.

In the paper "Recurrent Batch Normalization", authors use the running (or population) means and variance at each time step (until T_max), separately.
However, I couldn't find these separate running means and variances calculation in your code.
First, bnlstm constructed by nngraph using nn.BatchNormalization().
After bnlstm construction, clones of module are made by model_utils.clone_many_times().

As far as my understanding, in this situation, the sample means and variances are calculated at each time step, but the running means and variances are shared over time by model_utils.clone_many_times().

Could you kindly explain which part am I misunderstanding?

@zhiqiangdon
Copy link

I am also wondering the same question here. It seems you just do the normal BN initialization and usage. There is no code related to how separated mean and var are computed for different iterations? I don't think torch could automatically do this.

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants