-
Notifications
You must be signed in to change notification settings - Fork 293
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fully connected initialization #28
Comments
which file? which line? :) |
Hi, thanks for the response,
In wide-residual-networks/models/utils.lua the FCinit function, which is
used in wide-resnet.lua and in vgg.lua,
Why the fully connected layers aren't initialized in the same manner as the
convolutional layers?
where in the code is the FC layers initialization ?
Thanks!
…On Wed, Dec 7, 2016 at 7:38 PM, Remi ***@***.***> wrote:
which file? which line? :)
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#28 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AMoNPdWJWiYSbC4A45XQ4JLGBA4quMPWks5rFu8agaJpZM4LGqf7>
.
--
Andrey Isakov
|
FCinit and MSRinit applied to WideResNet I guess, it is just a matter of hyper parameter tuning. Maybe, the author could illuminate our thinking :p |
@Andreyisakov FC layers are initialized with Xavier, it doesn't affect the final accuracy https://github.com/torch/nn/blob/master/Linear.lua#L25 |
Today I think we are using Xavier Initialization with Uniform Distribution (Default Torch) to Fully Connected Layers and Kaiming Initialization with Gaussian Distribution (MSRinit Function) to Convolutional Layers. I don't see why not use the same Kaiming Initialization to both Convolutional and Fully Connected Layers, at least for the purpose of uniformity of treatment. The following paper shows that Kaiming Initialization is supposed to be better than Xavier Initialization, at least to Convolutional Layers. |
can you please explain why the fully connect layers weights are not initialize with MSRinit ,
how they are initialize ?
The text was updated successfully, but these errors were encountered: