Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MSR improvements #1

Open
wants to merge 2 commits into
base: msr_init
Choose a base branch
from
Open

MSR improvements #1

wants to merge 2 commits into from

Conversation

jjkjkj
Copy link

@jjkjkj jjkjkj commented Feb 18, 2015

Hi!
I want to collaborate in your development of MSR weight filler.

I added support for inner product layer. Current code doesn't work for ip because fan_in and fan_out are computed differently for ip and conv weight blobs.
Because ip and conv layers can have same blob dimension(100->100 ip, kernelsize = 100, in and out channels = 1, so we can't detect blob type by looking ant first 2 dimensions of blob) with different meaning, added optional FillerLayerType layer_type to FillerParameter message.

Added support for non-zero negative slope ReLU as described in original paper.

Also i added scale field to FillerParameter message.
Scale is needed in case scale of network inputs and outputs are not same ( in original paper they use scale<1 for top ip layers), or we want to manualy change amplitude of network's signal in any place.
Maybe Xavier filler also need support for this field.

Tests now expects all random variable's deviations not to exceed 5*std.

expecting not exceed 5*std of variable
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant