Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

EltAffine Functionality #2

Closed
Darwin2011 opened this issue Jan 8, 2016 · 2 comments
Closed

EltAffine Functionality #2

Darwin2011 opened this issue Jan 8, 2016 · 2 comments

Comments

@Darwin2011
Copy link

Hi, @ducha-aiki

I am trying to figure out BN implementations from the PR you test and there's no bias and shift implemented there.

I also notice that from your experiments, it seems that BN + Affine doesn't improve performance that much from the initial training stages.

And in your Caffe fork, https://github.com/ducha-aiki/caffe, there's another version of bn implementation as Caffe PR 1965, which implements shift and bias.

So may I know why such two operations is dropped in Caffe Upstream you test? Do they even hurt performance? Or what version should I choose to use?

Thanks a lot.

.

@ducha-aiki
Copy link
Owner

@Darwin2011 the caffe maintainers said that want layers to be smaller building block, so separate EA. EA could be implemented with dummydatalayer in very unelegant way, but with existing layers. So my version was dropped.

However, I am going to PR bias layer and scale layer, which together would EA and maintainers are ok with merging them, when ready.

@Darwin2011
Copy link
Author

Thanks a lot for your kind help.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants