-
Notifications
You must be signed in to change notification settings - Fork 18.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
SQUASH layer -- takes 1 or more inputs and produces no output #624
Conversation
I guess it can be useful if one want to discard some blobs and not appear as |
To join the naming party, how about I'd give this a vote for convenience, since depending on phase or purpose one might want different output blobs. |
I thought the SQUASH name would probably see some backlash; I'll change it to SILENCE when I get a chance. There is normally no need to compute backward, but I put the gradients in because they're very obvious -- if #686 is merged, the net will be able to figure out that it doesn't need to compute the backward pass of the SQUASH/SILENCE layer (as it has no top blobs, so it can neither produce a loss, nor send input to later layers that do produce a loss) so with that change the Backward implementation is only included for completeness -- it would only actually be executed if |
4278286
to
c01f07a
Compare
399eca6
to
280edaf
Compare
This is rebased, renamed, and ready-to-go, pending Travis. |
280edaf
to
9fefd85
Compare
Sweet, thanks for fixing this up! I can't figure out why the Travis build is erroring though... |
9fefd85
to
fbf6b13
Compare
It's because I didn't add a |
fbf6b13
to
d9975a9
Compare
This is useful for suppressing undesired outputs.
d9975a9
to
d8f56fb
Compare
SILENCE layer -- takes 1 or more inputs and produces no output (kudos to @longjon for fixup)
SILENCE layer -- takes 1 or more inputs and produces no output (kudos to @longjon for fixup)
SILENCE layer -- takes 1 or more inputs and produces no output (kudos to @longjon for fixup)
I used this to suppress outputs that I didn't want printed during training. Perhaps we shouldn't have a layer for this, I don't know. But here's a PR.