-
Notifications
You must be signed in to change notification settings - Fork 615
please add more activation functions #437
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Thanks @bionicles! So happy to accept PRs for activations such as swish, lisht, etc. I'm less sold on the value of aliasing |
I guess one nice behavior is being able to reference activations as strings rather than functions, which is mostly a convenience but still useful for reducing boilerplate when doing hyperparameter tuning. |
@seanpmorgan @kyleabeauchamp updated the code. yeah, for our architecture search project it's handy to just use strings, but yes, we can directly pass those functions
|
|
Please assign rrelu to me and it seems swish has been implemented in tensorflow.nn module.@seanpmorgan |
|
here are parametric linear, polynomial, and a parametric swish: (tends to blow up and make NaN tho)
|
also, here's Logistic Map, which is (if you believe wikipedia) a simple function on the "Edge of Chaos"
we could also re-sample "r" each call of the function:
|
We have already had C++/CUDA kernel for gelu activation, which is much faster than pure Python operations. |
@bionicles Thank you very much for all of these. I think a lot of these are now implemented or under review ( However, this issue format makes it very difficult for us to evaluate specific activations and determine who will be working on them. For that reason I'm going to close this issue...but feel free to open a single issue per missing activation that'd you would like to propose. Just a note I don't think we'll be accepting any of the alias'ed activations like (tf.sin). IMO if you're building architecture search you can quickly create a dictionary if you want string shortcuts. |
From the original list we are tracking Soft-argmax at #1364 |
@bionicles I actually quite interested in cahotic activation functions - #437 (comment) - Logistic Map I have couple of questions regarding this code snipset.
Thanks! |
@jvishnuvardhan @yongtang @seanpmorgan follow-up on the tf issue
System information
Describe the feature and the current behavior/state.
activations are high-yield because they dramatically influence performance for little code
Will this change the current api? How?
just adds more activations
Who will benefit with this feature?
people doing hyperparameter search can benefit especially
Any Other info.
here is an updated python file with some activations (converted the if/elif stuff into a lookup table at the bottom)
The text was updated successfully, but these errors were encountered: