Skip to content

Conversation

@AhmedMesih
Copy link

when I tested Tesla P100, using return F.mish(x) insted of return x * F.softplus(x).tanh() significantly reduced GPU memory usage.

@WongKinYiu
Copy link
Owner

thanks, mish-cuda even use less memory than nn.SiLU.
i think i have to wait ngc pytorch version 21.07 for testing if mish is supported or not.

@AlexeyAB
Copy link

AlexeyAB commented Jul 9, 2021

Yes, Mish was proposed 3 Sep 2019 pytorch/pytorch#25584 (comment) and 25 May 2021 it was added to Pytorch 1.9.0 pytorch/pytorch#58940
Also Mish is in the TensorFlow: https://www.tensorflow.org/addons/api_docs/python/tfa/activations/mish

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants