DTU DeepLearning: exercise 7

torch activation functions: sigmoid, relu, tanh, softplus. https://morvanzhou.github.io/tutorials/machine-learning/torch/2-03-activation/

torch.nn.activation functions: https://pytorch.org/docs/stable/nn.html#non-linear-activations-weighted-sum-nonlinearity

https://zhuanlan.zhihu.com/p/34998569

https://blog.csdn.net/roguesir/article/details/81263442

 

posted @ 2019-10-28 19:27  keeps_you_warm  阅读(122)  评论(0编辑  收藏  举报