- Neural Networks with R
- Giuseppe Ciaburro Balaji Venkateswaran
- 191字
- 2021-08-20 10:25:18
Which activation functions to use?
Given that neural networks are to support nonlinearity and more complexity, the activation function to be used has to be robust enough to have the following:
- It should be differential; we will see why we need differentiation in backpropagation. It should not cause gradients to vanish.
- It should be simple and fast in processing.
- It should not be zero centered.
The sigmoid is the most used activation function, but it suffers from the following setbacks:
- Since it uses logistic model, the computations are time consuming and complex
- It cause gradients to vanish and no signals pass through the neurons at some point of time
- It is slow in convergence
- It is not zero centered
These drawbacks are solved by ReLU. ReLU is simple and is faster to process. It does not have the vanishing gradient problem and has shown vast improvements compared to the sigmoid and tanh functions. ReLU is the most preferred activation function for neural networks and DL problems.
ReLU is used for hidden layers, while the output layer can use a softmax function for logistic problems and a linear function of regression problems.
推薦閱讀
- C++ Primer習題集(第5版)
- Python 深度學習
- Python自動化運維快速入門
- PHP 編程從入門到實踐
- MySQL數(shù)據(jù)庫基礎實例教程(微課版)
- VMware虛擬化技術
- Go并發(fā)編程實戰(zhàn)
- Building Serverless Web Applications
- Python自然語言理解:自然語言理解系統(tǒng)開發(fā)與應用實戰(zhàn)
- Visual Basic程序設計全程指南
- JavaScript編程精解(原書第2版)
- Mastering Embedded Linux Programming
- 零基礎學Java第2版
- Xamarin Cross-Platform Development Cookbook
- 例說FPGA:可直接用于工程項目的第一手經(jīng)驗