官术网_书友最值得收藏!

How to choose the right activation function

The activation function is decided depending upon the objective of the problem statement and the concerned properties. Some of the inferences are as follows:

  • Sigmoid functions work very well in the case of shallow networks and binary classifiers. Deeper networks may lead to vanishing gradients.

  • The ReLU function is the most widely used, and try using Leaky ReLU to avoid the case of dead neurons. Thus, start with ReLU, then move to another activation function if ReLU doesn't provide good results.

  • Use softmax in the outer layer for the multi-class classification.

  • Avoid using ReLU in the outer layer.

主站蜘蛛池模板: 洱源县| 佛教| 景洪市| 东至县| 正定县| 肥东县| 沧源| 苍梧县| 兴安县| 安图县| 新巴尔虎右旗| 南华县| 沂南县| 抚宁县| 岳阳市| 达拉特旗| 启东市| 花垣县| 阜城县| 景德镇市| 克拉玛依市| 宝鸡市| 于田县| 静宁县| 织金县| 濮阳县| 沛县| 印江| 武定县| 定陶县| 安平县| 江北区| 临安市| 西乌| 陆川县| 于田县| 木兰县| 宁河县| 泗阳县| 灵寿县| 辽阳县|