- Deep Learning with Keras
- Antonio Gulli Sujit Pal
- 267字
- 2021-07-02 23:58:02
Problems in training the perceptron and a solution
Let's consider a single neuron; what are the best choices for the weight w and the bias b? Ideally, we would like to provide a set of training examples and let the computer adjust the weight and the bias in such a way that the errors produced in the output are minimized. In order to make this a bit more concrete, let's suppose we have a set of images of cats and another separate set of images not containing cats. For the sake of simplicity, assume that each neuron looks at a single input pixel value. While the computer processes these images, we would like our neuron to adjust its weights and bias so that we have fewer and fewer images wrongly recognized as non-cats. This approach seems very intuitive, but it requires that a small change in weights (and/or bias) causes only a small change in outputs.
If we have a big output jump, we cannot progressively learn (rather than trying things in all possible directions—a process known as exhaustive search—without knowing if we are improving). After all, kids learn little by little. Unfortunately, the perceptron does not show this little-by-little behavior. A perceptron is either 0 or 1 and that is a big jump and it will not help it to learn, as shown in the following graph:

We need something different, smoother. We need a function that progressively changes from 0 to 1 with no discontinuity. Mathematically, this means that we need a continuous function that allows us to compute the derivative.
- 新型電腦主板關鍵電路維修圖冊
- Creating Dynamic UI with Android Fragments
- 電腦組裝與維修從入門到精通(第2版)
- 電腦維護365問
- Hands-On Machine Learning with C#
- 微服務分布式架構基礎與實戰:基于Spring Boot + Spring Cloud
- STM32嵌入式技術應用開發全案例實踐
- 筆記本電腦維修300問
- 計算機應用基礎案例教程(Windows 7+Office 2010)
- 分布式存儲系統:核心技術、系統實現與Go項目實戰
- 從企業級開發到云原生微服務:Spring Boot實戰
- 多媒體應用技術(第2版)
- 微服務架構實戰:基于Spring Boot、Spring Cloud、Docker
- FPGA進階開發與實踐
- Learn Qt 5