官术网_书友最值得收藏!

Learning the optimizer

In this method, we try to learn the optimizer. How do we generally optimize our neural network? We optimize our neural network by training on a large dataset and minimize the loss using gradient descent. But in the few-shot learning setting, gradient descent fails as we will have a smaller dataset. So, in this case, we will learn the optimizer itself. We will have two networks: a base network that actually tries to learn and a meta network that optimizes the base network. We will explore how exactly this works in the upcoming sections.

主站蜘蛛池模板: 广昌县| 于田县| 吴桥县| 斗六市| 涟水县| 夹江县| 长白| 新疆| 盐池县| 增城市| 湟源县| 阜康市| 罗甸县| 方山县| 萍乡市| 宿州市| 乡宁县| 横山县| 象州县| 汝城县| 阳山县| 洛阳市| 扎鲁特旗| 乌兰察布市| 漳州市| 宁德市| 西贡区| 汝州市| 依安县| 嘉义县| 广元市| 天台县| 张掖市| 江口县| 孟州市| 莎车县| 通河县| 万山特区| 普陀区| 贵定县| 育儿|