前往小程序,Get更优阅读体验!
立即前往
首页
学习
活动
专区
工具
TVP
发布
社区首页 >专栏 >Why we need activation function?

Why we need activation function?

作者头像
Steve Wang
发布2019-05-26 15:05:13
3410
发布2019-05-26 15:05:13
举报
文章被收录于专栏:从流域到海域从流域到海域

整理自吴恩达深度学习课程 https://mooc.study.163.com/learn/2001281002?tid=2001392029#/learn/content?type=detail&id=2001702018&cid=2001694026

Why we need activation function ?

Deep networks with many many layers, many many hidden layers and turns out that if you use a linear activation function or alternatively if you don’t have an activation function then no matter how many layers your network has, always doing is computing a linear activation function, so you might as well not have any hidden layers.

If you use the linear function here and sigmoid function here, then this model is no more expressive than standard logistic regression without any hidden layer.

The take-home is that a linear hidden layer is more or less useless,because the composition of two linear function is itself a linear function. So unless you throw a non-linearty in there, then you are not computing more interesting functions even as you good deeper in the network

I, blogger of this post, think that if you are suing linear function, you deep neural network will not get the higher level features we expect.

The hidden units should not use the linear activation functions, they could use ReLU or tanh or leaky ReLU or maybe somethings else. So the one place you might use as linear activation function is usually the output layer. But other that, using a linear activation function in a hidden layer except for some very special circumstances relating to compression. Using linear activation is extremely rare.

本文参与 腾讯云自媒体分享计划,分享自作者个人站点/博客。
原始发表:2018年11月24日,如有侵权请联系 cloudcommunity@tencent.com 删除

本文分享自 作者个人站点/博客 前往查看

如有侵权,请联系 cloudcommunity@tencent.com 删除。

本文参与 腾讯云自媒体分享计划  ,欢迎热爱写作的你一起参与!

评论
登录后参与评论
0 条评论
热度
最新
推荐阅读
目录
  • Why we need activation function ?
领券
问题归档专栏文章快讯文章归档关键词归档开发者手册归档开发者手册 Section 归档