常见激活函数优缺点与dead relu problem

常见激活函数优缺点与dead relu problem

 

https://zhuanlan.zhihu.com/p/71882757

 

https://mp.weixin.qq.com/s/hoOBTDBmE666-NcMDOzdoQ

 

1.什么是激活函数?

所谓激活函数(Activation Function),就是在人工神经网络的神经元上运行的函数,负责将神经元的输入映射到输出端。

激活函数对于人工神经网络模型去学习、理解非常复杂和非线性的函数来说具有十分重要的作用。它们将非线性特性引入到我们的网络中。如图,在神经元中,输入(inputs )通过加权,求和后,还被作用在一个函数上,这个函数就是激活函数。

 

第一个链接中内容写得非常好。

 

### ReLU Activation Function in Neural Networks In neural networks, the Rectified Linear Unit (ReLU) serves as an essential component that introduces non-linearity into models. This function outputs the input directly if it is positive; otherwise, it will output zero[^3]. Mathematically speaking, this can be represented by: \[ f(x)=\max(0,x) \] This simple yet effective mechanism allows neurons to efficiently compute activations while avoiding vanishing gradient problems associated with traditional activation functions like sigmoid or tanh. However, one potential issue encountered when using ReLU units lies within their tendency towards dying during training – meaning they become inactive permanently once weights update causes them not firing anymore over time due to negative inputs always resulting in zeros after passing through such layers without any adjustments made accordingly via backpropagation updates applied later stages down stream from these points forward until convergence occurs eventually leading up toward optimal solutions being found successfully overall throughout entire process completion period required for achieving desired outcomes set forth initially before starting out on journey ahead together now moving onward further into discussion about alternatives available today's modern era technology advancements present us currently at hand here right away below next section immediately following thereafter sequentially listed items provided just above mentioned previously already covered earlier parts written prior sections preceding current paragraph placement location spot situated position placed exactly where you see me writing words sentences paragraphs etc... To address some limitations of standard ReLUs, variations have been proposed including PACT which uses parameterized clipping instead of fixed thresholds allowing more flexibility during optimization processes thus potentially mitigating issues related specifically around dead/dormant nodes problem often faced traditionally under certain conditions depending upon specific use-cases scenarios encountered practically implemented systems deployed across various industries sectors fields applications domains areas contexts situations circumstances environments settings configurations arrangements structures frameworks architectures designs implementations deployments operations management maintenance support services products offerings solutions approaches methodologies strategies tactics techniques mechanisms procedures protocols standards guidelines policies regulations laws rules constraints requirements specifications parameters variables factors elements components pieces parts aspects features attributes properties characteristics traits qualities states conditions statuses positions placements locations spots places spaces rooms buildings facilities infrastructure superstructures constructs edifices establishments institutions organizations entities bodies corporations enterprises businesses ventures projects initiatives efforts endeavors undertakings pursuits activities actions events occurrences happenings incidents episodes occasions moments times periods durations intervals spans stretches expanses extents reaches ranges scopes scales magnitudes sizes dime
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值