img doc360img-src='http://image109.360doc.com/DownloadImg/2023/01/2800/259633446_38_20230128120340993.png' data-ratio='0.8819672131147541' data-type='png' data-w='1220' data-index='38' src='https://www.xkablog.com/uploads/202412/04/58876b0a855bf7f3.gif' _width='-30px' alt='图片'>
img doc360img-src='http://image109.360doc.com/DownloadImg/2023/01/2800/259633446_39_20230128120341227.png' data-ratio='0.7167530224525043' data-type='png' data-w='1158' data-index='39' src='https://www.xkablog.com/uploads/202412/04/58876b0a855bf7f3.gif' _width='-30px' alt='图片'>
Sigmoid和Tanh激活函数的导数在正负饱和区的梯度都会接近于0,这会造成梯度消失,而ReLU函数大于0部分都为常数保持梯度不衰减,不会产生梯度消失现象。