代码拉取完成,页面将自动刷新
# coding=utf-8
import torch
from torch import nn
from matplotlib import pyplot as plt
# alpha控制x<0的开合角度,alpha=0时,在x<0,y的斜率越小
# mu控制x<0的小范围内的圆滑度
# 当alpha=0时,mu->∞,图像越趋向于ReLU
class SMU(nn.Module):
'''
Implementation of SMU activation.
Shape:
- Input: (N, *) where * means, any number of additional
dimensions
- Output: (N, *), same shape as the input
Parameters:
- alpha: hyper parameter
References:
- See related paper:
https://arxiv.org/abs/2111.04682
Examples:
smu = SMU()
x = torch.Tensor([0.6,-0.3])
x = smu(x)
'''
def __init__(self, alpha = 0.25, mu=100000):
'''
Initialization.
INPUT:
- alpha: hyper parameter
aplha is initialized with zero value by default
'''
super(SMU,self).__init__()
self.alpha = alpha
# initialize mu
self.mu = torch.nn.Parameter(torch.tensor(float(mu)))
def forward(self, x):
return ((1+self.alpha)*x + (1-self.alpha)*x*torch.erf(self.mu*(1-self.alpha)*x))/2
class SMU1(nn.Module):
'''
Implementation of SMU-1 activation.
Shape:
- Input: (N, *) where * means, any number of additional
dimensions
- Output: (N, *), same shape as the input
Parameters:
- alpha: hyper parameter
References:
- See related paper:
https://arxiv.org/abs/2111.04682
Examples:
smu1 = SMU1()
x = torch.Tensor([0.6,-0.3])
x = smu1(x)
'''
def __init__(self, alpha = 0.25):
'''
Initialization.
INPUT:
- alpha: hyper parameter
aplha is initialized with zero value by default
'''
super(SMU1,self).__init__()
self.alpha = alpha
# initialize mu
self.mu = torch.nn.Parameter(torch.tensor(4.352665993287951e-9))
def forward(self, x):
return ((1+self.alpha)*x+torch.sqrt(torch.square(x-self.alpha*x)+torch.square(self.mu)))/2
def test_SMU(x):
smu_activation = SMU()
print(f'SMU = {smu_activation(x)}')
def test_SMU1(x):
smu1_activation=SMU1()
print(f'SMU1 = {smu1_activation(x)}')
def test():
x = torch.Tensor([0.6,-0.3])
test_SMU(x)
test_SMU1(x)
if __name__ == '__main__':
test()
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。