def __init__(self): super(SoftplusNet, self).__init__() self.softplus = P.Softplus()
def __init__(self): super(Mish, self).__init__() self.mul = P.Mul() self.tanh = P.Tanh() self.softplus = P.Softplus()
'desc_inputs': [[1, 3, 4, 4]], 'desc_bprop': [[1, 3, 4, 4]]}), ('ReLU6', { 'block': P.ReLU6(), 'desc_inputs': [[1, 3, 4, 4]], 'desc_bprop': [[1, 3, 4, 4]]}), ('ReLUV2', { 'block': P.ReLUV2(), 'desc_inputs': [[1, 3, 4, 4]], 'desc_bprop': [[1, 3, 4, 4], ([1, 1, 4, 4, 2], {'dtype': np.uint8})]}), ('ReLUGrad', { 'block': G.ReluGrad(), 'desc_inputs': [[1, 3, 4, 4], [1, 3, 4, 4]], 'skip': ['backward']}), ('Softplus', { 'block': P.Softplus(), 'desc_inputs': [[1, 3, 4, 4]], 'desc_bprop': [[1, 3, 4, 4]]}), ('SoftplusGrad', { 'block': G.SoftplusGrad(), 'desc_inputs': [[1, 3, 4, 4], [1, 3, 4, 4]], 'skip': ['backward']}), ('Elu', { 'block': P.Elu(), 'desc_inputs': [[2, 3, 4]], 'desc_bprop': [[2, 3, 4]]}), ('EluGrad', { 'block': G.EluGrad(), 'desc_inputs': [[2, 3, 4], [2, 3, 4]], 'desc_bprop': [[2, 3, 4]], 'skip': ['backward']}),