Exemple #1
0
 def __init__(self, params, accum=0.1, learning_rate=0.001, l1=0.0, l2=0.0,
              use_locking=False, loss_scale=1.0, weight_decay=0.0):
     super(ProximalAdagrad, self).__init__(learning_rate, params, weight_decay, loss_scale)
     _check_param_value(accum, l1, l2, use_locking, self.cls_name)
     self.accum = self.parameters.clone(prefix="accum", init=accum)
     self.l1 = Tensor(l1, mstype.float32)
     self.l2 = Tensor(l2, mstype.float32)
     self.hyper_map = C.HyperMap()
     self.opt = P.ApplyProximalAdagrad(use_locking=use_locking)
     self.sparse_opt = P.FusedSparseProximalAdagrad(use_locking=use_locking)
Exemple #2
0
 def __init__(self, params, accum=0.1, learning_rate=0.001, l1=0.0, l2=0.0,
              use_locking=False, loss_scale=1.0, weight_decay=0.0):
     super(ProximalAdagrad, self).__init__(learning_rate, params, weight_decay, loss_scale)
     if self.is_group:
         raise RuntimeError(f"The {self.cls_name} optimizer cannot support group setting.")
     _check_param_value(accum, l1, l2, use_locking, self.cls_name)
     self.accum = self.parameters.clone(prefix="accum", init=accum)
     self.l1 = Tensor(l1, mstype.float32)
     self.l2 = Tensor(l2, mstype.float32)
     self.weight_decay = weight_decay
     self.hyper_map = C.HyperMap()
     self.opt = P.ApplyProximalAdagrad(use_locking=use_locking)
     self.sparse_opt = inner.SparseApplyProximalAdagradNoReturn(use_locking=use_locking)
Exemple #3
0
 def __init__(self, var, accum):
     super(ApplyProximalAdagradNet, self).__init__()
     self.apply_proximal_adagrad = P.ApplyProximalAdagrad()
     self.var = Parameter(var, name="var")
     self.accum = Parameter(accum, name='accum')