FS-TFP/federatedscope/contrib/optimizer/example.py

18 lines
435 B
Python

from federatedscope.register import register_optimizer
def call_my_optimizer(model, type, lr, **kwargs):
try:
import torch.optim as optim
except ImportError:
optim = None
optimizer = None
if type == 'myoptimizer':
if optim is not None:
optimizer = optim.Adam(model.parameters(), lr=lr, **kwargs)
return optimizer
register_optimizer('myoptimizer', call_my_optimizer)