we need to improve FedAvg strategy to support more optimizers
Created by: honshj
FedAvg strategy supports only SGD and cannot be replaced. Other gradients optimizers like Adam, Momentum, AdaGrad can have the same param aggregation and update behaviors as SGD. We need to make them available and pluggable in FedAvg strategy, since a dozen of jobs need them.