# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserve. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import logging import numpy as np from collections import namedtuple import paddle import paddle.fluid as fluid from .utils.utils import get_paddle_version, remove_model_fn pd_ver = get_paddle_version() if pd_ver == 185: from .layers_old import SuperConv2D, SuperLinear Layer = paddle.fluid.dygraph.Layer DataParallel = paddle.fluid.dygraph.DataParallel else: from .layers import SuperConv2D, SuperLinear Layer = paddle.nn.Layer DataParallel = paddle.DataParallel from .layers_base import BaseBlock, Block from .utils.utils import search_idx from ...common import get_logger from ...core import GraphWrapper, dygraph2program from .get_sub_model import check_search_space, broadcast_search_space from paddle.fluid import core from paddle.fluid.framework import Variable import numbers _logger = get_logger(__name__, level=logging.INFO) __all__ = ['OFA', 'RunConfig', 'DistillConfig'] RunConfig = namedtuple( 'RunConfig', [ # int, batch_size in training, used to get current epoch, default: None 'train_batch_size', # list, the number of epoch of every task in training, default: None 'n_epochs', # list, initial learning rate of every task in traning, NOT used now. Default: None. 'init_learning_rate', # int, total images of train dataset, used to get current epoch, default: None 'total_images', # list, elactic depth of the model in training, default: None 'elastic_depth', # list, the number of sub-network to train per mini-batch data, used to get current epoch, default: None 'dynamic_batch_size', # the shape of weights in the skip_layers will not change in the training, default: None 'skip_layers' ]) RunConfig.__new__.__defaults__ = (None, ) * len(RunConfig._fields) DistillConfig = namedtuple( 'DistillConfig', [ # float, lambda scale of distillation loss, default: None. 'lambda_distill', # instance of model, instance of teacher model, default: None. 'teacher_model', # list(str), name of the layers which need a distillation, default: None. 'mapping_layers', # str, the path of teacher pretrained model, default: None. 'teacher_model_path', # instance of loss layer, the loss function used in distillation, if set to None, use mse_loss default, default: None. 'distill_fn', # str, define which op append between teacher model and student model used in distillation, choice in ['conv', 'linear', None], default: None. 'mapping_op' ]) DistillConfig.__new__.__defaults__ = (None, ) * len(DistillConfig._fields) class OFABase(Layer): def __init__(self, model): super(OFABase, self).__init__() self.model = model self._ofa_layers, self._elastic_task, self._key2name, self._layers = self.get_layers( ) def get_layers(self): ofa_layers = dict() layers = dict() key2name = dict() elastic_task = set() model_to_traverse = self.model._layers if isinstance( self.model, DataParallel) else self.model for name, sublayer in model_to_traverse.named_sublayers(): if isinstance(sublayer, BaseBlock): sublayer.set_supernet(self) if not sublayer.fixed: ofa_layers[name] = sublayer.candidate_config layers[sublayer.key] = sublayer.candidate_config key2name[sublayer.key] = name for k in sublayer.candidate_config.keys(): elastic_task.add(k) return ofa_layers, elastic_task, key2name, layers def forward(self, *inputs, **kwargs): raise NotImplementedError def layers_forward(self, block, *inputs, **kwargs): if getattr(self, 'current_config', None) != None: ### if block is fixed, donnot join key into candidate ### concrete config as parameter in kwargs if block.fixed == False and (self._skip_layers == None or (self._skip_layers != None and self._key2name[block.key] not in self._skip_layers)) and \ (block.fn.weight.name not in self._cannot_changed_layer): assert self._key2name[ block. key] in self.current_config, 'DONNT have {} layer in config.'.format( self._key2name[block.key]) config = self.current_config[self._key2name[block.key]] else: config = dict() config.update(kwargs) else: config = dict() _logger.debug(self.model, config) return block.fn(*inputs, **config) @property def ofa_layers(self): return self._ofa_layers @property def layers(self): return self._layers class OFA(OFABase): """ Convert the training progress to the Once-For-All training progress, a detailed description in the paper: `Once-for-All: Train One Network and Specialize it for Efficient Deployment`_ . This paper propose a training propgress named progressive shrinking (PS), which means we start with training the largest neural network with the maximum kernel size (i.e., 7), depth (i.e., 4), and width (i.e., 6). Next, we progressively fine-tune the network to support smaller sub-networks by gradually adding them into the sampling space (larger sub-networks may also be sampled). Specifically, after training the largest network, we first support elastic kernel size which can choose from {3, 5, 7} at each layer, while the depth and width remain the maximum values. Then, we support elastic depth and elastic width sequentially. Parameters: model(paddle.nn.Layer): instance of model. run_config(paddleslim.ofa.RunConfig, optional): config in ofa training, can reference `<>`_ . Default: None. distill_config(paddleslim.ofa.DistillConfig, optional): config of distilltion in ofa training, can reference `<>`_. Default: None. elastic_order(list, optional): define the training order, if it set to None, use the default order in the paper. Default: None. train_full(bool, optional): whether to train the largest sub-network only. Default: False. Examples: .. code-block:: python from paddle.vision.models import mobilenet_v1 from paddleslim.nas.ofa import OFA from paddleslim.nas.ofa.convert_super import Convert, supernet model = mobilenet_v1() sp_net_config = supernet(kernel_size=(3, 5, 7), expand_ratio=[1, 2, 4]) sp_model = Convert(sp_net_config).convert(model) ofa_model = OFA(sp_model) """ def __init__(self, model, run_config=None, distill_config=None, elastic_order=None, train_full=False): super(OFA, self).__init__(model) self.net_config = None self.run_config = run_config self.distill_config = distill_config self.elastic_order = elastic_order self.train_full = train_full self.iter = 0 self.dynamic_iter = 0 self.manual_set_task = False self.task_idx = 0 self._add_teacher = False self.netAs_param = [] self._mapping_layers = None self._build_ss = False self._broadcast = False self._skip_layers = None self._cannot_changed_layer = [] ### if elastic_order is none, use default order if self.elastic_order is not None: assert isinstance(self.elastic_order, list), 'elastic_order must be a list' if getattr(self.run_config, 'elastic_depth', None) != None: depth_list = list(set(self.run_config.elastic_depth)) depth_list.sort() self._ofa_layers['depth'] = depth_list self._layers['depth'] = depth_list if self.elastic_order is None: self.elastic_order = [] # zero, elastic resulotion, write in demo # first, elastic kernel size if 'kernel_size' in self._elastic_task: self.elastic_order.append('kernel_size') # second, elastic depth, such as: list(2, 3, 4) if getattr(self.run_config, 'elastic_depth', None) != None: depth_list = list(set(self.run_config.elastic_depth)) depth_list.sort() self._ofa_layers['depth'] = depth_list self._layers['depth'] = depth_list self.elastic_order.append('depth') # final, elastic width if 'expand_ratio' in self._elastic_task: self.elastic_order.append('width') if 'channel' in self._elastic_task and 'width' not in self.elastic_order: self.elastic_order.append('width') if getattr(self.run_config, 'n_epochs', None) != None: assert len(self.run_config.n_epochs) == len(self.elastic_order) for idx in range(len(run_config.n_epochs)): assert isinstance( run_config.n_epochs[idx], list), "each candidate in n_epochs must be list" if self.run_config.dynamic_batch_size != None: assert len(self.run_config.n_epochs) == len( self.run_config.dynamic_batch_size) if self.run_config.init_learning_rate != None: assert len(self.run_config.n_epochs) == len( self.run_config.init_learning_rate) for idx in range(len(run_config.n_epochs)): assert isinstance( run_config.init_learning_rate[idx], list ), "each candidate in init_learning_rate must be list" ### remove skip layers in search space if self.run_config != None and getattr(self.run_config, 'skip_layers', None) != None: self._skip_layers = self.run_config.skip_layers ### ================= add distill prepare ====================== if self.distill_config != None: self._add_teacher = True self._prepare_distill() self.model.train() def _prepare_distill(self): if self.distill_config.teacher_model == None: _logger.error( 'If you want to add distill, please input instance of teacher model' ) ### instance model by user can input super-param easily. assert isinstance(self.distill_config.teacher_model, Layer) # load teacher parameter if self.distill_config.teacher_model_path != None: param_state_dict, _ = paddle.load_dygraph( self.distill_config.teacher_model_path) self.distill_config.teacher_model.set_dict(param_state_dict) self.ofa_teacher_model = OFABase(self.distill_config.teacher_model) self.ofa_teacher_model.model.eval() # add hook if mapping layers is not None # if mapping layer is None, return the output of the teacher model, # if mapping layer is NOT None, add hook and compute distill loss about mapping layers. mapping_layers = getattr(self.distill_config, 'mapping_layers', None) if mapping_layers != None: if isinstance(self.model, DataParallel): for idx, name in enumerate(mapping_layers): if name[:7] != '_layers': mapping_layers[idx] = '_layers.' + name self._mapping_layers = mapping_layers self.netAs = [] for name, sublayer in self.model.named_sublayers(): if name in self._mapping_layers: if self.distill_config.mapping_op != None: if self.distill_config.mapping_op.lower() == 'conv2d': netA = SuperConv2D( getattr(sublayer, '_num_filters', sublayer._out_channels), getattr(sublayer, '_num_filters', sublayer._out_channels), 1) elif self.distill_config.mapping_op.lower() == 'linear': netA = SuperLinear( getattr(sublayer, '_output_dim', sublayer._out_features), getattr(sublayer, '_output_dim', sublayer._out_features)) else: raise NotImplementedError( "Not Support Op: {}".format( self.distill_config.mapping_op.lower())) else: netA = None if netA != None: self.netAs_param.extend(netA.parameters()) self.netAs.append(netA) def _reset_hook_before_forward(self): self.Tacts, self.Sacts = {}, {} self.hooks = [] if self._mapping_layers != None: def get_activation(mem, name): def get_output_hook(layer, input, output): mem[name] = output return get_output_hook def add_hook(net, mem, mapping_layers): for idx, (n, m) in enumerate(net.named_sublayers()): if n in mapping_layers: self.hooks.append( m.register_forward_post_hook( get_activation(mem, n))) add_hook(self.model, self.Sacts, self._mapping_layers) add_hook(self.ofa_teacher_model.model, self.Tacts, self._mapping_layers) def _remove_hook_after_forward(self): for hook in self.hooks: hook.remove() def _compute_epochs(self): if getattr(self, 'epoch', None) == None: assert self.run_config.total_images is not None, \ "if not use set_epoch() to set epoch, please set total_images in run_config." assert self.run_config.train_batch_size is not None, \ "if not use set_epoch() to set epoch, please set train_batch_size in run_config." assert self.run_config.n_epochs is not None, \ "if not use set_epoch() to set epoch, please set n_epochs in run_config." self.iter_per_epochs = self.run_config.total_images // self.run_config.train_batch_size epoch = self.iter // self.iter_per_epochs else: epoch = self.epoch return epoch def _sample_from_nestdict(self, cands, sample_type, task, phase): sample_cands = dict() for k, v in cands.items(): if isinstance(v, dict): sample_cands[k] = self._sample_from_nestdict( v, sample_type=sample_type, task=task, phase=phase) elif isinstance(v, list) or isinstance(v, set) or isinstance(v, tuple): if sample_type == 'largest': sample_cands[k] = v[-1] elif sample_type == 'smallest': sample_cands[k] = v[0] else: if k not in task: # sort and deduplication in candidate_config # fixed candidate not in task_list sample_cands[k] = v[-1] else: # phase == None -> all candidate; phase == number, append small candidate in each phase # phase only affect last task in current task_list if phase != None and k == task[-1]: start = -(phase + 2) else: start = 0 sample_cands[k] = np.random.choice(v[start:]) return sample_cands def _sample_config(self, task, sample_type='random', phase=None): config = self._sample_from_nestdict( self._ofa_layers, sample_type=sample_type, task=task, phase=phase) return config def set_task(self, task, phase=None): """ set task in the ofa training progress. Parameters: task(list(str)|str): spectial task in training progress. phase(int, optional): the search space is gradually increased, use this parameter to spectial the phase in current task, if set to None, means use the whole search space in training progress. Default: None. Examples: .. code-block:: python ofa_model.set_task('width') """ self.manual_set_task = True self.task = task self.phase = phase def set_epoch(self, epoch): """ set epoch in the ofa training progress. Parameters: epoch(int): spectial epoch in training progress. Examples: .. code-block:: python ofa_model.set_epoch(3) """ self.epoch = epoch def _progressive_shrinking(self): epoch = self._compute_epochs() phase_idx = None if len(self.elastic_order) != 1: assert self.run_config.n_epochs is not None, \ "if not use set_task() to set current task, please set n_epochs in run_config " \ "for to compute which task in this epoch." self.task_idx, phase_idx = search_idx(epoch, self.run_config.n_epochs) self.task = self.elastic_order[:self.task_idx + 1] if 'width' in self.task: ### change width in task to concrete config self.task.remove('width') if 'expand_ratio' in self._elastic_task: self.task.append('expand_ratio') if 'channel' in self._elastic_task: self.task.append('channel') return self._sample_config(task=self.task, phase=phase_idx) def calc_distill_loss(self): """ Calculate distill loss if there are distillation. Examples: .. code-block:: python dis_loss = ofa_model.calc_distill_loss() """ losses = [] assert len(self.netAs) > 0 for i, netA in enumerate(self.netAs): n = self.distill_config.mapping_layers[i] ### add for elastic depth if n not in self.Sacts.keys(): continue Tact = self.Tacts[n] Sact = self.Sacts[n] if isinstance(netA, SuperConv2D): Sact = netA( Sact, channel=getattr(netA, '_num_filters', netA._out_channels)) elif isinstance(netA, SuperLinear): Sact = netA( Sact, channel=getattr(netA, '_output_dim', netA._out_features)) else: Sact = Sact Sact = Sact[0] if isinstance(Sact, tuple) else Sact Tact = Tact[0] if isinstance(Tact, tuple) else Tact if self.distill_config.distill_fn == None: loss = fluid.layers.mse_loss(Sact, Tact.detach()) else: loss = distill_fn(Sact, Tact.detach()) losses.append(loss) if self.distill_config.lambda_distill != None: return sum(losses) * self.distill_config.lambda_distill return sum(losses) ### TODO: complete it def search(self, eval_func, condition): pass def _get_model_pruned_weight(self): pruned_param = {} for l_name, sublayer in self.model.named_sublayers(): if getattr(sublayer, 'cur_config', None) == None: continue assert 'prune_dim' in sublayer.cur_config, 'The laycer {} do not have prune_dim in cur_config.'.format( l_name) prune_shape = sublayer.cur_config['prune_dim'] for p_name, param in sublayer.named_parameters( include_sublayers=False): origin_param = param.value().get_tensor() param = np.array(origin_param).astype("float32") name = l_name + '.' + p_name if isinstance(prune_shape, list): if len(param.shape) == 4: pruned_param[name] = param[:prune_shape[0], : prune_shape[1], :, :] elif len(param.shape) == 2: pruned_param[name] = param[:prune_shape[0], : prune_shape[1]] else: if isinstance(sublayer, SuperLinear): pruned_param[name] = param[:prune_shape[1]] else: pruned_param[name] = param[:prune_shape[0]] else: pruned_param[name] = param[:prune_shape] return pruned_param def export(self, config, input_shapes, input_dtypes, origin_model=None, load_weights_from_supernet=True): """ Export the weights according origin model and sub model config. Parameters: origin_model(paddle.nn.Layer): the instance of original model. config(dict): the config of sub model, can get by OFA.get_current_config() or some special config, such as paddleslim.nas.ofa.utils.dynabert_config(width_mult). input_shapes(list|list(list)): the shape of all inputs. input_dtypes(list): the dtype of all inputs. load_weights_from_supernet(bool, optional): whether to load weights from SuperNet. Default: False. Examples: .. code-block:: python from paddle.vision.models import mobilenet_v1 origin_model = mobilenet_v1() config = {'conv2d_0': {'expand_ratio': 2}, 'conv2d_1': {'expand_ratio': 2}} origin_model = ofa_model.export(origin_model, config, input_shapes=[1, 3, 28, 28], input_dtypes=['float32']) """ self.set_net_config(config) self.model.eval() def build_input(input_size, dtypes): if isinstance(input_size, list) and all( isinstance(i, numbers.Number) for i in input_size): if isinstance(dtypes, list): dtype = dtypes[0] else: dtype = dtypes return paddle.cast(paddle.rand(list(input_size)), dtype) if isinstance(input_size, dict): inputs = {} if isinstance(dtypes, list): dtype = dtypes[0] else: dtype = dtypes for key, value in input_size.items(): inputs[key] = paddle.cast(paddle.rand(list(value)), dtype) return inputs if isinstance(input_size, list): return [ build_input(i, dtype) for i, dtype in zip(input_size, dtypes) ] data = build_input(input_shapes, input_dtypes) if isinstance(data, list): self.forward(*data) else: self.forward(data) super_model_state_dict = None if load_weights_from_supernet and origin_model != None: super_model_state_dict = remove_model_fn(origin_model, self.model.state_dict()) if origin_model == None: origin_model = self.model origin_model = origin_model._layers if isinstance( origin_model, DataParallel) else origin_model _logger.info("Start to get pruned params, please wait...") pruned_param = self._get_model_pruned_weight() pruned_state_dict = remove_model_fn(origin_model, pruned_param) _logger.info("Start to get pruned model, please wait...") for l_name, sublayer in origin_model.named_sublayers(): for p_name, param in sublayer.named_parameters( include_sublayers=False): name = l_name + '.' + p_name t_value = param.value().get_tensor() if name in pruned_state_dict: p = t_value._place() if p.is_cpu_place(): place = core.CPUPlace() elif p.is_cuda_pinned_place(): place = core.CUDAPinnedPlace() else: place = core.CUDAPlace(p.gpu_device_id()) t_value.set(pruned_state_dict[name], place) if super_model_state_dict != None and len(super_model_state_dict) != 0: for k, v in super_model_state_dict.items(): setattr(origin_model, k, v) return origin_model @property def get_current_config(self): return self.current_config def set_net_config(self, net_config): """ Set the config of the special sub-network to be trained. Parameters: net_config(dict): special the config of sug-network. Examples: .. code-block:: python config = {'conv2d_0': {'expand_ratio': 2}, 'conv2d_1': {'expand_ratio': 2}} ofa_model.set_net_config(config) """ self.net_config = net_config def _find_ele(self, inp, targets): def _roll_eles(target_list, types=(list, set, tuple)): if isinstance(target_list, types): for targ in target_list: for v in _roll_eles(targ, types): yield v else: yield target_list if inp in list(_roll_eles(targets)): return True else: return False def _clear_width(self, key): if 'expand_ratio' in self._ofa_layers[key]: self._ofa_layers[key].pop('expand_ratio') elif 'channel' in self._ofa_layers[key]: self._ofa_layers[key].pop('channel') if len(self._ofa_layers[key]) == 0: self._ofa_layers.pop(key) def _clear_search_space(self, *inputs, **kwargs): """ find shortcut in model, and clear up the search space """ input_shapes = [] input_dtypes = [] for n in inputs: if isinstance(n, Variable): input_shapes.append(n) input_dtypes.append(n.numpy().dtype) for key, val in kwargs.items(): if isinstance(val, Variable): input_shapes.append(val) input_dtypes.append(val.numpy().dtype) elif isinstance(val, dict): input_shape = {} input_dtype = {} for k, v in val.items(): input_shape[k] = v input_dtype[k] = v.numpy().dtype input_shapes.append(input_shape) input_dtypes.append(input_dtype) else: _logger.error( "Cannot figure out the type of inputs! Right now, the type of inputs can be only Variable or dict." ) ### find shortcut block using static model model_to_traverse = self.model._layers if isinstance( self.model, DataParallel) else self.model _st_prog = dygraph2program( model_to_traverse, inputs=input_shapes, dtypes=input_dtypes) self._same_ss, depthwise_conv, fixed_by_input, output_conv = check_search_space( GraphWrapper(_st_prog)) self._cannot_changed_layer = output_conv if self._same_ss != None: self._param2key = {} self._broadcast = True ### the name of sublayer is the key in search space ### param.name is the name in self._same_ss for name, sublayer in model_to_traverse.named_sublayers(): if isinstance(sublayer, BaseBlock): for param in sublayer.parameters(): if self._find_ele(param.name, self._same_ss): self._param2key[param.name] = name ### double clear same search space to avoid outputs weights in same ss. tmp_same_ss = [] for ss in self._same_ss: per_ss = [] for key in ss: if key not in self._param2key.keys(): continue ### if skip_layers and same ss both have same layer, ### the layers related to this layer need to add to skip_layers if self._skip_layers != None and self._param2key[ key] in self._skip_layers: self._skip_layers += [self._param2key[sk] for sk in ss] per_ss = [] break if self._param2key[key] in self._ofa_layers.keys() and \ ('expand_ratio' in self._ofa_layers[self._param2key[key]] or \ 'channel' in self._ofa_layers[self._param2key[key]]): per_ss.append(key) else: _logger.info("{} not in ss".format(key)) if len(per_ss) != 0: tmp_same_ss.append(per_ss) self._same_ss = tmp_same_ss ### if fixed_by_input layer in a same ss, ### layers in this same ss should all be fixed tmp_fixed_by_input = [] for ss in self._same_ss: for key in fixed_by_input: if key in ss: tmp_fixed_by_input += ss fixed_by_input += tmp_fixed_by_input ### clear layer in ofa_layers set by skip layers if self._skip_layers != None: for skip_layer in self._skip_layers: if skip_layer in self._ofa_layers.keys(): self._ofa_layers.pop(skip_layer) for per_ss in self._same_ss: for ss in per_ss[1:]: self._clear_width(self._param2key[ss]) self._cannot_changed_layer = sorted( set(output_conv + fixed_by_input + depthwise_conv)) ### clear depthwise convs from search space because of its output channel cannot change ### clear output convs from search space because of model output shape cannot change ### clear convs that operate with fixed input for name, sublayer in model_to_traverse.named_sublayers(): if isinstance(sublayer, BaseBlock): for param in sublayer.parameters(): if param.name in self._cannot_changed_layer and name in self._ofa_layers.keys( ): self._clear_width(name) def forward(self, *inputs, **kwargs): # ===================== teacher process ===================== teacher_output = None if self._add_teacher: self._reset_hook_before_forward() teacher_output = self.ofa_teacher_model.model.forward(*inputs, **kwargs) # ============================================================ # ==================== student process ===================== if not self._build_ss and self.net_config == None: self._clear_search_space(*inputs, **kwargs) self._build_ss = True if getattr(self.run_config, 'dynamic_batch_size', None) != None: self.dynamic_iter += 1 if self.dynamic_iter == self.run_config.dynamic_batch_size[ self.task_idx]: self.iter += 1 self.dynamic_iter = 0 if self.net_config == None: if self.train_full == True: self.current_config = self._sample_config( task=None, sample_type='largest') else: if self.manual_set_task == False: self.current_config = self._progressive_shrinking() else: self.current_config = self._sample_config( self.task, phase=self.phase) else: self.current_config = self.net_config _logger.debug("Current config is {}".format(self.current_config)) if 'depth' in self.current_config: kwargs['depth'] = self.current_config['depth'] if self._broadcast: broadcast_search_space(self._same_ss, self._param2key, self.current_config) student_output = self.model.forward(*inputs, **kwargs) if self._add_teacher: self._remove_hook_after_forward() return student_output, teacher_output