# Authors # * Peter Plantinga 2020 # Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Modified from speechbrain(https://github.com/speechbrain/speechbrain/blob/develop/speechbrain/nnet/containers.py). import inspect import paddle class Sequential(paddle.nn.LayerDict): """A sequence of modules with potentially inferring shape on construction. If layers are passed with names, these can be referenced with dot notation. Arguments --------- input_shape : iterable A list or tuple of ints or None, representing the expected shape of an input tensor. None represents a variable-length dimension. If no ``input_shape`` is passed, no shape inference will be performed. *layers, **named_layers The inputs are treated as a list of layers to be applied in sequence. The output shape of each layer is used to infer the shape of the following layer. If a tuple is returned, only the shape of the first element is used to determine input shape of the next layer (e.g. RNN returns output, hidden). Example ------- >>> inputs = paddle.rand(10, 40, 50) >>> model = Sequential(input_shape=inputs.shape) >>> model.append(Linear, n_neurons=100, layer_name="layer1") >>> model.append(Linear, n_neurons=200, layer_name="layer2") >>> outputs = model(inputs) >>> outputs.shape paddle.shape([10, 40, 200]) >>> outputs = model.layer1(inputs) >>> outputs.shape paddle.shape([10, 40, 100]) """ def __init__(self, *layers, input_shape=None, **named_layers): super().__init__() # Make sure either layers or input_shape is passed if not layers and input_shape is None and not named_layers: raise ValueError("Must pass either layers or input shape") # Keep track of what layers need "lengths" passed self.length_layers = [] # Replace None dimensions with arbitrary value self.input_shape = input_shape if input_shape and None in input_shape: self.input_shape = list(input_shape) for i, dim in enumerate(self.input_shape): # To reduce size of dummy tensors, use 1 for batch dim if i == 0 and dim is None: dim = 1 # Use 64 as nice round arbitrary value, big enough that # halving this dimension a few times doesn't reach 1 self.input_shape[i] = dim or 256 # Append non-named layers for layer in layers: self.append(layer) # Append named layers for name, layer in named_layers.items(): self.append(layer, layer_name=name) def append(self, layer, *args, layer_name=None, **kwargs): """Add a layer to the list of layers, inferring shape if necessary. Arguments --------- layer : A paddle.nn.Module class or object If the layer is a class, it should accept an argument called ``input_shape`` which will be inferred and passed. If the layer is a module object, it is added as-is. layer_name : str The name of the layer, for reference. If the name is in use, ``_{count}`` will be appended. *args, **kwargs These are passed to the layer if it is constructed. """ # Compute layer_name if layer_name is None: layer_name = str(len(self)) elif layer_name in self: index = 0 while f"{layer_name}_{index}" in self: index += 1 layer_name = f"{layer_name}_{index}" # Check if it needs to be constructed with input shape if self.input_shape: argspec = inspect.getfullargspec(layer) if "input_shape" in argspec.args + argspec.kwonlyargs: input_shape = self.get_output_shape() layer = layer(*args, input_shape=input_shape, **kwargs) # Finally, append the layer. try: self[layer_name] = layer # self.add_module(layer_name, layer) except TypeError: raise ValueError( "Must pass `input_shape` at initialization and use " "modules that take `input_shape` to infer shape when " "using `append()`.") def get_output_shape(self): """Returns expected shape of the output. Computed by passing dummy input constructed with the ``self.input_shape`` attribute. """ with paddle.no_grad(): dummy_input = paddle.zeros(self.input_shape) dummy_output = self(dummy_input) return dummy_output.shape def forward(self, x): """Applies layers in sequence, passing only the first element of tuples. Arguments --------- x : paddle.Tensor The input tensor to run through the network. """ for layer in self.values(): x = layer(x) if isinstance(x, tuple): x = x[0] return x