Conv Leaky ReLU
ConvLeakyReLU
Bases: Module
Class implements a Convolution followed by a Leaky ReLU activation layer.
Attributes layers (nn.Sequential): Sequential container that holds the Convolution and LeakyReLU layers.
Methods forward(x: torch.Tensor) -> torch.Tensor Passes the input through the Conv1d and LeakyReLU layers.
Source code in notebooks/experiments/conv_leaky_relu.py
9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 |
|
__init__(in_channels, out_channels, kernel_size, padding, leaky_relu_slope=LEAKY_RELU_SLOPE)
Args: in_channels (int): The number of channels in the input data. This could refer to different color channels (like RGB in an image) or different input features in a dataset.
out_channels (int): The number of channels in the output data. This typically corresponds to the number of filters applied on the input.
kernel_size (int): The size of the convolving kernel used in the convolution operation. This is usually an odd integer.
padding (int): The number of zero-padding pixels added on each side of the input data. This is used to control the spatial dimensions of the output data.
leaky_relu_slope (float, default=LEAKY_RELU_SLOPE): The slope of the function for negative values in a Leaky ReLU activation function. This controls the amount of "leakiness" or the degree to which the function allows negative values to pass through.
Source code in notebooks/experiments/conv_leaky_relu.py
20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 |
|
forward(x)
Defines the forward pass of the ConvLeakyReLU.
Args: x (torch.Tensor): The input tensor.
Returns: torch.Tensor: The output tensor after being passed through the Conv1d and LeakyReLU layers.
Source code in notebooks/experiments/conv_leaky_relu.py
47 48 49 50 51 52 53 54 55 56 |
|