Conv3d¶
- 
class torch.nn.Conv3d(in_channels, out_channels, kernel_size, stride=1, padding=0, dilation=1, groups=1, bias=True, padding_mode='zeros')[source]¶
- Applies a 3D convolution over an input signal composed of several input planes. - In the simplest case, the output value of the layer with input size and output can be precisely described as: - where is the valid 3D cross-correlation operator - This module supports TensorFloat32. - stridecontrols the stride for the cross-correlation.
- paddingcontrols the amount of implicit padding on both sides for- paddingnumber of points for each dimension.
- dilationcontrols the spacing between the kernel points; also known as the à trous algorithm. It is harder to describe, but this link has a nice visualization of what- dilationdoes.
- groupscontrols the connections between inputs and outputs.- in_channelsand- out_channelsmust both be divisible by- groups. For example,- At groups=1, all inputs are convolved to all outputs. 
- At groups=2, the operation becomes equivalent to having two conv layers side by side, each seeing half the input channels and producing half the output channels, and both subsequently concatenated. 
- At groups= - in_channels, each input channel is convolved with its own set of filters (of size ).
 
 - The parameters - kernel_size,- stride,- padding,- dilationcan either be:- a single - int– in which case the same value is used for the depth, height and width dimension
- a - tupleof three ints – in which case, the first int is used for the depth dimension, the second int for the height dimension and the third int for the width dimension
 - Note - When groups == in_channels and out_channels == K * in_channels, where K is a positive integer, this operation is also known as a “depthwise convolution”. - In other words, for an input of size , a depthwise convolution with a depthwise multiplier K can be performed with the arguments . - Note - In some circumstances when given tensors on a CUDA device and using CuDNN, this operator may select a nondeterministic algorithm to increase performance. If this is undesirable, you can try to make the operation deterministic (potentially at a performance cost) by setting - torch.backends.cudnn.deterministic = True. See Reproducibility for more information.- Parameters
- in_channels (int) – Number of channels in the input image 
- out_channels (int) – Number of channels produced by the convolution 
- stride (int or tuple, optional) – Stride of the convolution. Default: 1 
- padding (int or tuple, optional) – Zero-padding added to all three sides of the input. Default: 0 
- padding_mode (string, optional) – - 'zeros',- 'reflect',- 'replicate'or- 'circular'. Default:- 'zeros'
- dilation (int or tuple, optional) – Spacing between kernel elements. Default: 1 
- groups (int, optional) – Number of blocked connections from input channels to output channels. Default: 1 
- bias (bool, optional) – If - True, adds a learnable bias to the output. Default:- True
 
 - Shape:
- Input: 
- Output: where 
 
 - Variables
 - Examples: - >>> # With square kernels and equal stride >>> m = nn.Conv3d(16, 33, 3, stride=2) >>> # non-square kernels and unequal stride and with padding >>> m = nn.Conv3d(16, 33, (3, 5, 2), stride=(2, 1, 1), padding=(4, 2, 0)) >>> input = torch.randn(20, 16, 10, 50, 100) >>> output = m(input)