*Memos:
- My post explains Transposed Convolutional Layer.
- My post explains ConvTranspose2d().
- My post explains ConvTranspose3d().
- My post explains manual_seed().
- My post explains requires_grad.
ConvTranspose1d() can get the 2D or 3D tensor of the one or more elements computed by 1D transposed convolution from the 2D or 3D tensor of one or more elements as shown below:
*Memos:
- The 1st argument for initialization is
in_channels(Required-Type:float). *It must be1 <= x. - The 2nd argument for initialization is
out_channels(Required-Type:float). *It must be1 <= x. - The 3rd argument for initialization is
kernel_size(Required-Type:intortupleorlistofint). *It must be1 <= x. - The 4th argument for initialization is
stride(Optional-Default:1-Type:intortupleorlistofint). *It must be1 <= x. - The 5th argument for initialization is
padding(Optional-Default:0-Type:intortupleorlistofint). *It must be0 <= x. - The 6th argument for initialization is
output_padding(Optional-Default:0-Type:int,tupleorlistofint). *It must be0 <= x. - The 7th argument for initialization is
groups(Optional-Default:1-Type:int). *It must be1 <= x. - The 8th argument for initialization is
bias(Optional-Default:True-Type:bool). *My post explainsbiasargument. - The 9th argument for initialization is
dilation(Optional-Default:1-Type:intortupleorlistofint). *It must be1 <= x. - The 10th argument for initialization is
padding_mode(Optional-Default:'zeros'-Type:str). *Only'zeros'can be selected. - The 11th argument for initialization is
device(Optional-Default:None-Type:str,intor device()): *Memos:- If it's
None, get_default_device() is used. *My post explainsget_default_device()and set_default_device(). -
device=can be omitted. -
My post explains
deviceargument.
- If it's
- The 12th argument for initialization is
dtype(Optional-Default:None-Type:dtype): *Memos:- If it's
None, get_default_dtype() is used. *My post explainsget_default_dtype()and set_default_dtype(). -
dtype=can be omitted. -
My post explains
dtypeargument.
- If it's
- The 1st argument is
input(Required-Type:tensoroffloatorcomplex): *Memos:- It must be the 2D or 3D tensor of one or more elements.
- The number of the elements of the 2nd deepest dimension must be same as
in_channels. - Its
deviceanddtypemust be same asConvTranspose1d()'s. -
complexmust be set todtypeofConvTranspose1d()to use acomplextensor. - The tensor's
requires_gradwhich isFalseby default is set toTruebyConvTranspose1d().
-
convtran1d1.deviceandconvtran1d1.dtypedon't work.
import torch
from torch import nn
tensor1 = torch.tensor([[8., -3., 0., 1., 5., -2.]])
tensor1.requires_grad
# False
torch.manual_seed(42)
convtran1d1 = nn.ConvTranspose1d(in_channels=1, out_channels=3, kernel_size=1)
tensor2 = convtran1d1(input=tensor1)
tensor2
# tensor([[4.0616, -0.7939, 0.5304, 0.9718, 2.7374, -0.3525],
# [3.7071, -1.5641, -0.1265, 0.3527, 2.2695, -1.0849],
# [-0.9656, 0.5223, 0.1165, -0.0188, -0.5598, 0.3870]],
# grad_fn=<SqueezeBackward1>)
tensor2.requires_grad
# True
convtran1d1
# ConvTranspose1d(1, 3, kernel_size=(1,), stride=(1,))
convtran1d1.in_channels
# 1
convtran1d1.out_channels
# 3
convtran1d1.kernel_size
# (1,)
convtran1d1.stride
# (1,)
convtran1d1.padding
# (0,)
convtran1d1.output_padding
# (0,)
convtran1d1.groups
# 1
convtran1d1.bias
# Parameter containing:
# tensor([0.5304, -0.1265, 0.1165], requires_grad=True)
convtran1d1.dilation
# (1,)
convtran1d1.padding_mode
# 'zeros'
convtran1d1.weight
# Parameter containing:
# tensor([[[0.4414], [0.4792], [-0.1353]]], requires_grad=True)
torch.manual_seed(42)
convtran1d2 = nn.ConvTranspose1d(in_channels=3, out_channels=1, kernel_size=1)
convtran1d2(input=tensor2)
# tensor([[7.3270, -1.1089, 1.1918, 1.9587, 5.0263, -0.3420]],
# grad_fn=<SqueezeBackward1>)
torch.manual_seed(42)
convtran1d = nn.ConvTranspose1d(in_channels=1, out_channels=3,
kernel_size=1, stride=1, padding=0, output_padding=0,
groups=1, bias=True, dilation=1, padding_mode='zeros',
device=None, dtype=None)
convtran1d(input=tensor1)
# tensor([[4.0616, -0.7939, 0.5304, 0.9718, 2.7374, -0.3525],
# [3.7071, -1.5641, -0.1265, 0.3527, 2.2695, -1.0849],
# [-0.9656, 0.5223, 0.1165, -0.0188, -0.5598, 0.3870]],
# grad_fn=<SqueezeBackward1>)
my_tensor = torch.tensor([[8., -3., 0.],
[1., 5., -2.]])
torch.manual_seed(42)
convtran1d = nn.ConvTranspose1d(in_channels=2, out_channels=3,
kernel_size=1)
convtran1d(input=my_tensor)
# tensor([[3.7805, 1.0465, -1.3418],
# [4.0462, -1.7310, 0.5921],
# [-0.4566, 1.4973, 0.2760]], grad_fn=<SqueezeBackward1>)
my_tensor = torch.tensor([[8.], [-3.], [0.],
[1.], [5.], [-2.]])
torch.manual_seed(42)
convtran1d = nn.ConvTranspose1d(in_channels=6, out_channels=3, kernel_size=1)
convtran1d(input=my_tensor)
# tensor([[3.5429], [4.3629], [-0.3682]], grad_fn=<SqueezeBackward1>)
my_tensor = torch.tensor([[[8.], [-3.], [0.]],
[[1.], [5.], [-2.]]])
torch.manual_seed(42)
convtran1d = nn.ConvTranspose1d(in_channels=3, out_channels=3, kernel_size=1)
convtran1d(input=my_tensor)
# tensor([[[1.5166], [4.7150], [-1.3235]],
# [[3.2318], [-0.3296], [-0.4626]]],
# grad_fn=<ConvolutionBackward0>)
my_tensor = torch.tensor([[[8.+0.j], [-3.+0.j], [0.+0.j]],
[[1.+0.j], [5.+0.j], [-2.+0.j]]])
torch.manual_seed(42)
convtran1d = nn.ConvTranspose1d(in_channels=3, out_channels=3,
kernel_size=1, dtype=torch.complex64)
convtran1d(input=my_tensor)
# tensor([[[4.1050+2.9636j], [-2.8749+5.4459j], [-2.7520+0.9909j]],
# [[-2.0867+2.1653j], [1.5867-1.4921j], [1.2580+0.8691j]]],
# grad_fn=<AddBackward0>)
Top comments (0)