DEV Community

Super Kai (Kazuya Ito)
Super Kai (Kazuya Ito)

Posted on • Edited on

Set `bias` argument in PyTorch

Buy Me a Coffee

*Memos:

You can set out as shown below:

*Memos:

Linear(). *My post explains Linear():

import torch
from torch import nn

my_tensor = torch.tensor([8., -3., 0., 1., 5., -2.])

torch.manual_seed(42)

linear = nn.Linear(in_features=6, out_features=3, bias=True)

linear.bias
# Parameter containing:
# tensor([-0.1906, 0.1041, -0.1881], requires_grad=True)

linear(input=my_tensor)
# tensor([1.0529, -0.8833,  3.4542], grad_fn=<ViewBackward0>)

torch.manual_seed(42)

linear = nn.Linear(in_features=6, out_features=3, bias=False)

print(linear.bias)
# None

linear(input=my_tensor)
# tensor([1.2434, -0.9874, 3.6423], grad_fn=<SqueezeBackward4>)
Enter fullscreen mode Exit fullscreen mode

Conv1d(). *My post explains Conv1d():

import torch
from torch import nn

my_tensor = torch.tensor([[8., -3., 0., 1., 5., -2.]])

torch.manual_seed(42)

conv1d = nn.Conv1d(in_channels=1, out_channels=3,
                   kernel_size=1, bias=True)
conv1d.bias
# Parameter containing:
# tensor([0.9186, -0.2191, 0.2018], requires_grad=True)

conv1d(input=my_tensor)
# tensor([[7.0349, -1.3750, 0.9186, 1.6831, 4.7413, -0.6105],
#         [6.4210, -2.7091, -0.2191, 0.6109, 3.9309, -1.8791],
#         [-1.6724, 0.9046, 0.2018, -0.0325, -0.9696, 0.6703]],
#        grad_fn=<SqueezeBackward1>)

torch.manual_seed(42)

conv1d = nn.Conv1d(in_channels=1, out_channels=3,
                   kernel_size=1, bias=False)
print(conv1d.bias)
# None

conv1d(input=my_tensor)
# tensor([[6.1163, -2.2936, 0.0000, 0.7645, 3.8227, -1.5291],
#         [6.6401, -2.4900, 0.0000, 0.8300, 4.1500, -1.6600],
#         [-1.8742, 0.7028, -0.0000, -0.2343, -1.1714, 0.4685]],
#        grad_fn=<SqueezeBackward1>)
Enter fullscreen mode Exit fullscreen mode

ConvTranspose1d(). *My post explains ConvTranspose1d():

import torch
from torch import nn

my_tensor = torch.tensor([[8., -3., 0., 1., 5., -2.]])

torch.manual_seed(42)

convtran1d = nn.ConvTranspose1d(in_channels=1, out_channels=3,
                                kernel_size=1, bias=True)
convtran1d.bias
# Parameter containing:
# tensor([0.5304, -0.1265, 0.1165], requires_grad=True)

convtran1d(input=my_tensor)
# tensor([[4.0616, -0.7939, 0.5304, 0.9718, 2.7374, -0.3525],
#         [3.7071, -1.5641, -0.1265, 0.3527, 2.2695, -1.0849],
#         [-0.9656, 0.5223, 0.1165, -0.0188, -0.5598, 0.3870]],
#        grad_fn=<SqueezeBackward1>)

torch.manual_seed(42)

convtran1d = nn.ConvTranspose1d(in_channels=1, out_channels=3,
                                kernel_size=1, bias=False)
print(convtran1d.bias)
# None

convtran1d(input=my_tensor)
# tensor([[3.5313, -1.3242, 0.0000, 0.4414, 2.2070, -0.8828],
#         [3.8336, -1.4376, 0.0000, 0.4792, 2.3960, -0.9584],
#         [-1.0821, 0.4058, 0.0000, -0.1353, -0.6763, 0.2705]],
#        grad_fn=<SqueezeBackward1>)
Enter fullscreen mode Exit fullscreen mode

LayerNorm(). *My post explains LayerNorm():

import torch
from torch import nn

my_tensor = torch.tensor([8., -3., 0., 1., 5., -2.])

torch.manual_seed(42)

layernorm = nn.LayerNorm(normalized_shape=6, bias=True)

layernorm.bias
# Parameter containing:
# tensor([0., 0., 0., 0., 0., 0.], requires_grad=True)

layernorm(input=my_tensor)
# tensor([1.6830, -1.1651, -0.3884, -0.1295, 0.9062, -0.9062],
#        grad_fn=<NativeLayerNormBackward0>)

torch.manual_seed(42)

layernorm = nn.LayerNorm(normalized_shape=6, bias=False)

print(layernorm.bias)
# None

layernorm(input=my_tensor)
# tensor([1.6830, -1.1651, -0.3884, -0.1295, 0.9062, -0.9062],
#        grad_fn=<NativeLayerNormBackward0>)
Enter fullscreen mode Exit fullscreen mode

Postmark Image

Speedy emails, satisfied customers

Are delayed transactional emails costing you user satisfaction? Postmark delivers your emails almost instantly, keeping your customers happy and connected.

Sign up

Top comments (0)

Sentry image

See why 4M developers consider Sentry, “not bad.”

Fixing code doesn’t have to be the worst part of your day. Learn how Sentry can help.

Learn more