When building a simple perceptron neural network we usuall passes a 2D matrix of input of format (batch_size,features)
to a 2D weight matrix, similar to this simple neural network in numpy. I always assumed a Perceptron/Dense/Linear layer of a neural network only accepts an input of 2D format and outputs another 2D output. But recently I came across this pytorch model in which a Linear layer accepts a 3D input tensor and output another 3D tensor (o1 = self.a1(x)
).
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
class Net(nn.Module):
def __init__(self):
super().__init__()
self.a1 = nn.Linear(4,4)
self.a2 = nn.Linear(4,4)
self.a3 = nn.Linear(9,1)
def forward(self,x):
o1 = self.a1(x)
o2 = self.a2(x).transpose(1,2)
output = torch.bmm(o1,o2)
output = output.view(len(x),9)
output = self.a3(output)
return output
x = torch.randn(10,3,4)
y = torch.ones(10,1)
net = Net()
criterion = nn.MSELoss()
optimizer = optim.Adam(net.parameters())
for i in range(10):
net.zero_grad()
output = net(x)
loss = criterion(output,y)
loss.backward()
optimizer.step()
print(loss.item())
These are the question I have,
x = torch.randn(10,3,4)
, why is the pytorch nn.Linear
doesn't shows any error and gives a 3D output?Newer versions of PyTorch allows nn.Linear
to accept N-D input tensor, the only constraint is that the last dimension of the input tensor will equal in_features
of the linear layer. The linear transformation is then applied on the last dimension of the tensor.
For instance, if in_features=5
and out_features=10
and the input tensor x
has dimensions 2-3-5, then the output tensor will have dimensions 2-3-10.