I got the following error output in the PyTorch when sent model predictions into the model. Does anyone know what's going on?
Following are the architecture model that I created, in the error output, it shows the issue exists in the x = self.fc1(cls_hs) line.
class BERT_Arch(nn.Module):
def __init__(self, bert):
super(BERT_Arch, self).__init__()
self.bert = bert
# dropout layer
self.dropout = nn.Dropout(0.1)
# relu activation function
self.relu = nn.ReLU()
# dense layer 1
self.fc1 = nn.Linear(768,512)
# dense layer 2 (Output layer)
self.fc2 = nn.Linear(512,2)
#softmax activation function
self.softmax = nn.LogSoftmax(dim=1)
#define the forward pass
def forward(self, sent_id, mask):
#pass the inputs to the model
_, cls_hs = self.bert(sent_id, attention_mask=mask)
print(mask)
print(type(mask))
x = self.fc1(cls_hs)
x = self.relu(x)
x = self.dropout(x)
# output layer
x = self.fc2(x)
# apply softmax activation
x = self.softmax(x)
return x
/usr/local/lib/python3.6/dist-packages/torch/nn/functional.py in linear(input, weight, bias)
1686 if any([type(t) is not Tensor for t in tens_ops]) and has_torch_function(tens_ops):
1687 return handle_torch_function(linear, tens_ops, input, weight, bias=bias)
-> 1688 if input == 2 and bias is not None:
1689 print(input)
1690 # fused op is marginally faster
AttributeError: 'str' object has no attribute 'dim'
If you work with transformers==3.0.0, everything should work fine !
There were some updates in transformers==4.0.0
To get transformers==3.0.0, following command can be used:
!pip install transformers==3.0.0
As mentioned here, the newer versions returns a special dictionary instead of a tuple. You can either change this line:
_, cls_hs = self.bert(sent_id, attention_mask=mask)
to
_, cls_hs = self.bert(sent_id, attention_mask=mask, return_dict=False)
or to
cls_hs = self.bert(sent_id, attention_mask=mask)[1]
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With