add bias to attention v

pull/10/head
Eren Golge 2019-03-19 12:21:36 +01:00
parent 1898763028
commit 82cde95cfa
1 changed files with 1 additions and 1 deletions

View File

@ -118,7 +118,7 @@ class Attention(nn.Module):
attention_rnn_dim, attention_dim, bias=False, init_gain='tanh')
self.inputs_layer = Linear(
embedding_dim, attention_dim, bias=False, init_gain='tanh')
self.v = Linear(attention_dim, 1, bias=False)
self.v = Linear(attention_dim, 1, bias=True)
self.location_layer = LocationLayer(attention_location_n_filters,
attention_location_kernel_size,
attention_dim)