mirror of https://github.com/coqui-ai/TTS.git
use sigmoid for attention
parent
7e020d4084
commit
4431e04b48
|
@ -170,9 +170,9 @@ class AttentionRNNCell(nn.Module):
|
||||||
# Update the window
|
# Update the window
|
||||||
self.win_idx = torch.argmax(alignment,1).long()[0].item()
|
self.win_idx = torch.argmax(alignment,1).long()[0].item()
|
||||||
# Normalize context weight
|
# Normalize context weight
|
||||||
alignment = F.softmax(alignment, dim=-1)
|
# alignment = F.softmax(alignment, dim=-1)
|
||||||
# alignment = 5 * alignment
|
# alignment = 5 * alignment
|
||||||
# alignment = torch.sigmoid(alignment) / torch.sigmoid(alignment).sum(dim=1).unsqueeze(1)
|
alignment = torch.sigmoid(alignment) / torch.sigmoid(alignment).sum(dim=1).unsqueeze(1)
|
||||||
# Attention context vector
|
# Attention context vector
|
||||||
# (batch, 1, dim)
|
# (batch, 1, dim)
|
||||||
# c_i = \sum_{j=1}^{T_x} \alpha_{ij} h_j
|
# c_i = \sum_{j=1}^{T_x} \alpha_{ij} h_j
|
||||||
|
|
Loading…
Reference in New Issue