From 7d45e5d01163a665a3921144e5ed0cceff226f77 Mon Sep 17 00:00:00 2001 From: Eren Golge Date: Tue, 4 Jun 2019 00:40:03 +0200 Subject: [PATCH] remove print --- layers/common_layers.py | 1 - 1 file changed, 1 deletion(-) diff --git a/layers/common_layers.py b/layers/common_layers.py index 061134e2..b6f72bc1 100644 --- a/layers/common_layers.py +++ b/layers/common_layers.py @@ -207,7 +207,6 @@ class Attention(nn.Module): if not self.training and self.forward_attn_mask: _, n = prev_alpha.max(1) val, n2 = alpha.max(1) - print(True) for b in range(alignment.shape[0]): alpha[b, n[b] + 2:] = 0 alpha[b, :(n[b] - 1)] = 0 # ignore all previous states to prevent repetition.