From 53ca61db7a34c268ff05a2f670dbe6da7b3ec60a Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Tue, 7 Jun 2022 17:10:54 +0800 Subject: [PATCH] Reduce scale on decorrelation by 5, to 0.01 --- .../ASR/pruned_transducer_stateless2/decoder.py | 10 +--------- .../ASR/pruned_transducer_stateless5/conformer.py | 3 ++- 2 files changed, 3 insertions(+), 10 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless2/decoder.py b/egs/librispeech/ASR/pruned_transducer_stateless2/decoder.py index eba10f8c7..b6d94aaf1 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless2/decoder.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless2/decoder.py @@ -17,11 +17,7 @@ import torch import torch.nn as nn import torch.nn.functional as F -from scaling import ( - ScaledConv1d, - ScaledEmbedding, - Decorrelate, -) +from scaling import ScaledConv1d, ScaledEmbedding class Decoder(nn.Module): @@ -63,9 +59,6 @@ class Decoder(nn.Module): embedding_dim=decoder_dim, padding_idx=blank_id, ) - self.decorrelate = Decorrelate(apply_prob=0.25, - dropout_rate=0.05) - self.blank_id = blank_id assert context_size >= 1, context_size @@ -106,6 +99,5 @@ class Decoder(nn.Module): assert embedding_out.size(-1) == self.context_size embedding_out = self.conv(embedding_out) embedding_out = embedding_out.permute(0, 2, 1) - embedding_out = self.decorrelate(embedding_out) embedding_out = F.relu(embedding_out) return embedding_out diff --git a/egs/librispeech/ASR/pruned_transducer_stateless5/conformer.py b/egs/librispeech/ASR/pruned_transducer_stateless5/conformer.py index fd2278781..baf1441c3 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless5/conformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless5/conformer.py @@ -199,7 +199,8 @@ class ConformerEncoderLayer(nn.Module): ) self.dropout = torch.nn.Dropout(dropout) - self.decorrelate = Decorrelate(apply_prob=0.25, dropout_rate=0.05) + self.decorrelate = Decorrelate(apply_prob=0.25, + dropout_rate=0.01) def forward(