From cd36d149dfbf3730ade1e7b8bc2baccd77ac277f Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Mon, 29 May 2023 17:32:49 +0800 Subject: [PATCH] Reduce encoder-dim and num-heads of center stack. --- egs/libriheavy/LM/zipformer1/train.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/egs/libriheavy/LM/zipformer1/train.py b/egs/libriheavy/LM/zipformer1/train.py index f3f70481a..f470201aa 100755 --- a/egs/libriheavy/LM/zipformer1/train.py +++ b/egs/libriheavy/LM/zipformer1/train.py @@ -136,14 +136,14 @@ def add_model_arguments(parser: argparse.ArgumentParser): parser.add_argument( "--num-heads", type=str, - default="4,4,8,16,32,16,8,4,4", + default="4,4,8,16,16,16,8,4,4", help="Number of attention heads in the subformer encoder layers: a single int or comma-separated list.", ) parser.add_argument( "--encoder-dim", type=str, - default="256,384,512,768,1024,768,512,384,256", + default="256,384,512,768,768,768,512,384,256", help="Embedding dimension in encoder stacks: a single int or comma-separated list." )