-
Notifications
You must be signed in to change notification settings - Fork 41
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue about performance #7
Comments
Hi, This is Yu Wu, the mentor of Chengyi at MSRA. Chengyi is on vacation now, so she may not respond to you in a short time. As far as I known, Chengyi didn't try streaming SR on AISHELL-1 but her planning is to do it later. From my perspective, trigger attention algorithm is very tricky, so we have to do parameter tuning on different test sets to balance attention weight, ctc weight, and LM weight. I tried the algorithm on Microsoft internal dataset, and find the hyper-parameter is not a good setting for that dataset. I obtained absolute 10 WER gain by changing hyper-parameter |
Thank you very much. |
Chengyi told me her model achieves 6.0 on AISELL-1 yesterday with a conformer architecture as an encoder. |
Hi, I've done several experiments on AISHELL-1 dataset and got the following results on test set: I will release my code for Conformer and the pretrained model soon. |
May i ask how is it going for Streaming Conformer-chunk? I have been waiting for a few months, thanks! |
I'm also looking forward to use these models, have them been released? |
Sorry for the late reply. For streaming Conformer, you can simply set the conv module in the encoder to casual conv and the self-attention layers are as same as in streaming Transformer. I have updated the code for conformer and I will release my model next week. |
Where can I find the configs for these Aishell-1 experiments? |
I have finished training and decoding in AISHELL-1 dataset and got cer=12.4% in test set,and i found that my model.json which uses the default config is different from the one of Streaming_transformer-chunk32 with ESPnet Conv2d Encoder. It seems that my model lacks something,such as adaptive decoder.Can you release the result in AISHELL-1?
The text was updated successfully, but these errors were encountered: