Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Issue about performance #7

Open
Shengqiang-Li opened this issue Jul 18, 2020 · 8 comments
Open

Issue about performance #7

Shengqiang-Li opened this issue Jul 18, 2020 · 8 comments

Comments

@Shengqiang-Li
Copy link

Shengqiang-Li commented Jul 18, 2020

I have finished training and decoding in AISHELL-1 dataset and got cer=12.4% in test set,and i found that my model.json which uses the default config is different from the one of Streaming_transformer-chunk32 with ESPnet Conv2d Encoder. It seems that my model lacks something,such as adaptive decoder.Can you release the result in AISHELL-1?

@Shengqiang-Li Shengqiang-Li changed the title Issue about RNNLM Issue about prefix_recognize Jul 19, 2020
@Shengqiang-Li Shengqiang-Li changed the title Issue about prefix_recognize Issue about tensorboard Jul 19, 2020
@Shengqiang-Li Shengqiang-Li changed the title Issue about tensorboard Issue about performance Jul 19, 2020
@MarkWuNLP
Copy link
Contributor

I have finished training and decoding in AISHELL-1 dataset and got cer=12.4% in test set,and i found that my model.json which uses the default config is different from the one of Streaming_transformer-chunk32 with ESPnet Conv2d Encoder. It seems that my model lacks something,such as adaptive decoder.Can you release the result in AISHELL-1?

Hi, This is Yu Wu, the mentor of Chengyi at MSRA. Chengyi is on vacation now, so she may not respond to you in a short time. As far as I known, Chengyi didn't try streaming SR on AISHELL-1 but her planning is to do it later. From my perspective, trigger attention algorithm is very tricky, so we have to do parameter tuning on different test sets to balance attention weight, ctc weight, and LM weight. I tried the algorithm on Microsoft internal dataset, and find the hyper-parameter is not a good setting for that dataset. I obtained absolute 10 WER gain by changing hyper-parameter

@Shengqiang-Li
Copy link
Author

Thank you very much.

@MarkWuNLP
Copy link
Contributor

Thank you very much.

Chengyi told me her model achieves 6.0 on AISELL-1 yesterday with a conformer architecture as an encoder.

@cywang97
Copy link
Owner

I have finished training and decoding in AISHELL-1 dataset and got cer=12.4% in test set,and i found that my model.json which uses the default config is different from the one of Streaming_transformer-chunk32 with ESPnet Conv2d Encoder. It seems that my model lacks something,such as adaptive decoder.Can you release the result in AISHELL-1?

Hi, I've done several experiments on AISHELL-1 dataset and got the following results on test set:
Offline Transformer 5.6
Offline Transformer+lm 5.4
Streaming Transformer-chunk16 6.6
Offline Conformer 4.9
Streaming Conformer-chunk16 6.2

I will release my code for Conformer and the pretrained model soon.

@huangzj421
Copy link

I have finished training and decoding in AISHELL-1 dataset and got cer=12.4% in test set,and i found that my model.json which uses the default config is different from the one of Streaming_transformer-chunk32 with ESPnet Conv2d Encoder. It seems that my model lacks something,such as adaptive decoder.Can you release the result in AISHELL-1?

Hi, I've done several experiments on AISHELL-1 dataset and got the following results on test set:
Offline Transformer 5.6
Offline Transformer+lm 5.4
Streaming Transformer-chunk16 6.6
Offline Conformer 4.9
Streaming Conformer-chunk16 6.2

I will release my code for Conformer and the pretrained model soon.

May i ask how is it going for Streaming Conformer-chunk? I have been waiting for a few months, thanks!

@Some-random
Copy link

I have finished training and decoding in AISHELL-1 dataset and got cer=12.4% in test set,and i found that my model.json which uses the default config is different from the one of Streaming_transformer-chunk32 with ESPnet Conv2d Encoder. It seems that my model lacks something,such as adaptive decoder.Can you release the result in AISHELL-1?

Hi, I've done several experiments on AISHELL-1 dataset and got the following results on test set:
Offline Transformer 5.6
Offline Transformer+lm 5.4
Streaming Transformer-chunk16 6.6
Offline Conformer 4.9
Streaming Conformer-chunk16 6.2

I will release my code for Conformer and the pretrained model soon.

I'm also looking forward to use these models, have them been released?

@cywang97
Copy link
Owner

Sorry for the late reply. For streaming Conformer, you can simply set the conv module in the encoder to casual conv and the self-attention layers are as same as in streaming Transformer. I have updated the code for conformer and I will release my model next week.

@TIFOSI528
Copy link

I have finished training and decoding in AISHELL-1 dataset and got cer=12.4% in test set,and i found that my model.json which uses the default config is different from the one of Streaming_transformer-chunk32 with ESPnet Conv2d Encoder. It seems that my model lacks something,such as adaptive decoder.Can you release the result in AISHELL-1?

Hi, I've done several experiments on AISHELL-1 dataset and got the following results on test set:
Offline Transformer 5.6
Offline Transformer+lm 5.4
Streaming Transformer-chunk16 6.6
Offline Conformer 4.9
Streaming Conformer-chunk16 6.2

I will release my code for Conformer and the pretrained model soon.

Where can I find the configs for these Aishell-1 experiments?
Many thanks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants