Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Enabling the layer normalization decrease the performance of the model #44

Open
AzizCode92 opened this issue Sep 5, 2018 · 3 comments

Comments

@AzizCode92
Copy link
Contributor

I have tried to add layer normalization to the layers.py by setting layer_norm to True as a default value.
I used the "listener.py" as my encoder.
Those modifications decreased the performance of the model ( training time is longer, both accuracy and WER decreased dramatically).
I have seen other users complaining about such problems with tf.contrib.rnn.LayerNormBasicLSTMCell
in stackoverflow
Am I missing something else?

@vrenkens
Copy link
Owner

vrenkens commented Sep 6, 2018

I have experienced the same issues... Maybe it is better to use the tf.contrib.rnn.LSTMCell instead

@AzizCode92
Copy link
Contributor Author

AzizCode92 commented Dec 3, 2018

Hi Vincent, I have used the TIMIT dataset and I used the recipe of LAS and it worked fine for my case.
I have attached my validation loss (the blue one is without layer normalization and the orange one is with layer normalization).
screen shot 2018-12-03 at 12 56 53

Why it didn't work for you case?
P.S: I used the KALDI-ASR project to process the dataset

with best regards,
Aziz

@AzizCode92 AzizCode92 reopened this Dec 3, 2018
@AzizCode92
Copy link
Contributor Author

Previously, I have pointed that it didn't work with the librispeech but I do have doubts that it is due to some mistake I might have done during the processing.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants