-
Notifications
You must be signed in to change notification settings - Fork 321
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problem during preparing G #23
Comments
Seems that the G is too large and there is an overflow in k2. |
Hm. I'd like to see some debug information about what's in |
I tried to reproduce this case for getting shapes, but it takes too long (2-3 hours) and the script is just killled:
I tried to count via irstlm instead of kenlm. When I pruned kenlm model, it was reduced to 2.1 gb from 5 gb for 3-gram. However I pruned irstlm model (with threshold 3e-7) and now sizes are 68M and 85M for 3-gram and 4-gram respectively. In the end, preparing G is done successfully. |
Hi, I am trying to train a tdnn-lstm model on a Turkish data. I run an experiment successfully once (including decoding part) with smaller langauge model. Then I tried a new corpus for language modeling. During
preparing G
step, I got this error:How can I solve that? My arpa files are 2.1gb and 6.2gb for 3 gram and 4 gram respectively. Could it be related to size issue? My language models are prepared via Kenlm.
Is it relevant to icefall, or should I ask on K2 repository?
The text was updated successfully, but these errors were encountered: