Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

semi/ptm training is excessively, probably unnecessarily slow #42

Open
dhdaines opened this issue Aug 5, 2022 · 0 comments
Open

semi/ptm training is excessively, probably unnecessarily slow #42

dhdaines opened this issue Aug 5, 2022 · 0 comments

Comments

@dhdaines
Copy link
Contributor

dhdaines commented Aug 5, 2022

In theory, semi-continuous or PTM models are supposed to be fast! But training them is incredibly slow, especially the initial flat start. This is most likely due to some redundant or inefficient computation in the training code.

Training a 128-Gaussian PTM model on 100 hours of data on 16 CPUs takes approximately 4 hours, whereas training a 4000 senone, 16 Gaussian continuous model with LDA and MLLT takes only 1h25 (without LDA and MLLT it's less than an hour).

And of course the accuracy of said PTM model is quite atrocious.

One might argue that they are thoroughly obsolete, but actually they are maybe the only reason left to use CMU Sphinx, since they produce very small models.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

No branches or pull requests

1 participant