GPU-Accelerated Forward-Backward algorithm with Application to Lattice-Free MMI.

2021
We propose to express the forward-backward algorithm in terms of operations between sparse matrices in a specific semiring. This new perspective naturally leads to a GPU-friendly algorithm which is easy to implement in Julia or any programming languages with native support of semiring algebra. We use this new implementation to train a TDNN with the LF-MMI objective function and we compare the training time of our system with PyChain - a recently introduced C++/CUDA implementation of the LF-MMI loss. Our implementation is about two times faster while not having to use any approximation such as the "leaky-HMM".
    • Correction
    • Source
    • Cite
    • Save
    16
    References
    1
    Citations
    NaN
    KQI
    []
    Baidu
    map