Abstract: We present competitive results using a Transformer encoder-decoder-attention model for end-to-end speech recognition needing less training time compared to a similarly performing LSTM model.
Some results have been hidden because they may be inaccessible to you
Show inaccessible results