Export model with torch.jit.script()
In this section, we describe how to export a model via
torch.jit.script().
When to use it
If we want to use our trained model with torchscript,
we can use torch.jit.script().
Hint
See Export model with torch.jit.trace()
if you want to use torch.jit.trace().
How to export
We use https://github.com/k2-fsa/icefall/tree/master/egs/librispeech/ASR/pruned_transducer_stateless3 as an example in the following.
cd egs/librispeech/ASR
epoch=14
avg=1
./pruned_transducer_stateless3/export.py \
--exp-dir ./pruned_transducer_stateless3/exp \
--tokens data/lang_bpe_500/tokens.txt \
--epoch $epoch \
--avg $avg \
--jit 1
It will generate a file cpu_jit.pt in pruned_transducer_stateless3/exp.
Caution
Don’t be confused by cpu in cpu_jit.pt. We move all parameters
to CPU before saving it into a pt file; that’s why we use cpu
in the filename.
How to use the exported model
Please refer to the following pages for usage:
https://k2-fsa.github.io/sherpa/python/streaming_asr/emformer/index.html
https://k2-fsa.github.io/sherpa/python/streaming_asr/conv_emformer/index.html
https://k2-fsa.github.io/sherpa/python/streaming_asr/conformer/index.html
https://k2-fsa.github.io/sherpa/python/offline_asr/conformer/index.html
https://k2-fsa.github.io/sherpa/cpp/offline_asr/gigaspeech.html
https://k2-fsa.github.io/sherpa/cpp/offline_asr/wenetspeech.html