This is the runtime of WeNet.
We are going to support the following platforms:
- Various deep learning inference engines, such as LibTorch, ONNX, OpenVINO, TVM, and so on.
- Various OS, such as android, iOS, Harmony, and so on.
- Various AI chips, such as GPU, Horzion BPU, and so on.
- Various hardware platforms, such as Raspberry Pi.
- Various language binding, such as python and go.
Feel free to volunteer yourself if you are interested in trying out some items(they do not have to be on the list).
For each platform, we will create a subdirectory in runtime. Currently, we have:
- LibTorch: in c++, the default engine of WeNet.
- OnnxRuntime: in c++, the official runtime for onnx model.
- GPU: in python, powered by triton.
- android: in java, it shows an APP demo.
- Language binding
- binding/python: python is the first class for binding.
- binding/go: ongoing.