Is there a way to run multiple ONNX models in parallel and use multiple cores available?
Currently, I have trained two ONNX models and want to infer using them. I have used threading from Python but that doesn’t really use multiple cores.
After that I tried multiprocessing but that gives me below error:
can't pickle onnxruntime.capi.onnxruntime_pybind11_state.SessionOptions objects
Please let me know if there is any workaround to this?