When running model.predict(features,batch_size=n)
with a batch_size>1
, are those predictions made in parallel - aka across multiple cores at once?
I need to permute millions of columns, so I was hoping to infer many predictions simultaneously using multiprocessing
. However, there are major challenges with this: Keras + Tensorflow and Multiprocessing in Python
One of my friends suggested stacking the data for use with batch_size as a workaround to increase performance, but I am wondering what the potential gains would be as it would be a big rewrite.