Currently model inference only runs on single GPU.
Inference using multiple GPUs could be done by splitting the list of images in
dataset_to_infer into subsets, distributing subsets of
dataset_to_infer to multiple GPUs.
(In an image window based inference, for each image the tasks of inference on window can also be distributed to multiple GPUs, but it can be difficult to implement/maintain.)