This refers to one application's using multiple graphics-processing units, either in traditional (graphical) or general-purpose (GPGPU) applications.
Questions tagged [multi-gpu]
387 questions
163
votes
5 answers
Using Java with Nvidia GPUs (CUDA)
I'm working on a business project that is done in Java, and it needs huge computation power to compute business markets. Simple math, but with huge amount of data.
We ordered some CUDA GPUs to try it with and since Java is not supported by CUDA, I'm…

Hans
- 1,846
- 3
- 14
- 19
24
votes
1 answer
multi-GPU basic usage
How can I use two devices in order to improve for example
the performance of the following code (sum of vectors)?
Is it possible to use more devices "at the same time"?
If yes, how can I manage the allocations of the vectors on the global memory of…

micheletuttafesta
- 565
- 2
- 6
- 11
22
votes
3 answers
How'd multi-GPU programming work with Vulkan?
Would using multi-GPUs in Vulkan be something like making many command queues then dividing command buffers between them?
There are 2 problems:
In OpenGL, we use GLEW to get functions. With more than 1 GPU, each GPU has its own driver. How'd we use…

Egy Prot
- 345
- 1
- 2
- 6
19
votes
1 answer
Is there a way to programmatically select the rendering GPU in a multi-GPU environment? (Windows)
Question
I have an OpenGL application that will run in machines with diverse multi-GPU configurations (and possibly different Windows versions, from XP to 7). Is there a general way to select the specific GPU that will act as the OpenGL renderer…

Woundorf
- 213
- 3
- 10
12
votes
1 answer
OpenGL multi-GPU support
When we create the OpenGL context on PC, is there any way to choose which physical device or how many devices are used? Do the latest OpenGL (4.5) APIs support multi-GPU architecture? If I have two identical graphics cards (for example, two Nvidia…

user3677630
- 635
- 6
- 14
12
votes
3 answers
Ways to implement multi-GPU BN layers with synchronizing means and vars
I'd like to know the possible ways to implement batch normalization layers with synchronizing batch statistics when training with multi-GPU.
Caffe Maybe there are some variants of caffe that could do, like link. But for BN layer, my understanding…

LI Xuhong
- 2,339
- 2
- 17
- 32
12
votes
5 answers
How to resolve CGDirectDisplayID changing issues on newer multi-GPU Apple laptops in Core Foundation/IO Kit?
In Mac OS X, every display gets a unique CGDirectDisplayID number assigned to it. You can use CGGetActiveDisplayList() or [NSScreen screens] to access them, among others. Per Apple's docs:
A display ID can persist across
processes and system…

Dave
- 12,408
- 12
- 64
- 67
11
votes
2 answers
How to fix low volatile GPU-Util with Tensorflow-GPU and Keras?
I have a 4 GPU machine on which I run Tensorflow (GPU) with Keras. Some of my classification problems take several hours to complete.
nvidia-smi returns Volatile GPU-Util which never exceeds 25% on any of my 4 GPUs.
How can I increase GPU Util%…

Sharanya Arcot Desai
- 685
- 3
- 7
- 18
11
votes
5 answers
How to run Tensorflow Estimator on multiple GPUs with data parallelism
I have a standard tensorflow Estimator with some model and want to run it on multiple GPUs instead of just one. How can this be done using data parallelism?
I searched the Tensorflow Docs but did not find an example; only sentences saying that it…

andy
- 1,852
- 2
- 20
- 31
11
votes
3 answers
How to do multi GPU training with Keras?
I want my model to run on multiple GPU-sharing parameters but with different batches of data.
Can I do something like that with model.fit()? Is there any other alternative?

Rajarshee Mitra
- 1,876
- 28
- 23
11
votes
2 answers
tensorflow using 2 GPU at the same time
First, I'm still newbie in tensorflow. I'm using v0.9 and trying to use the 2 GPUs installed in the machine we have. So, here is what's happening:
When I launch a training data script on the machine, it works only on one of the 2 GPUs. It takes the…

Maystro
- 2,907
- 8
- 36
- 71
10
votes
5 answers
Pytorch Lightning duplicates main script in ddp mode
When I launch my main script on the cluster with ddp mode (2 GPU's), Pytorch Lightning duplicates whatever is executed in the main script, e.g. prints or other logic. I need some extended training logic, which I would like to handle myself. E.g. do…

dlsf
- 332
- 2
- 13
10
votes
0 answers
Efficient allreduce is not supported for 2 IndexedSlices
I am trying to run a Subclassed Keras Model on multiple GPUs. The code is running as expected, however, the following "warning" crops up during the execution of the code:
"Efficient allreduce is not supported for 2 IndexedSlices"
What does this…

Damayanti Sengupta
- 101
- 2
10
votes
2 answers
Tensorflow Java Multi-GPU inference
I have a server with multiple GPUs and want to make full use of them during model inference inside a java app.
By default tensorflow seizes all available GPUs, but uses only the first one.
I can think of three options to overcome this…

Alexander Lutsenko
- 2,130
- 8
- 14
9
votes
1 answer
Managing multiple GPUs with multiple users
I have a server (Ubuntu 16.04) with 4 GPUs. My team shares this, and our current approach is to containerize all of our work with Docker, and to restrict containers to GPUs using something like $ NV_GPU=0 nvidia-docker run -ti nvidia/cuda…

SocraticDatum
- 349
- 2
- 4
- 15