Questions tagged [multi-gpu]

This refers to one application's using multiple graphics-processing units, either in traditional (graphical) or general-purpose (GPGPU) applications.

387 questions
163
votes
5 answers

Using Java with Nvidia GPUs (CUDA)

I'm working on a business project that is done in Java, and it needs huge computation power to compute business markets. Simple math, but with huge amount of data. We ordered some CUDA GPUs to try it with and since Java is not supported by CUDA, I'm…
Hans
  • 1,846
  • 3
  • 14
  • 19
24
votes
1 answer

multi-GPU basic usage

How can I use two devices in order to improve for example the performance of the following code (sum of vectors)? Is it possible to use more devices "at the same time"? If yes, how can I manage the allocations of the vectors on the global memory of…
micheletuttafesta
  • 565
  • 2
  • 6
  • 11
22
votes
3 answers

How'd multi-GPU programming work with Vulkan?

Would using multi-GPUs in Vulkan be something like making many command queues then dividing command buffers between them? There are 2 problems: In OpenGL, we use GLEW to get functions. With more than 1 GPU, each GPU has its own driver. How'd we use…
Egy Prot
  • 345
  • 1
  • 2
  • 6
19
votes
1 answer

Is there a way to programmatically select the rendering GPU in a multi-GPU environment? (Windows)

Question I have an OpenGL application that will run in machines with diverse multi-GPU configurations (and possibly different Windows versions, from XP to 7). Is there a general way to select the specific GPU that will act as the OpenGL renderer…
Woundorf
  • 213
  • 3
  • 10
12
votes
1 answer

OpenGL multi-GPU support

When we create the OpenGL context on PC, is there any way to choose which physical device or how many devices are used? Do the latest OpenGL (4.5) APIs support multi-GPU architecture? If I have two identical graphics cards (for example, two Nvidia…
user3677630
  • 635
  • 6
  • 14
12
votes
3 answers

Ways to implement multi-GPU BN layers with synchronizing means and vars

I'd like to know the possible ways to implement batch normalization layers with synchronizing batch statistics when training with multi-GPU. Caffe Maybe there are some variants of caffe that could do, like link. But for BN layer, my understanding…
LI Xuhong
  • 2,339
  • 2
  • 17
  • 32
12
votes
5 answers

How to resolve CGDirectDisplayID changing issues on newer multi-GPU Apple laptops in Core Foundation/IO Kit?

In Mac OS X, every display gets a unique CGDirectDisplayID number assigned to it. You can use CGGetActiveDisplayList() or [NSScreen screens] to access them, among others. Per Apple's docs: A display ID can persist across processes and system…
Dave
  • 12,408
  • 12
  • 64
  • 67
11
votes
2 answers

How to fix low volatile GPU-Util with Tensorflow-GPU and Keras?

I have a 4 GPU machine on which I run Tensorflow (GPU) with Keras. Some of my classification problems take several hours to complete. nvidia-smi returns Volatile GPU-Util which never exceeds 25% on any of my 4 GPUs. How can I increase GPU Util%…
11
votes
5 answers

How to run Tensorflow Estimator on multiple GPUs with data parallelism

I have a standard tensorflow Estimator with some model and want to run it on multiple GPUs instead of just one. How can this be done using data parallelism? I searched the Tensorflow Docs but did not find an example; only sentences saying that it…
andy
  • 1,852
  • 2
  • 20
  • 31
11
votes
3 answers

How to do multi GPU training with Keras?

I want my model to run on multiple GPU-sharing parameters but with different batches of data. Can I do something like that with model.fit()? Is there any other alternative?
Rajarshee Mitra
  • 1,876
  • 28
  • 23
11
votes
2 answers

tensorflow using 2 GPU at the same time

First, I'm still newbie in tensorflow. I'm using v0.9 and trying to use the 2 GPUs installed in the machine we have. So, here is what's happening: When I launch a training data script on the machine, it works only on one of the 2 GPUs. It takes the…
Maystro
  • 2,907
  • 8
  • 36
  • 71
10
votes
5 answers

Pytorch Lightning duplicates main script in ddp mode

When I launch my main script on the cluster with ddp mode (2 GPU's), Pytorch Lightning duplicates whatever is executed in the main script, e.g. prints or other logic. I need some extended training logic, which I would like to handle myself. E.g. do…
dlsf
  • 332
  • 2
  • 13
10
votes
0 answers

Efficient allreduce is not supported for 2 IndexedSlices

I am trying to run a Subclassed Keras Model on multiple GPUs. The code is running as expected, however, the following "warning" crops up during the execution of the code: "Efficient allreduce is not supported for 2 IndexedSlices" What does this…
10
votes
2 answers

Tensorflow Java Multi-GPU inference

I have a server with multiple GPUs and want to make full use of them during model inference inside a java app. By default tensorflow seizes all available GPUs, but uses only the first one. I can think of three options to overcome this…
Alexander Lutsenko
  • 2,130
  • 8
  • 14
9
votes
1 answer

Managing multiple GPUs with multiple users

I have a server (Ubuntu 16.04) with 4 GPUs. My team shares this, and our current approach is to containerize all of our work with Docker, and to restrict containers to GPUs using something like $ NV_GPU=0 nvidia-docker run -ti nvidia/cuda…
SocraticDatum
  • 349
  • 2
  • 4
  • 15
1
2 3
25 26