Questions tagged [gpu]

Acronym for "Graphics Processing Unit". For programming traditional graphical applications, see the tag entry for "graphics programming". For general-purpose programming using GPUs, see the tag entry for "gpgpu". For specific GPU programming technologies, see the popular tag entries for "opencl", "cuda" and "thrust".

Acronym for "Graphics Processing Unit". For programming traditional graphical applications, see the tag entry for . For general-purpose programming using GPUs, see the tag entry for . For specific GPU programming technologies, see the popular tag entries for , and .

More information on GPU at http://en.wikipedia.org/wiki/Graphics_processing_unit

8854 questions
521
votes
18 answers

How do I check if PyTorch is using the GPU?

How do I check if PyTorch is using the GPU? The nvidia-smi command can detect GPU activity, but I want to check it directly from inside a Python script.
vvvvv
  • 25,404
  • 19
  • 49
  • 81
518
votes
19 answers

Nvidia NVML Driver/library version mismatch

When I run nvidia-smi, I get the following message: Failed to initialize NVML: Driver/library version mismatch An hour ago I received the same message and uninstalled my CUDA library and I was able to run nvidia-smi, getting the following…
etal
  • 12,914
  • 4
  • 13
  • 16
441
votes
31 answers

How to tell if tensorflow is using gpu acceleration from inside python shell?

I have installed tensorflow in my ubuntu 16.04 using the second answer here with ubuntu's builtin apt cuda installation. Now my question is how can I test if tensorflow is really using gpu? I have a gtx 960m gpu. When I import tensorflow this is the…
Tamim Addari
  • 7,591
  • 9
  • 40
  • 59
284
votes
3 answers

Once upon a time, when > was faster than < ... Wait, what?

I am reading an awesome OpenGL tutorial. It's really great, trust me. The topic I am currently at is Z-buffer. Aside from explaining what's it all about, the author mentions that we can perform custom depth tests, such as GL_LESS, GL_ALWAYS, etc. He…
Armen Tsirunyan
  • 130,161
  • 59
  • 324
  • 434
233
votes
16 answers

How to get current available GPUs in tensorflow?

I have a plan to use distributed TensorFlow, and I saw TensorFlow can use GPUs for training and testing. In a cluster environment, each machine could have 0 or 1 or more GPUs, and I want to run my TensorFlow graph into GPUs on as many machines as…
Sangwon Kim
  • 2,435
  • 2
  • 13
  • 9
186
votes
9 answers

Intro to GPU programming

Everyone has this huge massively parallelized supercomputer on their desktop in the form of a graphics card GPU. What is the "hello world" equivalent of the GPU community? What do I do, where do I go, to get started programming the GPU for the…
Adam Davis
  • 91,931
  • 60
  • 264
  • 330
163
votes
5 answers

Using Java with Nvidia GPUs (CUDA)

I'm working on a business project that is done in Java, and it needs huge computation power to compute business markets. Simple math, but with huge amount of data. We ordered some CUDA GPUs to try it with and since Java is not supported by CUDA, I'm…
Hans
  • 1,846
  • 3
  • 14
  • 19
146
votes
13 answers

Utilizing the GPU with c#

I am trying to get more processing power out of my grid. I am using all cpus/cores, is it possible to utilize the GPU with C#. Anyone know any libraries or got any sample code?
csharpdevguy568
137
votes
3 answers

How do I choose grid and block dimensions for CUDA kernels?

This is a question about how to determine the CUDA grid, block and thread sizes. This is an additional question to the one posted here. Following this link, the answer from talonmies contains a code snippet (see below). I don't understand the…
user1292251
  • 1,655
  • 3
  • 16
  • 16
134
votes
7 answers

GPU Emulator for CUDA programming without the hardware

Question: Is there an emulator for a Geforce card that would allow me to program and test CUDA without having the actual hardware? Info: I'm looking to speed up a few simulations of mine in CUDA, but my problem is that I'm not always around my…
Narcolapser
  • 5,895
  • 15
  • 46
  • 56
133
votes
9 answers

Google Colaboratory: misleading information about its GPU (only 5% RAM available to some users)

update: this question is related to Google Colab's "Notebook settings: Hardware accelerator: GPU". This question was written before the "TPU" option was added. Reading multiple excited announcements about Google Colaboratory providing free Tesla K80…
stason
  • 5,409
  • 4
  • 34
  • 48
127
votes
9 answers

Is it possible to run CUDA on AMD GPUs?

I'd like to extend my skill set into GPU computing. I am familiar with raytracing and realtime graphics(OpenGL), but the next generation of graphics and high performance computing seems to be in GPU computing or something like it. I currently use an…
Lee Jacobs
  • 1,692
  • 3
  • 12
  • 21
109
votes
6 answers

Can I run CUDA on Intel's integrated graphics processor?

I have a very simple Toshiba Laptop with i3 processor. Also, I do not have any expensive graphics card. In the display settings, I see Intel(HD) Graphics as display adapter. I am planning to learn some cuda programming. But, I am not sure, if I can…
Ankit
  • 6,772
  • 11
  • 48
  • 84
105
votes
2 answers

nvidia-smi Volatile GPU-Utilization explanation?

I know that nvidia-smi -l 1 will give the GPU usage every one second (similarly to the following). However, I would appreciate an explanation on what Volatile GPU-Util really means. Is that the number of used SMs over total SMs, or the occupancy, or…
user3813674
  • 2,553
  • 2
  • 15
  • 26
96
votes
4 answers

How does OpenGL work at the lowest level?

I understand how to write OpenGL/DirectX programs, and I know the maths and the conceptual stuff behind it, but I'm curious how the GPU-CPU communication works on a low level. Say I've got an OpenGL program written in C that displays a triangle and…
Benno
  • 5,288
  • 5
  • 42
  • 60
1
2 3
99 100