Questions tagged [multi-process-service]

The Multi-Process Service (MPS) is an alternative, binary-compatible implementation of the CUDA Application Programming Interface (API).

The MPS runtime architecture is designed to transparently enable co-operative multi-process CUDA applications, typically MPI jobs, to utilize Hyper-Q capabilities on the latest NVIDIA (Kepler-based) GPUs. Hyper-Q allows CUDA kernels to be processed concurrently on the same GPU; this can benefit performance when the GPU compute capacity is underutilized by a single application process.

Online documentation is available here.

4 questions
2
votes
0 answers

How to run mpirun on vps properly?

i tried to run mpirun on my vps to do computation for my thesis. but when i tried to run this command, an error occurs mpirun -np 3 "myProcess" -parallel the error says mpirun has detected an attempt to run as root. Running at root is *strongly*…
deb97
  • 61
  • 1
  • 6
0
votes
0 answers

limiting access to GPU threads using CUDA MPS(Multi-Process Service)

I am using CUDA MPS (Multi-Process Service) and trying to understand how the number of active GPU threads affects GPU utilization and execution time. I have written a Python script that multiplies two arrays of length 1,000,000. However, I am…
0
votes
1 answer

Fine grained Kernel scheduling with MPS

I am working on using NVIDIA CUDA Multi Process Service (MPS) for running multiple TensorFlow inference jobs using the same GPU. For my use-case, when GPU is being shared by more than one processes, I sometimes need to prioritize execution of…
0
votes
2 answers

How to reduce CUDA context size (Multi-Process Service)

I followed Robert Crovella's example on how to use Nvidia's Multi-Process Service. According to docs: 2.1.2. Reduced on-GPU context storage Without MPS each CUDA processes using a GPU allocates separate storage and scheduling resources on the…
alex
  • 10,900
  • 15
  • 70
  • 100