I am using ubuntu 16.04, tensorflow 1.3
A network with ~ 17M weights
Experiments
image size 400x1000, batch size 4, during graph construction:
failed to alloc 34359738368 bytes on host: CUDA_ERROR_OUT_OF_MEMORY
image size 300x750, batch size 4, during graph construction:
failed to alloc 34359738368 bytes on host: CUDA_ERROR_OUT_OF_MEMORY
image size 300x740, batch size 1, during graph construction:
failed to alloc 34359738368 bytes on host: CUDA_ERROR_OUT_OF_MEMORY
So, the memory requested is the same for all the three experiment. My question is does 17M weights really need such a huge amount of memory? And why the required memory doesn't change with different images sizes and batch sizes ?