0

I have been testing with a word2vec model. This word2vec model for some reason doesn't use the gpu much. My performance is roughly 1 epoch every 30 seconds with a ~2000 samples dataset.

This doesn't seem normal. There are researchers that have gigabytes of training data, and I doubt they are waiting for months for the training to finish.

My GPU is a gtx 970. The memory usage is around 10% (Note that I have a few programs open too)

The problem might be the batches itself, although I am not sure.

Basically I run a method at the start of the training, and then while training I iterate over the entries in that list.

This is roughly how I do this. Is my approach wrong? (I would guess that it's not suitable for huge datasets)

batch_method(batch_size=x) # I tested with different sizes, all seem to train fine, from 2 to 512.
for epo in self.epochs_num:
    for batch in self.batch_list:
        for input,target in batch:
        ...
G. Ramistella
  • 1,327
  • 1
  • 9
  • 19

0 Answers0