The OpenCV Haar cascade classifier seems to use 24x24 images of faces as its positive training data. I have two questions regarding this:
- What are the consideration that go into selecting the training image size, besides the fact that larger training images require more processing?
- For non-square images, some people have chosen to keep one dimension at 24px, and expand the other dimension as necessary (to, say 100-200px). Is this the correct strategy?
- How does one go about deciding the size of the training images (this is a variant of question 1)