I have trained a Tesseract 4 LSTM model against a set of ~30,000 ground truth images that I generated (as opposed to using "real" images from scanned works, of which I do not have enough to reliably train a model).
The model works well (or at least better than eng
, on which it is based). The image generation script has several parameters that I can adjust, but I'd like to do that in a more "ordered" way than just eyeballing the output, so I'd like to generate metrics based on accuracy across a (much smaller) set of real-world images.
However, it is not clear to me how you can take a set of line images and ground truth text files and generate the required files to run lstmeval
on the new model. How do you generate the data to feed to lstmeval
when the evaluation images are not related to the images actually used to train the model in the first place?