There is the following business logic.
User uploads a file (xls, csv, google sheet). Files can be about 80K lines.
From the data in each row, records are created in the database (PostgreSQL
). Each record is indexed in ElasticSearch
.
It takes quite a long time. Therefore, I make data processing, writing to the database in celery
.
How else can speed up data processing, creating records in the database? Multiprocesses
? Threads
? Celery chunks
?