out of my two weeks of Azure experience. I want to split files based on a size. For example there is a table with 200k rows I would like to set a parameter to split that table into multiple files with a limit of 100Mb per file (if that makes sense). It will return N
number of files depending of the table size. something like:
my_file_1ofN.csv
I was walking through the documentation, blogs and videos and could do some POC with Azure Functions, Azure Batch, and Databricks with a python script in my personal account. The problem is the company doesn't let me use any of these approaches.
So I split the file using the number of partitions but these files are with a different sizes depending on the table and the partition.
Is there a way to accomplish this? I'm experimenting with lookups
and foreach
activities in the pipeline now but with not good results.
Any idea or clue will be welcome. Thanks!!