Hive Partitioned Tables have a folder structure with partition date as the folder. I have explored loading externally partitioned tables directly to bigquery which is possible.
What I would like to know is if this feature is possible to do with dataflow
since I am going to be running some feature transforms and such using dataflow
before loading the data into bigquery
. what I have found is if I add the partition date as a column then partitioning is possible but I am looking for a direct method with which I wouldn't be adding the column during transforms but directly while loading data into bigquery.
Is such a thing possible?