I'm really new in this whole data engineering whilst I'm taking this matter as my thesis project, so bear with me.
I'm currently developing a big data platform for a battery storage system that already has CloudSQL services that collect data every 15 seconds (so it already in GCP environment). My job is to COPY THEM AND TRANSFER TO BIGQUERY EACH TIME A DATA CAME IN (including preparing the data using Dataprep) which will then be implemented to machine learning.
I have dug up several ways, one of them using Dataflow, tried once but it was done manually. using jdbc to bigquery
jobs. In order to fulfill my needs (running the jobs regularly), I was recommended using Cloud Composer.
On the other hand, I got another source that uses PubSub which triggers jobs to Dataflow. The latter approach seems more promising but, still, it's better to know both of the worlds. Any suggestion will definitely helps...