Daily, I ran an ELT process every 20 minutes (Talend), it takes some API REST response files (JSON), and some other JSON fields in a PostgreSQL server.
I've been processing with Talend, using JPATH to extract and "flat" the rows, so I get a CSV that I can freely load into BQ.
I don't feel this is the best way to do it since I iterate the pipes like 2000 times to make sure all the values will make a row in the end file, and then trim the NULL ones to just to load the ones with the proper value. This is dirty and slow.
I've always thought to switch to a cloud-based ETL (Stich, Fivetran, Xplenty, Alooma) but I don't know if those tools will be able to do this or load the JSON into BQ.
I've also explored the option of loading the JSON as a string in GQ and then use SCALAR to get the info I need in pure KPI rows, but I don't know if it will work.
The JSON format I need to load it's not a plan and some key/value, it's like this:
https://i.stack.imgur.com/3dgPs.png
(Zoom on "Properties")
https://i.stack.imgur.com/3dgPs.png
(It's test data, so no real information is here, but still, some of the test emails and names have been censored)
Side note, I've tried to contact Alooma guys, email, LinkedIn, even some google employees friends of mine, it's impossible, just want to try their product or even pay for it, they are like gone when Google bought them, did they die?
(i anyone interested this is a sample of 10 hits of the array: https://www.codepile.net/pile/q2ypDOxQ )