I'm looking at MS Fabric to:
- Ingest data from multiple sources including xls / csv sharepoint online files and azure SQL databases.
- Store this data in onelake
- Use synapse (or a workflow in data factory) to process the various sources data into a data warehouse
- use dataflows to read the DW data into smaller datasets to be used by multiple power bi reports.
My initial problem is how to read the sharepoint online files.
I can do this easily in the power bi service with a dataflow by choosing a web api connector, and pasting in the file path from sharepoint and choosing my organisational account to authenticate.
However, as I understand it, I can't then write that dataflow to onelake, to do that, I need to use Data Factory connection / copy data tasks.
How?
There are loads fewer connectors / data sources in DataFactory, so I assume that I need to use one of the generic protocols (http?) of the copy data activity, or the web activity to connect to my source sharepoint file(s)??
I've tried doing that, and using the sharepoint file path as the url, but I'm not getting anywhere - basically I haven't got a clue how to do it, and I can't find documentation that's helpful.
The closest I can find is how to access a sharepoint online list, but I don't need that, I need to access the actual files.
Has anyone got an idiots guide of how to do this?
Cheers