I need to read in specific parquet files with spark, I know this can be done like so:
sqlContext
.read
.parquet("s3://bucket/key", "s3://bucket/key")
Right now I have a List[String] object with all these s3 paths in it but I don't know how I can pass this programmatically to the parquet function in Scala? There are way to many files to do it manually, any ideas how to get the files into the parquet function programmatically?