I need to read in a specific partition range using pyspark. I have seen various posts such as as this, that when using scala you can do the following:
val dataframe = sqlContext
.read
.parquet("file:///your/path/data=jDD/year=2015/month=10/day={5,6}/*")
val dataframe = sqlContext
.read
.parquet("file:///your/path/data=jDD/year=2015/month=10/day=[5-10]/*")
When using pyspark, the first method using {}
brackets works, this reads in specific partitions. However, I can't get the range method using []
to work.
I'm wondering is the syntax different for pyspark or is it just not supported?