I have source file compressed in delflate.gz
format. While loading the data into Spark data frame it failed with ArrayOutofBound
exception.
val cf = spark.read.option("header", "false").option("delimiter", "\u0001").option("codec", "deflate").csv("path/xxx.deflate.gz")
cf.show()
Error:
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 15.0 failed 4 times, most recent failure: Lost task 0.3 in stage 15.0 (TID 871, 10.180.255.33, executor 0): java.lang.ArrayIndexOutOfBoundsException: 63