How to convert org.apache.spark.rdd.RDD[(String, String)] = org.apache.spark.rdd.RDD[String, String] in Spark core with Scala
INPUT (2020-01-19,ERROR) (2020-01-19,INFO)
OUTPUT 2020-01-19,ERROR 2020-01-19,INFO
How to convert org.apache.spark.rdd.RDD[(String, String)] = org.apache.spark.rdd.RDD[String, String] in Spark core with Scala
INPUT (2020-01-19,ERROR) (2020-01-19,INFO)
OUTPUT 2020-01-19,ERROR 2020-01-19,INFO
How to convert org.apache.spark.rdd.RDD[(String, String)] = org.apache.spark.rdd.RDD[String, String] in Spark core with Scala
You can't convert to org.apache.spark.rdd.RDD[String, String]
, because as you can clearly see from the documentation of org.apache.spark.rdd.RDD[T]
, it only takes one type parameter, not two. You can't convert to something that doesn't exist.
It looks like you come from Java spark. In Scala, the equivalent of JavaPairRDD<String, String>
is simply RDD[(String, String)]
: In Spark Scala, RDD[(K, V)]
s are extended with pair RDD behaviors thanks to implicit conversion (see PairRDDFunctions)