I've always read that Scala is much faster than pyspark for many operations, but recently I've read in a blog that since the release of Spark 2 the performance differences are much lower.
Is this maybe due to the Dataframe introduction? Does that mean that operations on dataframe take the same time with Scala and pyspark?
Does exist a detailed and recent performance report about Scala/pyspark differences?