0

I am processing my cassandra data in spark . I am using "spark-cassandra-connector " that gets data from cassandra into an RDD .

I wan to use spark 2.* s datasets , as datasets will improve my performance .Any Idea how can I do that ?

Any code snippet will be a great help

Sandeep Das
  • 1,010
  • 9
  • 22

1 Answers1

3

Use

spark.read.format("org.apache.spark.sql.cassandra")
   .options(Map("keyspace" -> "your_keyspake", "table" -> "your_table"))
   .load.filter(conditions)

You don't have to convert from rdd to dataset.

Alex Ott
  • 80,552
  • 8
  • 87
  • 132
M. Alexandru
  • 614
  • 5
  • 20