I would like to transform the values of a column into multiple columns of a dataframe in pyspark on databricks.
e.g
from pyspark.sql import SparkSession
spark = SparkSession.builder.getOrCreate()
df = spark._sc.parallelize([["dapd", "shop", "retail"],
["dapd", "shop", "on-line"],
["dapd", "payment", "credit"],
["wrfr", "shop", "supermarket"],
["wrfr", "shop", "brand store"],
["wrfr", "payment", "cash"]]).toDF(["id", "value1", "value2"])
I need to transform it to:
id, shop payment
dapd retail|on-line credit
wrfr supermarket|brand store cash
I am not sure how I can do this in pyspark ?
Thanks,