I'm struggling with creating dummy columns in a PySpark dataframe.
If I have a data frame with 10 columns (1 ID column, 9 object/string columns with n categories) In Python, I can simply do :
cols = list(df.columns)
cols.remove('ID')
df = pd.get_dummies(df[cols])
However, I cannot find a single resource where I can create the same result the above code references.