I am using Tensorflow,python, Pandas to create a logistic regression model similar to this link
Instead of MNist dataset, I am using my own dataset. I use Pandas to create the dataframes, replace nulls with fillna function and then convert it into tensor dataset using from_tensor_slices.
I have many CATEGORICAL_COLUMNS and I am using get_dummies to do the OneHotEncoding (along with LabelEncoding). But the issue is, my categorical columns have big vocabulary list (Ex: Zipcode...i have thousands of zipcodes in my data). So when I create columns using "get_dummies", the Zip code is creating lots of new columns for me.
Is this a good approach? Is this the way I should approach this kind of dataset ?