I have built a decision tree model using Pyspark and I want to deploy that model using the docker container. I am using spark 1.6.0. The data is stored in the Hive tables and is located in my local machine. Is there a way to connect PySpark from my docker container to the hive tables in my local machine?
The data in my hive tables might get updated so I don't want to mount a drive or just copy the folder from local to my container but establish a connection between PySpark and Hive tables.