I have a package initialized with setuptools loaded in pyspark with a setup.py file like:
setup(
name='mypackagename',
version='0.9.10',
...
This package installs/ships, loads, and runs its code fine on the spark cluster. However, I am unable to log the version via get the package version within the package. This works fine if I load the library in normal python, but when the library is loaded in pyspark pkg_resources.get_distribution('mypackagename')
raises DistributionNotFound
exception.
Is there a better way to get the package version in pyspark?