I want to define an environment variable in Databricks init script and then read it in Pyspark notebook. I wrote this:
dbutils.fs.put("/databricks/scripts/initscript.sh","""
#!/bin/bash
export env="dev"
pip install pretty-html-table==0.9.14
""", True)
the pyspark code:
import os
environment=os.getenv("env")
it gives:
TypeError: can only concatenate str (not "NoneType") to str
but the Pyspark notebook is not able to read the environment variable properly
Any idea how to fix this ?