I have a data normalization process that exists in python but now needs to scale. This process currently runs via a job-specific configuration file containing a list of transforming functions that need to be applied to a table of data for that job. The transforming functions are mutually exclusive and can be applied in any order. All transforming functions live in a library and only get imported and applied to the data when they are listed in the job-specific configuration file. Different jobs will have different required functions listed in the configuration for that job, but all functions will exist in the library.
In the most general sense, how might a process like this be handled by AWS Glue? I don't need a technical example as much as a high level overview. Simply looking to be aware of some options. Thanks!