To my understanding, the data-lake solution is used for storing everything from raw-data in the original format to processed data. I have not able to understand the concept of metadata-management in the (Azure) data-lake though. What are best-practices for dealing with metadata in the data-lake?
Are there any mechanisms to read metadata automatically (e.g. from header files) and if yes, are there any ways to view and edit this metadata (maybe an API to do it programatically)? I am worried that without proper management, the 'lake' will just turn to a "data-grave". One solution may be to create an own database where I myself store the metadata for each file. Are there any other more state-of-the-art approaches?