I'm trying to summarize how much data has been written to a folder in my Data Lake. What is the best way to do this? Should I use a U-SQL job? HDInsights?
There are two ways to do this:
Hope this helps
José
You can use Python code to loop through the files. Refer here: https://cloudarchitected.com/2019/05/computing-total-storage-size-of-a-folder-in-azure-data-lake-storage-gen2/
In case you would like to quickly cross check this:
Download the Azure Storage Explorer from Windows Application https://azure.microsoft.com/en-in/features/storage-explorer/
Open the folder which you would like to view the size details.
On the top bar menu choose More -> Folder Statistics will help you get the details of the Directory including the size in bytes. Refer the attachment [sample snapshot of the Azure Storage Explorer Menu[1]][1]
[1]: https://i.stack.imgur.com/R1DuZ.jpg
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With