I want the latest updated file from one of my HDFS directories. The code should basically loop through the directories and sub directories and the get the latest file path with the file name.I was able to get the latest file in local file system but not sure how to do it for HDFS one.
find /tmp/sdsa -type f -print0 | xargs -0 stat --format '%Y :%y %n' | sort -nr | cut -d: -f2- | head
The above code is working for local file system. I am able to get the date , time and file name from HDFS, but how do I get the latest file using these 3 parameters?
this is the code I tried:
hadoop fs -ls -R /tmp/apps | awk -F" " '{print $6" "$7" "$8}'
Any help will be appreciated.
Thanks in advance.
This command is used to empty the trash available in an HDFS system. Syntax: $ hadoop fs –expunge.
tail. Displays last kilobyte of the file to stdout. -f option can be used as in Unix.
The File System (FS) shell includes various shell-like commands that directly interact with the Hadoop Distributed File System (HDFS) as well as other file systems that Hadoop supports, such as Local FS, WebHDFS, S3 FS, and others. The FS shell is invoked by: bin/hadoop fs <args>
You can't UPDATE any existing record in HDFS, but yes, you can surely make another copy of the data (with the modifications/updates) in the HDFS and can remove the previous original copy.
This one worked for me:
hadoop fs -ls -R /tmp/app | awk -F" " '{print $6" "$7" "$8}' | sort -nr | head -1 | cut -d" " -f3
The output is the entire file path.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With