after a lot of read and write operation to the hdfs , (i don't know the exact operation that cause this problem). these two files : dncp_block_verification.log.curr , dncp_block_verification.log.prev are more than 200 000 000 000 byte each.
what operation to hdfs may cause these file grow fast?
from the internet I know that I could shotdown the hdfs and delete the log,but it is not the good solution. how to avoid this problem? thank you very much
The block scanner is what is causing the files to grow. Here's a link to an article explaining the behavior: http://aosabook.org/en/hdfs.html (Section 8.3.5).
The bug which causes this has been fixed in HDFS 2.6.0
I've run into a similar situation with my 20 datanode cluster. I've seen several reports of this reportedly being a bug. I'm seeing this behavior with CDH 5.0.2 which runs HDFS 2.3.x.
I had 1 node out of 20 use 100% of all available space because of the two dncp_block_verification logs. The other nodes seem to have very typical log sizes and I can't figure out why this one node had an issue.
I manually deleted the two files and that resolved my issue with no adverse behavior. Unfortunately, I don't know of a fix or automated solution to this problem.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With