The situation is the following:
If I manually intervene and kill the application, the disk space is cleaned up. Now I can manually restart the application and it's fine.
I wish I could tell the automated retry to clean up the disk. Alternatively I suppose it could count that used disk as part of the new allocation, since it belongs to the application anyway.
I'll happily take any solution you can offer. I don't know much about YARN. It's an Apache Spark application started with spark-submit
in yarn-client
mode. The files that fill up the disk are the shuffle spill files.
So here's what happens:
local directories and won't be deleted on JVM exit when using the external shuffle service.
This directory can be cleaned via:
yarn.nodemanager.delete.debug-delay-sec
=0. Otherwise there is some unresolved yarn bug
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With