When we submit application to Spark, and after performing any operation Spark Web UI displays Job and Stages like count at MyJob.scala:15
. But in my application there are multiple count
and save
operations are there. So it is very difficult to understand UI. Instead of count at MyJob.scala:15
, can we add custom description to give more detailed information to job.
While googling found https://issues.apache.org/jira/browse/SPARK-3468 and https://github.com/apache/spark/pull/2342, author attached image, with detailed description like 'Count', 'Cache and Count', 'Job with delays'. So can we achieve same? I am using Spark 2.0.0.
use the sc.setJobGroup
:
Examples:
python:
In [28]: sc.setJobGroup("my job group id", "job description goes here") In [29]: lines = sc.parallelize([1,2,3,4]) In [30]: lines.count() Out[30]: 4
Scala:
scala> sc.setJobGroup("my job group id", "job description goes here") scala> val lines = sc.parallelize(List(1,2,3,4)) scala> lines.count() res3: Long = 4
SparkUI:
I hope this is what you are looking for.
Note that new Zeppelin 0.8 loose his tracking hook if you change the JobGroup name, and can't display his job progress bar (job still working, no effect on the job itself)
You can use
sc.setLocalProperty("callSite.short","my job description") sc.setLocalProperty("callSite.long","my job details long description")
instead
See How to change job/stage description in web UI? for some screen captures and scala syntax
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With