I'm running a Databricks job running few thousands of jobs and stages. When investigating Spark UI after the cluster is terminated, the UI only shows some of the latest jobs/stages instead of showing all of them (2002 in this example).
In my investigation I've found and set the below Spark Conf but still no luck as seen in the attached photo (using below Spark Conf).
spark.sql.ui.retainedExecutions 10000
spark.ui.retainedTasks 1000000
spark.ui.retainedStages 10000
spark.ui.retainedJobs 10000
Is there a way to keep all historical jobs/stages/tasks?
If not, how can one successfully debug a process after cluster termination?
After some research I understood it's a known issue though there is a workaround to be able to view and investigate the full content of the jobs/stages after the cluster is terminated. Please do the following in order to apply the workaround:
* For simplicity, let's call the large process we want to explore its SparkUI, the "Large Process".
spark.ui.retainedTasks 10000000
spark.ui.retainedJobs 1000000
spark.ui.retainedStages 10000000
spark.sql.ui.retainedExecutions 1000
Read more here.