Spark Jobs

The Spark Jobs page provides a Sankey diagram with jobs by users and by queue. The page also provides detailed information about Spark jobs categorized by name of the job, users running the job, queue in which the job is run, status of the job, and other criteria.

Go to Spark > Jobs to view the Spark Jobs page.

The default time is Last 24 hrs. To change the timeline, click the down arrow in the time selection menu.

The default grouping is Queue. To change the grouping, perform the following:

  1. Click Group-By drop-down menu in the top right corner of the page, the drop-down menu is displayed.

  2. Select the type of grouping you want to display from the following list:

    1. Queue
    2. User
    3. Name
    4. Status
  3. The details by selected type is grouped in the table.

  4. (Optional)To ungroup the queries, select Ungroup.

The following tables provides description of the metrics displayed for each job group:

MetricDescription
Group NameThe name of the group.
Job CountThe total number of jobs associated by a particular group.
Failed JobsThe total number of failed jobs associated by a particular group.
Running JobsThe total number of running jobs associated by a particular group.
Avg Duration/JobThe average time taken to execute a job by a particular group.
Total DurationThe total time taken to execute jobs by a particular group.
Avg Mem/JobThe average amount of memory consumed by a particular job group.
Total MemThe total memory consumed by a particular job group.
Avg VCore/JobThe average number of VCores used by a particular job group.
Total VCoreThe total number of VCores used by a jobs by a particular group.

Click on the group row to view the details of the jobs in that group. The following tables provides description of the metrics displayed for each job:

MetricDescription
App IDThe application ID of the job. To display the concurrency of the job, click ... icon and click View Concurrency.
Attempt IDThe number of attempts made by the job.
UserThe name of the user who executed the job.
NameThe name of the spark job. To copy the name of the job, hover over the name you want to copy and click the icon. The name of the job is copied to clipboard.
Final StatusThe final state of the Spark job. The following are the status: Failed, Finishing, Killed, Running, Succeeded. Click the log icon to view the Logs for a specific job.
Dynamic ExecuterDisplays the execution condition for Spark Jobs such as True or False. Note The Simulation tab will not appear in the Reports if the Dynamic Executor is set to True.
QueueThe name of the queue that the job belongs to.
DurationThe total time taken to execute the job.
Started TimeThe starting time of the job
Finished TimeThe job's completion time.
MemoryThe amount of memory consumed in executing the job.
VCoresThe number of VCores used in job processing.

Click the App ID to view more details about the job in the Spark Job Details page.

Filtering Spark Jobs

To filter the data you want to view in the table, click the filter drop-down menu and select the filters. The following are the drop-down menu options:

  • User
  • Name
  • Final Status
  • Queue

You can search for the required filter using the Search box provided in the drop-down menu. Click Reset to clear the applied filter.

Searching Spark Jobs

You can use the search box to search for a job. For example, the name of the parameter visible in the UI is User. However, internally, the parameter name is effective_user. Hence, if you want to search for queries with User = Mark, you need to type effective_user:Mark.

Click Reset to clear the applied filter.

Type to search, ESC to discard
Type to search, ESC to discard
Type to search, ESC to discard