The Hadoop Big Data platform accepts jobs submitted by clients. In a multiuser environment, multiple jobs can be submitted and run simultaneously. The management of Hadoop jobs include checking job status, changing the priority of jobs, killing a running job, and so on. In this recipe, we will outline the steps to do these job management tasks.
We assume that our Hadoop cluster has been configured properly and all the Hadoop daemons are running without any issues. We also assume that a regular user can submit Hadoop jobs to the cluster.
Log in to the master node from the cluster administrator machine with the following command:
ssh hduser@master
Perform the following steps to check the status of Hadoop jobs:
List all the running jobs using the following command:
hadoop job -list
We will be able to get an output similar to the following:
1 jobs currently running JobId State StartTime UserName Priority SchedulingInfo job_201302152353_0001...