| Modifier and Type | Class | Description |
|---|---|---|
static class |
JobImpl.InitTransition |
|
static class |
JobImpl.StartTransition |
| Modifier and Type | Field | Description |
|---|---|---|
org.apache.hadoop.mapred.JobConf |
conf |
|
static java.lang.String |
JOB_KILLED_DIAG |
|
org.apache.hadoop.fs.Path |
remoteJobConfFile |
|
protected static org.apache.hadoop.yarn.state.StateMachineFactory<JobImpl,JobStateInternal,JobEventType,JobEvent> |
stateMachineFactory |
| Constructor | Description |
|---|---|
JobImpl(org.apache.hadoop.mapreduce.v2.api.records.JobId jobId,
org.apache.hadoop.yarn.api.records.ApplicationAttemptId applicationAttemptId,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.yarn.event.EventHandler eventHandler,
TaskAttemptListener taskAttemptListener,
org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager jobTokenSecretManager,
org.apache.hadoop.security.Credentials jobCredentials,
org.apache.hadoop.yarn.util.Clock clock,
java.util.Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo> completedTasksFromPreviousRun,
MRAppMetrics metrics,
org.apache.hadoop.mapreduce.OutputCommitter committer,
boolean newApiCommitter,
java.lang.String userName,
long appSubmitTime,
java.util.List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> amInfos,
AppContext appContext,
JobStateInternal forcedState,
java.lang.String forcedDiagnostic) |
| Modifier and Type | Method | Description |
|---|---|---|
protected void |
addDiagnostic(java.lang.String diag) |
|
protected void |
addTask(Task task) |
|
boolean |
checkAccess(org.apache.hadoop.security.UserGroupInformation callerUGI,
org.apache.hadoop.mapreduce.JobACL jobOperation) |
|
protected JobStateInternal |
checkReadyForCommit() |
|
void |
constructFinalFullcounters() |
|
static java.lang.String |
escapeString(java.lang.String data) |
|
org.apache.hadoop.mapreduce.Counters |
getAllCounters() |
Get all the counters of this job.
|
java.util.List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> |
getAMInfos() |
|
int |
getCompletedMaps() |
|
int |
getCompletedReduces() |
|
org.apache.hadoop.fs.Path |
getConfFile() |
|
java.util.List<java.lang.String> |
getDiagnostics() |
|
int |
getFailedMaps() |
|
int |
getFailedReduces() |
|
protected org.apache.hadoop.fs.FileSystem |
getFileSystem(org.apache.hadoop.conf.Configuration conf) |
Create the default file System for this job.
|
org.apache.hadoop.mapreduce.v2.api.records.JobId |
getID() |
|
JobStateInternal |
getInternalState() |
|
java.util.Map<org.apache.hadoop.mapreduce.JobACL,org.apache.hadoop.security.authorize.AccessControlList> |
getJobACLs() |
|
int |
getKilledMaps() |
|
int |
getKilledReduces() |
|
org.apache.hadoop.mapred.TaskCompletionEvent[] |
getMapAttemptCompletionEvents(int startIndex,
int maxEvents) |
|
float |
getMaxAllowedFetchFailuresFraction() |
|
int |
getMaxFetchFailuresNotifications() |
|
java.lang.String |
getName() |
|
float |
getProgress() |
|
java.lang.String |
getQueueName() |
|
org.apache.hadoop.mapreduce.v2.api.records.JobReport |
getReport() |
|
org.apache.hadoop.mapreduce.v2.api.records.JobState |
getState() |
|
protected org.apache.hadoop.yarn.state.StateMachine<JobStateInternal,JobEventType,JobEvent> |
getStateMachine() |
|
Task |
getTask(org.apache.hadoop.mapreduce.v2.api.records.TaskId taskID) |
|
org.apache.hadoop.mapreduce.v2.api.records.TaskAttemptCompletionEvent[] |
getTaskAttemptCompletionEvents(int fromEventId,
int maxEvents) |
|
java.util.Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> |
getTasks() |
|
java.util.Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> |
getTasks(org.apache.hadoop.mapreduce.v2.api.records.TaskType taskType) |
|
int |
getTotalMaps() |
|
int |
getTotalReduces() |
|
java.lang.String |
getUserName() |
|
void |
handle(JobEvent event) |
|
static org.apache.hadoop.mapreduce.Counters |
incrTaskCounters(org.apache.hadoop.mapreduce.Counters counters,
java.util.Collection<Task> tasks) |
|
boolean |
isUber() |
|
org.apache.hadoop.conf.Configuration |
loadConfFile() |
|
protected void |
scheduleTasks(java.util.Set<org.apache.hadoop.mapreduce.v2.api.records.TaskId> taskIDs,
boolean recoverTaskOutput) |
|
void |
setJobPriority(org.apache.hadoop.yarn.api.records.Priority priority) |
|
void |
setQueueName(java.lang.String queueName) |
public static final java.lang.String JOB_KILLED_DIAG
public org.apache.hadoop.mapred.JobConf conf
public org.apache.hadoop.fs.Path remoteJobConfFile
protected static final org.apache.hadoop.yarn.state.StateMachineFactory<JobImpl,JobStateInternal,JobEventType,JobEvent> stateMachineFactory
public JobImpl(org.apache.hadoop.mapreduce.v2.api.records.JobId jobId,
org.apache.hadoop.yarn.api.records.ApplicationAttemptId applicationAttemptId,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.yarn.event.EventHandler eventHandler,
TaskAttemptListener taskAttemptListener,
org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager jobTokenSecretManager,
org.apache.hadoop.security.Credentials jobCredentials,
org.apache.hadoop.yarn.util.Clock clock,
java.util.Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo> completedTasksFromPreviousRun,
MRAppMetrics metrics,
org.apache.hadoop.mapreduce.OutputCommitter committer,
boolean newApiCommitter,
java.lang.String userName,
long appSubmitTime,
java.util.List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> amInfos,
AppContext appContext,
JobStateInternal forcedState,
java.lang.String forcedDiagnostic)
protected org.apache.hadoop.yarn.state.StateMachine<JobStateInternal,JobEventType,JobEvent> getStateMachine()
public org.apache.hadoop.mapreduce.v2.api.records.JobId getID()
public boolean checkAccess(org.apache.hadoop.security.UserGroupInformation callerUGI,
org.apache.hadoop.mapreduce.JobACL jobOperation)
checkAccess in interface Jobpublic Task getTask(org.apache.hadoop.mapreduce.v2.api.records.TaskId taskID)
public int getCompletedMaps()
getCompletedMaps in interface Jobpublic int getCompletedReduces()
getCompletedReduces in interface Jobpublic org.apache.hadoop.mapreduce.Counters getAllCounters()
JobgetAllCounters in interface Jobpublic static org.apache.hadoop.mapreduce.Counters incrTaskCounters(org.apache.hadoop.mapreduce.Counters counters,
java.util.Collection<Task> tasks)
public org.apache.hadoop.mapreduce.v2.api.records.TaskAttemptCompletionEvent[] getTaskAttemptCompletionEvents(int fromEventId,
int maxEvents)
getTaskAttemptCompletionEvents in interface Jobpublic org.apache.hadoop.mapred.TaskCompletionEvent[] getMapAttemptCompletionEvents(int startIndex,
int maxEvents)
getMapAttemptCompletionEvents in interface Jobpublic java.util.List<java.lang.String> getDiagnostics()
getDiagnostics in interface Jobpublic org.apache.hadoop.mapreduce.v2.api.records.JobReport getReport()
public float getProgress()
getProgress in interface Jobpublic java.util.Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> getTasks()
public java.util.Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> getTasks(org.apache.hadoop.mapreduce.v2.api.records.TaskType taskType)
public org.apache.hadoop.mapreduce.v2.api.records.JobState getState()
protected void scheduleTasks(java.util.Set<org.apache.hadoop.mapreduce.v2.api.records.TaskId> taskIDs,
boolean recoverTaskOutput)
public void handle(JobEvent event)
handle in interface org.apache.hadoop.yarn.event.EventHandler<JobEvent>@Private public JobStateInternal getInternalState()
protected void addTask(Task task)
protected org.apache.hadoop.fs.FileSystem getFileSystem(org.apache.hadoop.conf.Configuration conf)
throws java.io.IOException
conf - the conf objectjava.io.IOExceptionprotected JobStateInternal checkReadyForCommit()
public java.lang.String getUserName()
getUserName in interface Jobpublic java.lang.String getQueueName()
getQueueName in interface Jobpublic void setQueueName(java.lang.String queueName)
setQueueName in interface Jobpublic org.apache.hadoop.fs.Path getConfFile()
getConfFile in interface Jobpublic int getTotalMaps()
getTotalMaps in interface Jobpublic int getTotalReduces()
getTotalReduces in interface Jobpublic java.util.Map<org.apache.hadoop.mapreduce.JobACL,org.apache.hadoop.security.authorize.AccessControlList> getJobACLs()
getJobACLs in interface Jobpublic java.util.List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> getAMInfos()
getAMInfos in interface Jobpublic static java.lang.String escapeString(java.lang.String data)
@Private public void constructFinalFullcounters()
protected void addDiagnostic(java.lang.String diag)
public org.apache.hadoop.conf.Configuration loadConfFile()
throws java.io.IOException
loadConfFile in interface JobJob.getConfFile().java.io.IOException - on any error trying to load the conf file.public float getMaxAllowedFetchFailuresFraction()
public int getMaxFetchFailuresNotifications()
public void setJobPriority(org.apache.hadoop.yarn.api.records.Priority priority)
setJobPriority in interface Jobpublic int getFailedMaps()
getFailedMaps in interface Jobpublic int getFailedReduces()
getFailedReduces in interface Jobpublic int getKilledMaps()
getKilledMaps in interface Jobpublic int getKilledReduces()
getKilledReduces in interface JobCopyright © 2008–2025 Apache Software Foundation. All rights reserved.