| Modifier and Type | Class and Description |
|---|---|
static class |
JobImpl.InitTransition |
static class |
JobImpl.StartTransition |
| Modifier and Type | Field and Description |
|---|---|
org.apache.hadoop.mapred.JobConf |
conf |
static String |
JOB_KILLED_DIAG |
org.apache.hadoop.fs.Path |
remoteJobConfFile |
protected static org.apache.hadoop.yarn.state.StateMachineFactory<JobImpl,JobStateInternal,JobEventType,JobEvent> |
stateMachineFactory |
| Constructor and Description |
|---|
JobImpl(org.apache.hadoop.mapreduce.v2.api.records.JobId jobId,
org.apache.hadoop.yarn.api.records.ApplicationAttemptId applicationAttemptId,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.yarn.event.EventHandler eventHandler,
TaskAttemptListener taskAttemptListener,
org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager jobTokenSecretManager,
org.apache.hadoop.security.Credentials jobCredentials,
org.apache.hadoop.yarn.util.Clock clock,
Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo> completedTasksFromPreviousRun,
MRAppMetrics metrics,
org.apache.hadoop.mapreduce.OutputCommitter committer,
boolean newApiCommitter,
String userName,
long appSubmitTime,
List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> amInfos,
AppContext appContext,
JobStateInternal forcedState,
String forcedDiagnostic) |
| Modifier and Type | Method and Description |
|---|---|
protected void |
addDiagnostic(String diag) |
protected void |
addTask(Task task) |
boolean |
checkAccess(org.apache.hadoop.security.UserGroupInformation callerUGI,
org.apache.hadoop.mapreduce.JobACL jobOperation) |
protected JobStateInternal |
checkReadyForCommit() |
void |
constructFinalFullcounters() |
static String |
escapeString(String data) |
org.apache.hadoop.mapreduce.Counters |
getAllCounters()
Get all the counters of this job.
|
List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> |
getAMInfos() |
int |
getCompletedMaps() |
int |
getCompletedReduces() |
org.apache.hadoop.fs.Path |
getConfFile() |
List<String> |
getDiagnostics() |
int |
getFailedMaps() |
int |
getFailedReduces() |
protected org.apache.hadoop.fs.FileSystem |
getFileSystem(org.apache.hadoop.conf.Configuration conf)
Create the default file System for this job.
|
org.apache.hadoop.mapreduce.v2.api.records.JobId |
getID() |
JobStateInternal |
getInternalState() |
Map<org.apache.hadoop.mapreduce.JobACL,org.apache.hadoop.security.authorize.AccessControlList> |
getJobACLs() |
int |
getKilledMaps() |
int |
getKilledReduces() |
org.apache.hadoop.mapred.TaskCompletionEvent[] |
getMapAttemptCompletionEvents(int startIndex,
int maxEvents) |
float |
getMaxAllowedFetchFailuresFraction() |
int |
getMaxFetchFailuresNotifications() |
String |
getName() |
float |
getProgress() |
String |
getQueueName() |
org.apache.hadoop.mapreduce.v2.api.records.JobReport |
getReport() |
org.apache.hadoop.mapreduce.v2.api.records.JobState |
getState() |
protected org.apache.hadoop.yarn.state.StateMachine<JobStateInternal,JobEventType,JobEvent> |
getStateMachine() |
Task |
getTask(org.apache.hadoop.mapreduce.v2.api.records.TaskId taskID) |
org.apache.hadoop.mapreduce.v2.api.records.TaskAttemptCompletionEvent[] |
getTaskAttemptCompletionEvents(int fromEventId,
int maxEvents) |
Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> |
getTasks() |
Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> |
getTasks(org.apache.hadoop.mapreduce.v2.api.records.TaskType taskType) |
int |
getTotalMaps() |
int |
getTotalReduces() |
String |
getUserName() |
void |
handle(JobEvent event) |
static org.apache.hadoop.mapreduce.Counters |
incrTaskCounters(org.apache.hadoop.mapreduce.Counters counters,
Collection<Task> tasks) |
boolean |
isUber() |
org.apache.hadoop.conf.Configuration |
loadConfFile() |
protected void |
scheduleTasks(Set<org.apache.hadoop.mapreduce.v2.api.records.TaskId> taskIDs,
boolean recoverTaskOutput) |
void |
setJobPriority(org.apache.hadoop.yarn.api.records.Priority priority) |
void |
setQueueName(String queueName) |
public static final String JOB_KILLED_DIAG
public org.apache.hadoop.mapred.JobConf conf
public org.apache.hadoop.fs.Path remoteJobConfFile
protected static final org.apache.hadoop.yarn.state.StateMachineFactory<JobImpl,JobStateInternal,JobEventType,JobEvent> stateMachineFactory
public JobImpl(org.apache.hadoop.mapreduce.v2.api.records.JobId jobId,
org.apache.hadoop.yarn.api.records.ApplicationAttemptId applicationAttemptId,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.yarn.event.EventHandler eventHandler,
TaskAttemptListener taskAttemptListener,
org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager jobTokenSecretManager,
org.apache.hadoop.security.Credentials jobCredentials,
org.apache.hadoop.yarn.util.Clock clock,
Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo> completedTasksFromPreviousRun,
MRAppMetrics metrics,
org.apache.hadoop.mapreduce.OutputCommitter committer,
boolean newApiCommitter,
String userName,
long appSubmitTime,
List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> amInfos,
AppContext appContext,
JobStateInternal forcedState,
String forcedDiagnostic)
protected org.apache.hadoop.yarn.state.StateMachine<JobStateInternal,JobEventType,JobEvent> getStateMachine()
public org.apache.hadoop.mapreduce.v2.api.records.JobId getID()
public boolean checkAccess(org.apache.hadoop.security.UserGroupInformation callerUGI,
org.apache.hadoop.mapreduce.JobACL jobOperation)
checkAccess in interface Jobpublic Task getTask(org.apache.hadoop.mapreduce.v2.api.records.TaskId taskID)
public int getCompletedMaps()
getCompletedMaps in interface Jobpublic int getCompletedReduces()
getCompletedReduces in interface Jobpublic org.apache.hadoop.mapreduce.Counters getAllCounters()
JobgetAllCounters in interface Jobpublic static org.apache.hadoop.mapreduce.Counters incrTaskCounters(org.apache.hadoop.mapreduce.Counters counters,
Collection<Task> tasks)
public org.apache.hadoop.mapreduce.v2.api.records.TaskAttemptCompletionEvent[] getTaskAttemptCompletionEvents(int fromEventId,
int maxEvents)
getTaskAttemptCompletionEvents in interface Jobpublic org.apache.hadoop.mapred.TaskCompletionEvent[] getMapAttemptCompletionEvents(int startIndex,
int maxEvents)
getMapAttemptCompletionEvents in interface Jobpublic List<String> getDiagnostics()
getDiagnostics in interface Jobpublic org.apache.hadoop.mapreduce.v2.api.records.JobReport getReport()
public float getProgress()
getProgress in interface Jobpublic Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> getTasks(org.apache.hadoop.mapreduce.v2.api.records.TaskType taskType)
public org.apache.hadoop.mapreduce.v2.api.records.JobState getState()
protected void scheduleTasks(Set<org.apache.hadoop.mapreduce.v2.api.records.TaskId> taskIDs, boolean recoverTaskOutput)
public void handle(JobEvent event)
handle in interface org.apache.hadoop.yarn.event.EventHandler<JobEvent>@InterfaceAudience.Private public JobStateInternal getInternalState()
protected void addTask(Task task)
protected org.apache.hadoop.fs.FileSystem getFileSystem(org.apache.hadoop.conf.Configuration conf)
throws IOException
conf - the conf objectIOExceptionprotected JobStateInternal checkReadyForCommit()
public String getUserName()
getUserName in interface Jobpublic String getQueueName()
getQueueName in interface Jobpublic void setQueueName(String queueName)
setQueueName in interface Jobpublic org.apache.hadoop.fs.Path getConfFile()
getConfFile in interface Jobpublic int getTotalMaps()
getTotalMaps in interface Jobpublic int getTotalReduces()
getTotalReduces in interface Jobpublic Map<org.apache.hadoop.mapreduce.JobACL,org.apache.hadoop.security.authorize.AccessControlList> getJobACLs()
getJobACLs in interface Jobpublic List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> getAMInfos()
getAMInfos in interface Job@InterfaceAudience.Private public void constructFinalFullcounters()
protected void addDiagnostic(String diag)
public org.apache.hadoop.conf.Configuration loadConfFile()
throws IOException
loadConfFile in interface JobJob.getConfFile().IOException - on any error trying to load the conf file.public float getMaxAllowedFetchFailuresFraction()
public int getMaxFetchFailuresNotifications()
public void setJobPriority(org.apache.hadoop.yarn.api.records.Priority priority)
setJobPriority in interface Jobpublic int getFailedMaps()
getFailedMaps in interface Jobpublic int getFailedReduces()
getFailedReduces in interface Jobpublic int getKilledMaps()
getKilledMaps in interface Jobpublic int getKilledReduces()
getKilledReduces in interface JobCopyright © 2008–2023 Apache Software Foundation. All rights reserved.