org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob Java Examples
The following examples show how to use
org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: JobControl.java From hadoop with Apache License 2.0 | 5 votes |
static ArrayList<Job> castToJobList(List<ControlledJob> cjobs) { ArrayList<Job> ret = new ArrayList<Job>(); for (ControlledJob job : cjobs) { ret.add((Job)job); } return ret; }
Example #2
Source File: ValueAggregatorJob.java From hadoop with Apache License 2.0 | 5 votes |
public static JobControl createValueAggregatorJobs(String args[], Class<? extends ValueAggregatorDescriptor>[] descriptors) throws IOException { JobControl theControl = new JobControl("ValueAggregatorJobs"); ArrayList<ControlledJob> dependingJobs = new ArrayList<ControlledJob>(); Configuration conf = new Configuration(); if (descriptors != null) { conf = setAggregatorDescriptors(descriptors); } Job job = createValueAggregatorJob(conf, args); ControlledJob cjob = new ControlledJob(job, dependingJobs); theControl.addJob(cjob); return theControl; }
Example #3
Source File: JobControl.java From big-c with Apache License 2.0 | 5 votes |
static ArrayList<Job> castToJobList(List<ControlledJob> cjobs) { ArrayList<Job> ret = new ArrayList<Job>(); for (ControlledJob job : cjobs) { ret.add((Job)job); } return ret; }
Example #4
Source File: ValueAggregatorJob.java From big-c with Apache License 2.0 | 5 votes |
public static JobControl createValueAggregatorJobs(String args[], Class<? extends ValueAggregatorDescriptor>[] descriptors) throws IOException { JobControl theControl = new JobControl("ValueAggregatorJobs"); ArrayList<ControlledJob> dependingJobs = new ArrayList<ControlledJob>(); Configuration conf = new Configuration(); if (descriptors != null) { conf = setAggregatorDescriptors(descriptors); } Job job = createValueAggregatorJob(conf, args); ControlledJob cjob = new ControlledJob(job, dependingJobs); theControl.addJob(cjob); return theControl; }
Example #5
Source File: PigJobControl.java From spork with Apache License 2.0 | 5 votes |
private State checkState(ControlledJob j) { try { return (State)checkState.invoke(j); } catch (Exception e) { throw new RuntimeException(e); } }
Example #6
Source File: PigJobControl.java From spork with Apache License 2.0 | 5 votes |
private State submit(ControlledJob j) { try { return (State)submit.invoke(j); } catch (Exception e) { throw new RuntimeException(e); } }
Example #7
Source File: PigJobControl.java From spork with Apache License 2.0 | 5 votes |
@SuppressWarnings("unchecked") private LinkedList<ControlledJob> getJobs(Field field) { try { return (LinkedList<ControlledJob>)field.get(this); } catch (Exception e) { throw new RuntimeException(e); } }
Example #8
Source File: JobControlRun.java From MapReduce-Demo with MIT License | 4 votes |
public static void main(String[] args) throws IOException { String namenode_ip = "192.168.17.10"; String hdfs = "hdfs://" + namenode_ip + ":9000"; Configuration conf = new Configuration(); conf.set("fs.defaultFS", hdfs); conf.set("mapreduce.app-submission.cross-platform", "true"); Job job1 = Job.getInstance(conf,"Decompose"); job1.setJarByClass(JobControlRun.class); job1.setJar("export\\mutualFriend.jar"); job1.setMapperClass(DecomposeFriendsMapper.class); job1.setReducerClass(DecomposeFriendsReducer.class); job1.setOutputKeyClass(Text.class); job1.setOutputValueClass(Text.class); Path input = new Path(hdfs+"/workspace/mutualFriends/data"); Path output1 = new Path(hdfs+"/workspace/mutualFriends/output_Dec"); FileInputFormat.addInputPath(job1, input); FileOutputFormat.setOutputPath(job1, output1); FileSystem fs = FileSystem.get(conf); if (fs.exists(output1)) { fs.delete(output1, true); System.out.println("我被删了");// 打印可见只被删了一次,有点怪 } // ControlledJob作业控制容器 ControlledJob ctrJob1=new ControlledJob(conf); ctrJob1.setJob(job1);// job1加入控制容器 Job job2 = Job.getInstance(conf, "Merge"); job2.setJarByClass(JobControlRun.class); job2.setJar("export\\mutualFriend.jar"); job2.setMapperClass(MergeFriendsMapper.class); job2.setReducerClass(MergeFriendsReducer.class); job2.setOutputKeyClass(Text.class); job2.setOutputValueClass(Text.class); Path input2 = new Path(hdfs+"/workspace/mutualFriends/output_Dec"); Path output2 = new Path(hdfs+"/workspace/mutualFriends/output_Meg"); FileInputFormat.addInputPath(job2, input2); FileOutputFormat.setOutputPath(job2, output2); if (fs.exists(output2)) { fs.delete(output2, true); } ControlledJob ctrJob2 = new ControlledJob(conf); ctrJob2.setJob(job2);// job2加入作业控制容器 // 添加作业依赖,表明job2依赖job1执行 ctrJob2.addDependingJob(ctrJob1); // 定义作业主控制容器,监控、调度job1,job2 JobControl jobControl=new JobControl("JobControl"); jobControl.addJob(ctrJob1); jobControl.addJob(ctrJob2); // 启动作业线程 Thread T=new Thread(jobControl); T.start(); while(true){ if(jobControl.allFinished()){// 等待作业全部结束 System.out.println(jobControl.getSuccessfulJobList());// 打印成功job信息 jobControl.stop(); break; } } /** * 打印控制信息如下 * [job name: Decompose job id: JobControl0 job state: SUCCESS job mapred id: job_local445604445_0001 job message: just initialized job has no depending job: , job name: Merge job id: JobControl1 job state: SUCCESS job mapred id: job_local1897659504_0002 job message: just initialized job has 1 dependeng jobs: depending job 0: Decompose ] */ }
Example #9
Source File: Job.java From hadoop with Apache License 2.0 | 4 votes |
/** * Construct a job. * @param jobConf a mapred job configuration representing a job to be executed. * @param dependingJobs an array of jobs the current job depends on */ @SuppressWarnings("unchecked") public Job(JobConf jobConf, ArrayList<?> dependingJobs) throws IOException { super(org.apache.hadoop.mapreduce.Job.getInstance(jobConf), (List<ControlledJob>) dependingJobs); }
Example #10
Source File: Job.java From big-c with Apache License 2.0 | 4 votes |
/** * Construct a job. * @param jobConf a mapred job configuration representing a job to be executed. * @param dependingJobs an array of jobs the current job depends on */ @SuppressWarnings("unchecked") public Job(JobConf jobConf, ArrayList<?> dependingJobs) throws IOException { super(org.apache.hadoop.mapreduce.Job.getInstance(jobConf), (List<ControlledJob>) dependingJobs); }