July 6, 2023
WordCount
String line=value.toString(); StringTokenizer token = new StringTokenizer(line); while(token.hasMoreTokens()) { word.set(token.nextToken()); out.collect(word, one);
int sum=0; while(value.hasNext()) { sum+=value.next().get(); } out.collect(key, new IntWritable(sum));
private final static IntWritable one = new IntWritable(1); private Text word = new Text();
public static void main(String[] args) throws IOException { // TODO Auto-generated method stub JobConf conf = new JobConf(WC.class); conf.setJobName("WC"); conf.setOutputKeyClass(Text.class) ; conf.setOutputValueClass(IntWritable.class); conf.setMapperClass(Map.class); conf.setCombinerClass(Reduce.class); conf.setReducerClass(Reduce.class); conf.setInputFormat(TextInputFormat.class); conf.setOutputFormat(TextOutputFormat.class); FileInputFormat.setInputPaths(conf, new Path(args[0])); FileOutputFormat.setOutputPath(conf,new Path(args[1])); JobClient.runJob(conf); } }
package test; import java.io.IOException; import java.util.Iterator; import java.util.StringTokenizer; import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.Text; import org.apache.hadoop.mapred.FileInputFormat; import org.apache.hadoop.mapred.FileOutputFormat; import org.apache.hadoop.mapred.JobClient; import org.apache.hadoop.mapred.JobConf; import org.apache.hadoop.mapred.MapReduceBase; import org.apache.hadoop.mapred.Mapper; import org.apache.hadoop.mapred.OutputCollector; import org.apache.hadoop.mapred.Reducer; import org.apache.hadoop.mapred.Reporter; import org.apache.hadoop.mapred.TextInputFormat; import org.apache.hadoop.mapred.TextOutputFormat; import org.apache.hadoop.fs.Path; public class WC { public static class Map extends MapReduceBase implements Mapper<LongWritable, Text, Text, IntWritable>{ private final static IntWritable one = new IntWritable(1); private Text word = new Text(); @Override public void map(LongWritable key, Text value, OutputCollector<Text, IntWritable> out, Reporter arg3) throws IOException { // TODO Auto-generated method stub String line=value.toString(); StringTokenizer token = new StringTokenizer(line); while(token.hasMoreTokens()) { word.set(token.nextToken()); out.collect(word, one); } } } public static class Reduce extends MapReduceBase implements Reducer<Text, IntWritable,Text, IntWritable>{ @Override public void reduce(Text key, Iterator<IntWritable> value, OutputCollector<Text, IntWritable> out, Reporter arg3) throws IOException { // TODO Auto-generated method stub int sum=0; while(value.hasNext()) { sum+=value.next().get(); } out.collect(key, new IntWritable(sum)); }} public static void main(String[] args) throws IOException { // TODO Auto-generated method stub JobConf conf = new JobConf(WC.class); conf.setJobName("WC"); conf.setOutputKeyClass(Text.class) ; conf.setOutputValueClass(IntWritable.class); conf.setMapperClass(Map.class); conf.setCombinerClass(Reduce.class); conf.setReducerClass(Reduce.class); conf.setInputFormat(TextInputFormat.class); conf.setOutputFormat(TextOutputFormat.class); FileInputFormat.setInputPaths(conf, new Path(args[0])); FileOutputFormat.setOutputPath(conf,new Path(args[1])); JobClient.runJob(conf); } }