Uses of Interface
org.apache.hadoop.mapred.Reporter

Packages that use Reporter
org.apache.hadoop.contrib.index.example   
org.apache.hadoop.contrib.index.mapred   
org.apache.hadoop.contrib.utils.join   
org.apache.hadoop.examples Hadoop example code. 
org.apache.hadoop.examples.dancing This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop. 
org.apache.hadoop.mapred A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. 
org.apache.hadoop.mapred.join Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map. 
org.apache.hadoop.mapred.lib Library of generally useful mappers, reducers, and partitioners. 
org.apache.hadoop.mapred.lib.aggregate Classes for performing various counting and aggregations. 
org.apache.hadoop.streaming Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g. 
 

Uses of Reporter in org.apache.hadoop.contrib.index.example
 

Methods in org.apache.hadoop.contrib.index.example with parameters of type Reporter
 RecordReader<DocumentID,LineDocTextAndOp> LineDocInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 void IdentityLocalAnalysis.map(DocumentID key, DocumentAndOp value, OutputCollector<DocumentID,DocumentAndOp> output, Reporter reporter)
           
 void LineDocLocalAnalysis.map(DocumentID key, LineDocTextAndOp value, OutputCollector<DocumentID,DocumentAndOp> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.contrib.index.mapred
 

Methods in org.apache.hadoop.contrib.index.mapred with parameters of type Reporter
 void IndexUpdateMapper.map(K key, V value, OutputCollector<Shard,IntermediateForm> output, Reporter reporter)
          Map a key-value pair to a shard-and-intermediate form pair.
 void IndexUpdateCombiner.reduce(Shard key, Iterator<IntermediateForm> values, OutputCollector<Shard,IntermediateForm> output, Reporter reporter)
           
 void IndexUpdateReducer.reduce(Shard key, Iterator<IntermediateForm> values, OutputCollector<Shard,Text> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.contrib.utils.join
 

Fields in org.apache.hadoop.contrib.utils.join declared as Reporter
protected  Reporter DataJoinReducerBase.reporter
           
protected  Reporter DataJoinMapperBase.reporter
           
 

Methods in org.apache.hadoop.contrib.utils.join with parameters of type Reporter
protected  void DataJoinReducerBase.collect(Object key, TaggedMapOutput aRecord, OutputCollector output, Reporter reporter)
          The subclass can overwrite this method to perform additional filtering and/or other processing logic before a value is collected.
 void DataJoinReducerBase.map(Object arg0, Object arg1, OutputCollector arg2, Reporter arg3)
           
 void DataJoinMapperBase.map(Object key, Object value, OutputCollector output, Reporter reporter)
           
 void DataJoinReducerBase.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)
           
 void DataJoinMapperBase.reduce(Object arg0, Iterator arg1, OutputCollector arg2, Reporter arg3)
           
 

Uses of Reporter in org.apache.hadoop.examples
 

Methods in org.apache.hadoop.examples with parameters of type Reporter
 RecordReader<MultiFileWordCount.WordOffset,Text> MultiFileWordCount.MyInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 void SleepJob.map(IntWritable key, IntWritable value, OutputCollector<IntWritable,IntWritable> output, Reporter reporter)
           
 void WordCount.MapClass.map(LongWritable key, Text value, OutputCollector<Text,IntWritable> output, Reporter reporter)
           
 void PiEstimator.PiMapper.map(LongWritable key, Writable val, OutputCollector<LongWritable,LongWritable> out, Reporter reporter)
          Map method.
 void MultiFileWordCount.MapClass.map(MultiFileWordCount.WordOffset key, Text value, OutputCollector<Text,IntWritable> output, Reporter reporter)
           
 void SleepJob.reduce(IntWritable key, Iterator<IntWritable> values, OutputCollector<IntWritable,IntWritable> output, Reporter reporter)
           
 void PiEstimator.PiReducer.reduce(LongWritable key, Iterator<LongWritable> values, OutputCollector<WritableComparable,Writable> output, Reporter reporter)
          Reduce method.
 void WordCount.Reduce.reduce(Text key, Iterator<IntWritable> values, OutputCollector<Text,IntWritable> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.examples.dancing
 

Methods in org.apache.hadoop.examples.dancing with parameters of type Reporter
 void DistributedPentomino.PentMap.map(WritableComparable key, Text value, OutputCollector<Text,Text> output, Reporter reporter)
          Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order).
 

Uses of Reporter in org.apache.hadoop.mapred
 

Fields in org.apache.hadoop.mapred declared as Reporter
static Reporter Reporter.NULL
          A constant of Reporter type that does nothing.
 

Methods in org.apache.hadoop.mapred with parameters of type Reporter
 void TextOutputFormat.LineRecordWriter.close(Reporter reporter)
           
 void RecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 RecordReader<LongWritable,Text> TextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 RecordReader<K,V> SequenceFileInputFilter.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Create a record reader for the given split
 RecordReader<Text,Text> SequenceFileAsTextInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<K,V> SequenceFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<BytesWritable,BytesWritable> SequenceFileAsBinaryInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
abstract  RecordReader<K,V> MultiFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<Text,Text> KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
abstract  RecordReader<K,V> FileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<K,V> InputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Get the RecordReader for the given InputSplit.
 void Mapper.map(K1 key, V1 value, OutputCollector<K2,V2> output, Reporter reporter)
          Maps a single input key/value pair into an intermediate key/value pair.
 void Reducer.reduce(K2 key, Iterator<V2> values, OutputCollector<K3,V3> output, Reporter reporter)
          Reduces values for a given key.
 void MapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 void MapRunnable.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
          Start mapping input <key, value> pairs.
 

Uses of Reporter in org.apache.hadoop.mapred.join
 

Methods in org.apache.hadoop.mapred.join with parameters of type Reporter
 ComposableRecordReader<K,TupleWritable> CompositeInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.
 ComposableRecordReader<K,V> ComposableInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.mapred.lib
 

Methods in org.apache.hadoop.mapred.lib with parameters of type Reporter
 RecordReader<LongWritable,Text> NLineInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 void TokenCountMapper.map(K key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
           
 void RegexMapper.map(K key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
           
 void IdentityMapper.map(K key, V val, OutputCollector<K,V> output, Reporter reporter)
          The identify function.
 void FieldSelectionMapReduce.map(K key, V val, OutputCollector<Text,Text> output, Reporter reporter)
          The identify function.
 void InverseMapper.map(K key, V value, OutputCollector<V,K> output, Reporter reporter)
          The inverse function.
 void LongSumReducer.reduce(K key, Iterator<LongWritable> values, OutputCollector<K,LongWritable> output, Reporter reporter)
           
 void IdentityReducer.reduce(K key, Iterator<V> values, OutputCollector<K,V> output, Reporter reporter)
          Writes all keys and values directly to output.
 void FieldSelectionMapReduce.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
           
 void MultithreadedMapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.mapred.lib.aggregate
 

Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Reporter
 void ValueAggregatorReducer.map(K1 arg0, V1 arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorMapper.map(K1 key, V1 value, OutputCollector<Text,Text> output, Reporter reporter)
          the map function.
 void ValueAggregatorCombiner.map(K1 arg0, V1 arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorReducer.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
           
 void ValueAggregatorMapper.reduce(Text arg0, Iterator<Text> arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorCombiner.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
          Combines values for a given key.
 

Uses of Reporter in org.apache.hadoop.streaming
 

Methods in org.apache.hadoop.streaming with parameters of type Reporter
 RecordReader<Text,Text> StreamInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 void PipeMapper.map(Object key, Object value, OutputCollector output, Reporter reporter)
           
 void PipeReducer.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)
           
 

Constructors in org.apache.hadoop.streaming with parameters of type Reporter
StreamBaseRecordReader(FSDataInputStream in, FileSplit split, Reporter reporter, JobConf job, FileSystem fs)
           
StreamXmlRecordReader(FSDataInputStream in, FileSplit split, Reporter reporter, JobConf job, FileSystem fs)
           
 



Copyright © 2008 The Apache Software Foundation