|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use Text | |
---|---|
org.apache.hadoop.contrib.index.example | |
org.apache.hadoop.contrib.index.mapred | |
org.apache.hadoop.contrib.utils.join | |
org.apache.hadoop.examples | Hadoop example code. |
org.apache.hadoop.examples.dancing | This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop. |
org.apache.hadoop.io | Generic i/o code for use when reading and writing data to the network, to databases, and to files. |
org.apache.hadoop.mapred | A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. |
org.apache.hadoop.mapred.lib | Library of generally useful mappers, reducers, and partitioners. |
org.apache.hadoop.mapred.lib.aggregate | Classes for performing various counting and aggregations. |
org.apache.hadoop.streaming | Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g. |
Uses of Text in org.apache.hadoop.contrib.index.example |
---|
Methods in org.apache.hadoop.contrib.index.example that return Text | |
---|---|
Text |
LineDocTextAndOp.getText()
Get the text that represents a document. |
Uses of Text in org.apache.hadoop.contrib.index.mapred |
---|
Methods in org.apache.hadoop.contrib.index.mapred that return Text | |
---|---|
Text |
DocumentID.getText()
The text of the document id. |
Methods in org.apache.hadoop.contrib.index.mapred that return types with arguments of type Text | |
---|---|
RecordWriter<Shard,Text> |
IndexUpdateOutputFormat.getRecordWriter(FileSystem fs,
JobConf job,
String name,
Progressable progress)
|
Method parameters in org.apache.hadoop.contrib.index.mapred with type arguments of type Text | |
---|---|
void |
IndexUpdateReducer.reduce(Shard key,
Iterator<IntermediateForm> values,
OutputCollector<Shard,Text> output,
Reporter reporter)
|
Uses of Text in org.apache.hadoop.contrib.utils.join |
---|
Fields in org.apache.hadoop.contrib.utils.join declared as Text | |
---|---|
protected Text |
DataJoinMapperBase.inputTag
|
static Text |
DataJoinReducerBase.NUM_OF_VALUES_FIELD
|
static Text |
DataJoinReducerBase.SOURCE_TAGS_FIELD
|
protected Text |
TaggedMapOutput.tag
|
Methods in org.apache.hadoop.contrib.utils.join that return Text | |
---|---|
protected abstract Text |
DataJoinMapperBase.generateGroupKey(TaggedMapOutput aRecord)
Generate a map output key. |
protected abstract Text |
DataJoinMapperBase.generateInputTag(String inputFile)
Determine the source tag based on the input file name. |
Text |
TaggedMapOutput.getTag()
|
Methods in org.apache.hadoop.contrib.utils.join with parameters of type Text | |
---|---|
void |
TaggedMapOutput.setTag(Text tag)
|
Uses of Text in org.apache.hadoop.examples |
---|
Methods in org.apache.hadoop.examples that return Text | |
---|---|
Text |
MultiFileWordCount.MultiFileLineRecordReader.createValue()
|
Methods in org.apache.hadoop.examples that return types with arguments of type Text | |
---|---|
ArrayList<Map.Entry<Text,Text>> |
AggregateWordCount.WordCountPlugInClass.generateKeyValPairs(Object key,
Object val)
|
ArrayList<Map.Entry<Text,Text>> |
AggregateWordCount.WordCountPlugInClass.generateKeyValPairs(Object key,
Object val)
|
ArrayList<Map.Entry<Text,Text>> |
AggregateWordHistogram.AggregateWordHistogramPlugin.generateKeyValPairs(Object key,
Object val)
Parse the given value, generate an aggregation-id/value pair per word. |
ArrayList<Map.Entry<Text,Text>> |
AggregateWordHistogram.AggregateWordHistogramPlugin.generateKeyValPairs(Object key,
Object val)
Parse the given value, generate an aggregation-id/value pair per word. |
RecordReader<MultiFileWordCount.WordOffset,Text> |
MultiFileWordCount.MyInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
Methods in org.apache.hadoop.examples with parameters of type Text | |
---|---|
void |
WordCount.MapClass.map(LongWritable key,
Text value,
OutputCollector<Text,IntWritable> output,
Reporter reporter)
|
void |
MultiFileWordCount.MapClass.map(MultiFileWordCount.WordOffset key,
Text value,
OutputCollector<Text,IntWritable> output,
Reporter reporter)
|
boolean |
MultiFileWordCount.MultiFileLineRecordReader.next(MultiFileWordCount.WordOffset key,
Text value)
|
void |
WordCount.Reduce.reduce(Text key,
Iterator<IntWritable> values,
OutputCollector<Text,IntWritable> output,
Reporter reporter)
|
Method parameters in org.apache.hadoop.examples with type arguments of type Text | |
---|---|
void |
WordCount.MapClass.map(LongWritable key,
Text value,
OutputCollector<Text,IntWritable> output,
Reporter reporter)
|
void |
MultiFileWordCount.MapClass.map(MultiFileWordCount.WordOffset key,
Text value,
OutputCollector<Text,IntWritable> output,
Reporter reporter)
|
void |
WordCount.Reduce.reduce(Text key,
Iterator<IntWritable> values,
OutputCollector<Text,IntWritable> output,
Reporter reporter)
|
Uses of Text in org.apache.hadoop.examples.dancing |
---|
Methods in org.apache.hadoop.examples.dancing with parameters of type Text | |
---|---|
void |
DistributedPentomino.PentMap.map(WritableComparable key,
Text value,
OutputCollector<Text,Text> output,
Reporter reporter)
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order). |
Method parameters in org.apache.hadoop.examples.dancing with type arguments of type Text | |
---|---|
void |
DistributedPentomino.PentMap.map(WritableComparable key,
Text value,
OutputCollector<Text,Text> output,
Reporter reporter)
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order). |
void |
DistributedPentomino.PentMap.map(WritableComparable key,
Text value,
OutputCollector<Text,Text> output,
Reporter reporter)
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order). |
Uses of Text in org.apache.hadoop.io |
---|
Methods in org.apache.hadoop.io that return Text | |
---|---|
Text |
SequenceFile.Metadata.get(Text name)
|
Methods in org.apache.hadoop.io that return types with arguments of type Text | |
---|---|
TreeMap<Text,Text> |
SequenceFile.Metadata.getMetadata()
|
TreeMap<Text,Text> |
SequenceFile.Metadata.getMetadata()
|
Methods in org.apache.hadoop.io with parameters of type Text | |
---|---|
Text |
SequenceFile.Metadata.get(Text name)
|
void |
Text.set(Text other)
copy a text. |
void |
SequenceFile.Metadata.set(Text name,
Text value)
|
Constructors in org.apache.hadoop.io with parameters of type Text | |
---|---|
Text(Text utf8)
Construct from another text. |
Constructor parameters in org.apache.hadoop.io with type arguments of type Text | |
---|---|
SequenceFile.Metadata(TreeMap<Text,Text> arg)
|
|
SequenceFile.Metadata(TreeMap<Text,Text> arg)
|
Uses of Text in org.apache.hadoop.mapred |
---|
Methods in org.apache.hadoop.mapred that return Text | |
---|---|
Text |
SequenceFileAsTextRecordReader.createKey()
|
Text |
KeyValueLineRecordReader.createKey()
|
Text |
SequenceFileAsTextRecordReader.createValue()
|
Text |
LineRecordReader.createValue()
|
Text |
KeyValueLineRecordReader.createValue()
|
Methods in org.apache.hadoop.mapred that return types with arguments of type Text | |
---|---|
RecordReader<LongWritable,Text> |
TextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
Methods in org.apache.hadoop.mapred with parameters of type Text | |
---|---|
boolean |
LineRecordReader.next(LongWritable key,
Text value)
Read a line. |
boolean |
SequenceFileAsTextRecordReader.next(Text key,
Text value)
Read key/value pair in a line. |
boolean |
KeyValueLineRecordReader.next(Text key,
Text value)
Read key/value pair in a line. |
int |
LineRecordReader.LineReader.readLine(Text str)
Read from the InputStream into the given Text. |
int |
LineRecordReader.LineReader.readLine(Text str,
int maxLineLength)
Read from the InputStream into the given Text. |
int |
LineRecordReader.LineReader.readLine(Text str,
int maxLineLength,
int maxBytesToConsume)
Read from the InputStream into the given Text. |
Uses of Text in org.apache.hadoop.mapred.lib |
---|
Methods in org.apache.hadoop.mapred.lib that return types with arguments of type Text | |
---|---|
RecordReader<LongWritable,Text> |
NLineInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
Methods in org.apache.hadoop.mapred.lib with parameters of type Text | |
---|---|
void |
TokenCountMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
RegexMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Method parameters in org.apache.hadoop.mapred.lib with type arguments of type Text | |
---|---|
void |
TokenCountMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
RegexMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
FieldSelectionMapReduce.map(K key,
V val,
OutputCollector<Text,Text> output,
Reporter reporter)
The identify function. |
void |
FieldSelectionMapReduce.map(K key,
V val,
OutputCollector<Text,Text> output,
Reporter reporter)
The identify function. |
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Uses of Text in org.apache.hadoop.mapred.lib.aggregate |
---|
Fields in org.apache.hadoop.mapred.lib.aggregate declared as Text | |
---|---|
static Text |
ValueAggregatorDescriptor.ONE
|
Methods in org.apache.hadoop.mapred.lib.aggregate that return types with arguments of type Text | |
---|---|
static Map.Entry<Text,Text> |
ValueAggregatorBaseDescriptor.generateEntry(String type,
String id,
Text val)
|
static Map.Entry<Text,Text> |
ValueAggregatorBaseDescriptor.generateEntry(String type,
String id,
Text val)
|
ArrayList<Map.Entry<Text,Text>> |
ValueAggregatorBaseDescriptor.generateKeyValPairs(Object key,
Object val)
Generate 1 or 2 aggregation-id/value pairs for the given key/value pair. |
ArrayList<Map.Entry<Text,Text>> |
ValueAggregatorBaseDescriptor.generateKeyValPairs(Object key,
Object val)
Generate 1 or 2 aggregation-id/value pairs for the given key/value pair. |
ArrayList<Map.Entry<Text,Text>> |
ValueAggregatorDescriptor.generateKeyValPairs(Object key,
Object val)
Generate a list of aggregation-id/value pairs for the given key/value pair. |
ArrayList<Map.Entry<Text,Text>> |
ValueAggregatorDescriptor.generateKeyValPairs(Object key,
Object val)
Generate a list of aggregation-id/value pairs for the given key/value pair. |
ArrayList<Map.Entry<Text,Text>> |
UserDefinedValueAggregatorDescriptor.generateKeyValPairs(Object key,
Object val)
Generate a list of aggregation-id/value pairs for the given key/value pairs by delegating the invocation to the real object. |
ArrayList<Map.Entry<Text,Text>> |
UserDefinedValueAggregatorDescriptor.generateKeyValPairs(Object key,
Object val)
Generate a list of aggregation-id/value pairs for the given key/value pairs by delegating the invocation to the real object. |
Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Text | |
---|---|
static Map.Entry<Text,Text> |
ValueAggregatorBaseDescriptor.generateEntry(String type,
String id,
Text val)
|
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
Method parameters in org.apache.hadoop.mapred.lib.aggregate with type arguments of type Text | |
---|---|
void |
ValueAggregatorReducer.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorReducer.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.map(K1 key,
V1 value,
OutputCollector<Text,Text> output,
Reporter reporter)
the map function. |
void |
ValueAggregatorMapper.map(K1 key,
V1 value,
OutputCollector<Text,Text> output,
Reporter reporter)
the map function. |
void |
ValueAggregatorCombiner.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorCombiner.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
Uses of Text in org.apache.hadoop.streaming |
---|
Methods in org.apache.hadoop.streaming that return Text | |
---|---|
Text |
StreamBaseRecordReader.createKey()
|
Text |
StreamBaseRecordReader.createValue()
|
Methods in org.apache.hadoop.streaming that return types with arguments of type Text | |
---|---|
RecordReader<Text,Text> |
StreamInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
StreamInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
Methods in org.apache.hadoop.streaming with parameters of type Text | |
---|---|
boolean |
StreamXmlRecordReader.next(Text key,
Text value)
|
abstract boolean |
StreamBaseRecordReader.next(Text key,
Text value)
Read a record. |
static int |
UTF8ByteArrayUtils.readLine(LineRecordReader.LineReader lineReader,
Text out)
Read a utf8 encoded line from a data input stream. |
static void |
UTF8ByteArrayUtils.splitKeyVal(byte[] utf,
int start,
int length,
Text key,
Text val,
int splitPos)
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos. |
static void |
UTF8ByteArrayUtils.splitKeyVal(byte[] utf,
Text key,
Text val,
int splitPos)
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |