K
- The type of the Trevni key to read.V
- The type of the Trevni value to read.
A subset schema to be read may be specified with
AvroJob.setInputKeySchema(org.apache.hadoop.mapreduce.Job, org.apache.avro.Schema)
and
AvroJob.setInputValueSchema(org.apache.hadoop.mapreduce.Job, org.apache.avro.Schema)
.
public class AvroTrevniKeyValueInputFormat<K,V> extends FileInputFormat<AvroKey<K>,AvroValue<V>>
InputFormat
for Trevni files.
This implement was modeled off
AvroKeyValueInputFormat
to allow for easy
transition
A MapReduce InputFormat that reads from Trevni container files of key/value generic records.
Trevni container files that container generic records with the two fields 'key' and 'value' are expected. The contents of the 'key' field will be used as the job input key, and the contents of the 'value' field will be used as the job output value.
FileInputFormat.Counter
DEFAULT_LIST_STATUS_NUM_THREADS, INPUT_DIR, INPUT_DIR_NONRECURSIVE_IGNORE_SUBDIRS, INPUT_DIR_RECURSIVE, LIST_STATUS_NUM_THREADS, NUM_INPUT_FILES, PATHFILTER_CLASS, SPLIT_MAXSIZE, SPLIT_MINSIZE
Constructor and Description |
---|
AvroTrevniKeyValueInputFormat() |
Modifier and Type | Method and Description |
---|---|
RecordReader<AvroKey<K>,AvroValue<V>> |
createRecordReader(InputSplit split,
TaskAttemptContext context) |
addInputPath, addInputPathRecursively, addInputPaths, computeSplitSize, getBlockIndex, getFormatMinSplitSize, getInputDirRecursive, getInputPathFilter, getInputPaths, getMaxSplitSize, getMinSplitSize, getSplits, isSplitable, listStatus, makeSplit, makeSplit, setInputDirRecursive, setInputPathFilter, setInputPaths, setInputPaths, setMaxInputSplitSize, setMinInputSplitSize
public RecordReader<AvroKey<K>,AvroValue<V>> createRecordReader(InputSplit split, TaskAttemptContext context) throws IOException, InterruptedException
createRecordReader
in class InputFormat<AvroKey<K>,AvroValue<V>>
IOException
InterruptedException
Copyright © 2009–2022 The Apache Software Foundation. All rights reserved.