public class AccumuloFileOutputFormat extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat<Key,Value>
Key), as this is an important requirement of Accumulo data files.
 
 The output path to be created must be specified via FileOutputFormat.setOutputPath(Job, Path). This is inherited from
 FileOutputFormat.setOutputPath(Job, Path). Other methods from FileOutputFormat are not supported and may be ignored or cause failures. Using
 other Hadoop configuration options that affect the behavior of the underlying files directly in the Job's configuration may work, but are not directly
 supported at this time.
| Modifier and Type | Field and Description | 
|---|---|
| protected static org.apache.log4j.Logger | log | 
| Constructor and Description | 
|---|
| AccumuloFileOutputFormat() | 
| Modifier and Type | Method and Description | 
|---|---|
| protected static org.apache.accumulo.core.conf.AccumuloConfiguration | getAccumuloConfiguration(org.apache.hadoop.mapreduce.JobContext context)This helper method provides an AccumuloConfiguration object constructed from the Accumulo defaults, and overridden with Accumulo properties that have been
 stored in the Job's configuration. | 
| protected static Instance | getInstance(org.apache.hadoop.conf.Configuration conf)Deprecated. 
 since 1.5.0; This  OutputFormatdoes not communicate with Accumulo. If this is needed, subclasses must implement their own
             configuration. | 
| org.apache.hadoop.mapreduce.RecordWriter<Key,Value> | getRecordWriter(org.apache.hadoop.mapreduce.TaskAttemptContext context) | 
| protected static void | handleBlockSize(org.apache.hadoop.conf.Configuration conf)Deprecated. 
 since 1.5.0; Retrieve the relevant block size from  getAccumuloConfiguration(JobContext)and configure hadoop's
             io.seqfile.compress.blocksize with the same value. No longer needed, asRFiledoes not use this field. | 
| static void | setBlockSize(org.apache.hadoop.conf.Configuration conf,
            int blockSize)Deprecated. 
 since 1.5.0; Use  setFileBlockSize(Job, long),setDataBlockSize(Job, long), orsetIndexBlockSize(Job, long)instead. | 
| static void | setCompressionType(org.apache.hadoop.mapreduce.Job job,
                  String compressionType)Sets the compression type to use for data blocks. | 
| static void | setDataBlockSize(org.apache.hadoop.mapreduce.Job job,
                long dataBlockSize)Sets the size for data blocks within each file. Data blocks are a span of key/value pairs stored in the file that are compressed and indexed as a group. | 
| static void | setFileBlockSize(org.apache.hadoop.mapreduce.Job job,
                long fileBlockSize)Sets the size for file blocks in the file system; file blocks are managed, and replicated, by the underlying file system. | 
| static void | setFileType(org.apache.hadoop.conf.Configuration conf,
           String type)Deprecated. 
 since 1.5.0; This method does nothing. Only 'rf' type is supported. | 
| static void | setIndexBlockSize(org.apache.hadoop.mapreduce.Job job,
                 long indexBlockSize)Sets the size for index blocks within each file; smaller blocks means a deeper index hierarchy within the file, while larger blocks mean a more shallow
 index hierarchy within the file. | 
| static void | setReplication(org.apache.hadoop.mapreduce.Job job,
              int replication)Sets the file system replication factor for the resulting file, overriding the file system default. | 
| static void | setZooKeeperInstance(org.apache.hadoop.conf.Configuration conf,
                    String instanceName,
                    String zooKeepers)Deprecated. 
 since 1.5.0; This  OutputFormatdoes not communicate with Accumulo. If this is needed, subclasses must implement their own
             configuration. | 
checkOutputSpecs, getCompressOutput, getDefaultWorkFile, getOutputCommitter, getOutputCompressorClass, getOutputName, getOutputPath, getPathForWorkFile, getUniqueFile, getWorkOutputPath, setCompressOutput, setOutputCompressorClass, setOutputName, setOutputPathprotected static org.apache.accumulo.core.conf.AccumuloConfiguration getAccumuloConfiguration(org.apache.hadoop.mapreduce.JobContext context)
context - the Hadoop context for the configured jobpublic static void setCompressionType(org.apache.hadoop.mapreduce.Job job,
                      String compressionType)
job - the Hadoop job instance to be configuredcompressionType - one of "none", "gz", "lzo", or "snappy"public static void setDataBlockSize(org.apache.hadoop.mapreduce.Job job,
                    long dataBlockSize)
Making this value smaller may increase seek performance, but at the cost of increasing the size of the indexes (which can also affect seek performance).
job - the Hadoop job instance to be configureddataBlockSize - the block size, in bytespublic static void setFileBlockSize(org.apache.hadoop.mapreduce.Job job,
                    long fileBlockSize)
job - the Hadoop job instance to be configuredfileBlockSize - the block size, in bytespublic static void setIndexBlockSize(org.apache.hadoop.mapreduce.Job job,
                     long indexBlockSize)
job - the Hadoop job instance to be configuredindexBlockSize - the block size, in bytespublic static void setReplication(org.apache.hadoop.mapreduce.Job job,
                  int replication)
job - the Hadoop job instance to be configuredreplication - the number of replicas for produced filespublic org.apache.hadoop.mapreduce.RecordWriter<Key,Value> getRecordWriter(org.apache.hadoop.mapreduce.TaskAttemptContext context) throws IOException
getRecordWriter in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat<Key,Value>IOException@Deprecated protected static void handleBlockSize(org.apache.hadoop.conf.Configuration conf)
getAccumuloConfiguration(JobContext) and configure hadoop's
             io.seqfile.compress.blocksize with the same value. No longer needed, as RFile does not use this field.@Deprecated public static void setFileType(org.apache.hadoop.conf.Configuration conf, String type)
@Deprecated public static void setBlockSize(org.apache.hadoop.conf.Configuration conf, int blockSize)
setFileBlockSize(Job, long), setDataBlockSize(Job, long), or setIndexBlockSize(Job, long) instead.@Deprecated public static void setZooKeeperInstance(org.apache.hadoop.conf.Configuration conf, String instanceName, String zooKeepers)
OutputFormat does not communicate with Accumulo. If this is needed, subclasses must implement their own
             configuration.@Deprecated protected static Instance getInstance(org.apache.hadoop.conf.Configuration conf)
OutputFormat does not communicate with Accumulo. If this is needed, subclasses must implement their own
             configuration.Copyright © 2011-2016 The Apache Software Foundation. All Rights Reserved.