public class HadoopDruidIndexerConfig extends Object
Modifier and Type | Class and Description |
---|---|
static class |
HadoopDruidIndexerConfig.IndexJobCounters |
Modifier and Type | Field and Description |
---|---|
static String |
CONFIG_PROPERTY |
static Charset |
javaNativeCharset |
static com.fasterxml.jackson.databind.ObjectMapper |
jsonMapper |
protected static Properties |
properties |
static com.google.common.base.Joiner |
tabJoiner |
static com.google.common.base.Splitter |
tabSplitter |
Constructor and Description |
---|
HadoopDruidIndexerConfig(HadoopIngestionSpec spec) |
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapreduce.Job |
addInputPaths(org.apache.hadoop.mapreduce.Job job) |
void |
addJobProperties(org.apache.hadoop.mapreduce.Job job) |
static HadoopDruidIndexerConfig |
fromConfiguration(org.apache.hadoop.conf.Configuration conf) |
static HadoopDruidIndexerConfig |
fromFile(File file) |
static HadoopDruidIndexerConfig |
fromMap(Map<String,Object> argSpec) |
static HadoopDruidIndexerConfig |
fromSpec(HadoopIngestionSpec spec) |
static HadoopDruidIndexerConfig |
fromString(String str) |
com.google.common.base.Optional<Iterable<Bucket>> |
getAllBuckets() |
com.google.common.base.Optional<Bucket> |
getBucket(io.druid.data.input.InputRow inputRow)
Get the proper bucket for some input row.
|
String |
getDataSource() |
GranularitySpec |
getGranularitySpec() |
com.google.common.base.Optional<List<org.joda.time.Interval>> |
getIntervals() |
long |
getMaxPartitionSize() |
io.druid.data.input.impl.StringInputRowParser |
getParser() |
PartitionsSpec |
getPartitionsSpec() |
HadoopIngestionSpec |
getSchema() |
com.google.common.base.Optional<Set<org.joda.time.Interval>> |
getSegmentGranularIntervals() |
HadoopyShardSpec |
getShardSpec(Bucket bucket) |
Long |
getTargetPartitionSize() |
String |
getWorkingPath() |
void |
intoConfiguration(org.apache.hadoop.mapreduce.Job job) |
boolean |
isCombineText() |
boolean |
isDeterminingPartitions() |
boolean |
isIgnoreInvalidRows() |
boolean |
isOverwriteFiles() |
boolean |
isPersistInHeap() |
boolean |
isUpdaterJobSpecSet() |
org.apache.hadoop.fs.Path |
makeDescriptorInfoDir() |
org.apache.hadoop.fs.Path |
makeDescriptorInfoPath(io.druid.timeline.DataSegment segment) |
org.apache.hadoop.fs.Path |
makeGroupedDataDir() |
org.apache.hadoop.fs.Path |
makeIntermediatePath()
Make the intermediate path for this job run.
|
org.apache.hadoop.fs.Path |
makeIntervalInfoPath() |
org.apache.hadoop.fs.Path |
makeSegmentOutputPath(org.apache.hadoop.fs.FileSystem fileSystem,
Bucket bucket) |
org.apache.hadoop.fs.Path |
makeSegmentPartitionInfoPath(org.joda.time.Interval bucketInterval) |
void |
setGranularitySpec(GranularitySpec granularitySpec) |
void |
setShardSpecs(Map<org.joda.time.DateTime,List<HadoopyShardSpec>> shardSpecs) |
void |
setVersion(String version) |
void |
verify() |
public static final String CONFIG_PROPERTY
public static final Charset javaNativeCharset
public static final com.google.common.base.Splitter tabSplitter
public static final com.google.common.base.Joiner tabJoiner
public static final com.fasterxml.jackson.databind.ObjectMapper jsonMapper
protected static final Properties properties
public HadoopDruidIndexerConfig(HadoopIngestionSpec spec)
public static HadoopDruidIndexerConfig fromSpec(HadoopIngestionSpec spec)
public static HadoopDruidIndexerConfig fromMap(Map<String,Object> argSpec)
public static HadoopDruidIndexerConfig fromFile(File file)
public static HadoopDruidIndexerConfig fromString(String str)
public static HadoopDruidIndexerConfig fromConfiguration(org.apache.hadoop.conf.Configuration conf)
public HadoopIngestionSpec getSchema()
public String getDataSource()
public GranularitySpec getGranularitySpec()
public void setGranularitySpec(GranularitySpec granularitySpec)
public PartitionsSpec getPartitionsSpec()
public boolean isOverwriteFiles()
public boolean isIgnoreInvalidRows()
public void setVersion(String version)
public void setShardSpecs(Map<org.joda.time.DateTime,List<HadoopyShardSpec>> shardSpecs)
public com.google.common.base.Optional<List<org.joda.time.Interval>> getIntervals()
public boolean isDeterminingPartitions()
public Long getTargetPartitionSize()
public long getMaxPartitionSize()
public boolean isUpdaterJobSpecSet()
public boolean isCombineText()
public io.druid.data.input.impl.StringInputRowParser getParser()
public HadoopyShardSpec getShardSpec(Bucket bucket)
public org.apache.hadoop.mapreduce.Job addInputPaths(org.apache.hadoop.mapreduce.Job job) throws IOException
IOException
public com.google.common.base.Optional<Bucket> getBucket(io.druid.data.input.InputRow inputRow)
inputRow
- an InputRowpublic com.google.common.base.Optional<Set<org.joda.time.Interval>> getSegmentGranularIntervals()
public boolean isPersistInHeap()
public String getWorkingPath()
public org.apache.hadoop.fs.Path makeIntermediatePath()
public org.apache.hadoop.fs.Path makeSegmentPartitionInfoPath(org.joda.time.Interval bucketInterval)
public org.apache.hadoop.fs.Path makeIntervalInfoPath()
public org.apache.hadoop.fs.Path makeDescriptorInfoDir()
public org.apache.hadoop.fs.Path makeGroupedDataDir()
public org.apache.hadoop.fs.Path makeDescriptorInfoPath(io.druid.timeline.DataSegment segment)
public org.apache.hadoop.fs.Path makeSegmentOutputPath(org.apache.hadoop.fs.FileSystem fileSystem, Bucket bucket)
public void addJobProperties(org.apache.hadoop.mapreduce.Job job)
public void intoConfiguration(org.apache.hadoop.mapreduce.Job job)
public void verify()
Copyright © 2015. All rights reserved.