public abstract class AbstractHdfsBolt extends BaseRichBolt
Modifier and Type | Field and Description |
---|---|
protected OutputCollector |
collector |
protected String |
configKey |
protected FileNameFormat |
fileNameFormat |
protected Integer |
fileRetryCount |
protected org.apache.hadoop.fs.FileSystem |
fs |
protected String |
fsUrl |
protected org.apache.hadoop.conf.Configuration |
hdfsConfig |
protected Integer |
maxOpenFiles |
protected long |
offset |
protected Partitioner |
partitioner |
protected List<RotationAction> |
rotationActions |
protected Map<String,Integer> |
rotationCounterMap |
protected FileRotationPolicy |
rotationPolicy |
protected Timer |
rotationTimer |
protected SyncPolicy |
syncPolicy |
protected Integer |
tickTupleInterval |
protected Object |
writeLock |
protected Map<String,Writer> |
writers |
Constructor and Description |
---|
AbstractHdfsBolt() |
Modifier and Type | Method and Description |
---|---|
void |
cleanup()
Called when an IBolt is going to be shutdown.
|
void |
declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer)
Declare the output schema for all the streams of this topology.
|
protected abstract void |
doPrepare(Map<String,Object> conf,
TopologyContext topologyContext,
OutputCollector collector) |
void |
execute(Tuple tuple)
Process a single tuple of input.
|
protected org.apache.hadoop.fs.Path |
getBasePathForNextFile(Tuple tuple) |
Map<String,Object> |
getComponentConfiguration()
Declare configuration specific to this component.
|
protected abstract String |
getWriterKey(Tuple tuple) |
protected abstract Writer |
makeNewWriter(org.apache.hadoop.fs.Path path,
Tuple tuple) |
void |
prepare(Map<String,Object> conf,
TopologyContext topologyContext,
OutputCollector collector)
Marked as final to prevent override.
|
protected void |
rotateOutputFile(Writer writer) |
protected List<RotationAction> rotationActions
protected OutputCollector collector
protected transient org.apache.hadoop.fs.FileSystem fs
protected SyncPolicy syncPolicy
protected FileRotationPolicy rotationPolicy
protected FileNameFormat fileNameFormat
protected String fsUrl
protected String configKey
protected transient Object writeLock
protected transient Timer rotationTimer
protected long offset
protected Integer fileRetryCount
protected Integer tickTupleInterval
protected Integer maxOpenFiles
protected Partitioner partitioner
protected transient org.apache.hadoop.conf.Configuration hdfsConfig
protected void rotateOutputFile(Writer writer) throws IOException
IOException
public final void prepare(Map<String,Object> conf, TopologyContext topologyContext, OutputCollector collector)
Marked as final to prevent override. Subclasses should implement the doPrepare() method.
conf
- The Storm configuration for this bolt. This is the configuration provided to the topology merged in with cluster configuration on this machine.topologyContext
- This object can be used to get information about this task’s place within the topology, including the task id and component id of this task, input and output information, etc.collector
- The collector is used to emit tuples from this bolt. Tuples can be emitted at any time, including the prepare and cleanup methods. The collector is thread-safe and should be saved as an instance variable of this bolt object.public final void execute(Tuple tuple)
IBolt
Process a single tuple of input. The Tuple object contains metadata on it about which component/stream/task it came from. The values of the Tuple can be accessed using Tuple#getValue. The IBolt does not have to process the Tuple immediately. It is perfectly fine to hang onto a tuple and process it later (for instance, to do an aggregation or join).
Tuples should be emitted using the OutputCollector provided through the prepare method. It is required that all input tuples are acked or failed at some point using the OutputCollector. Otherwise, Storm will be unable to determine when tuples coming off the spouts have been completed.
For the common case of acking an input tuple at the end of the execute method, see IBasicBolt which automates this.
tuple
- The input tuple to be processed.public Map<String,Object> getComponentConfiguration()
IComponent
Declare configuration specific to this component. Only a subset of the “topology.*” configs can be overridden. The component configuration can be further overridden when constructing the topology using TopologyBuilder
getComponentConfiguration
in interface IComponent
getComponentConfiguration
in class BaseComponent
public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer)
IComponent
Declare the output schema for all the streams of this topology.
outputFieldsDeclarer
- this is used to declare output stream ids, output fields, and whether or not each output stream is a direct streampublic void cleanup()
IBolt
Called when an IBolt is going to be shutdown. Storm will make a best-effort attempt to call this if the worker shutdown is orderly. The Config.SUPERVISOR_WORKER_SHUTDOWN_SLEEP_SECS
setting controls how long orderly shutdown is allowed to take. There is no guarantee that cleanup will be called if shutdown is not orderly, or if the shutdown exceeds the time limit.
The one context where cleanup is guaranteed to be called is when a topology is killed when running Storm in local mode.
cleanup
in interface IBolt
cleanup
in class BaseRichBolt
protected org.apache.hadoop.fs.Path getBasePathForNextFile(Tuple tuple)
protected abstract void doPrepare(Map<String,Object> conf, TopologyContext topologyContext, OutputCollector collector) throws IOException
IOException
protected abstract Writer makeNewWriter(org.apache.hadoop.fs.Path path, Tuple tuple) throws IOException
IOException
Copyright © 2022 The Apache Software Foundation. All rights reserved.