There are three distinct layers to Storm's codebase.
First, Storm was designed from the very beginning to be compatible with multiple languages. Nimbus is a Thrift service and topologies are defined as Thrift structures. The usage of Thrift allows Storm to be used from any language.
Second, all of Storm's interfaces are specified as Java interfaces. So even though there's a lot of Clojure in Storm's implementation, all usage must go through the Java API. This means that every feature of Storm is always available via Java.
Third, Storm's implementation is largely in Clojure. Line-wise, Storm is about half Java code, half Clojure code. But Clojure is much more expressive, so in reality the great majority of the implementation logic is in Clojure.
The following sections explain each of these layers in more detail.
The first place to look to understand the structure of Storm's codebase is the storm.thrift file.
Storm uses this fork of Thrift (branch 'storm') to produce the generated code. This "fork" is actually Thrift 7 with all the Java packages renamed to be
org.apache.thrift7. Otherwise, it's identical to Thrift 7. This fork was done because of the lack of backwards compatibility in Thrift and the need for many people to use other versions of Thrift in their Storm topologies.
Every spout or bolt in a topology is given a user-specified identifier called the "component id". The component id is used to specify subscriptions from a bolt to the output streams of other spouts or bolts. A StormTopology structure contains a map from component id to component for each type of component (spouts and bolts).
Spouts and bolts have the same Thrift definition, so let's just take a look at the Thrift definition for bolts. It contains a
ComponentObject struct and a
ComponentObject defines the implementation for the bolt. It can be one of three types:
ShellComponentobject that indicates the implementation is in another language. Specifying a bolt this way will cause Storm to instantiate a ShellBolt object to handle the communication between the JVM-based worker process and the non-JVM-based implementation of the component.
JavaObjectstructure which tells Storm the classname and constructor arguments to use to instantiate that bolt. This is useful if you want to define a topology in a non-JVM language. This way, you can make use of JVM-based spouts and bolts without having to create and serialize a Java object yourself.
ComponentCommon defines everything else for this component. This includes:
Note that the structure spouts also have a
ComponentCommon field, and so spouts can also have declarations to consume other input streams. Yet the Storm Java API does not provide a way for spouts to consume other streams, and if you put any input declarations there for a spout you would get an error when you tried to submit the topology. The reason that spouts have an input declarations field is not for users to use, but for Storm itself to use. Storm adds implicit streams and bolts to the topology to set up the acking framework, and two of these implicit streams are from the acker bolt to each spout in the topology. The acker sends "ack" or "fail" messages along these streams whenever a tuple tree is detected to be completed or failed. The code that transforms the user's topology into the runtime topology is located here.
The interfaces for Storm are generally specified as Java interfaces. The main interfaces are:
The strategy for the majority of the interfaces is to:
You can see this strategy at work with the BaseRichSpout class.
Spouts and bolts are serialized into the Thrift definition of the topology as described above.
One subtle aspect of the interfaces is the difference between
IRichSpout. The main difference between them is the addition of the
declareOutputFields method in the "Rich" versions of the interfaces. The reason for the split is that the output fields declaration for each output stream needs to be part of the Thrift struct (so it can be specified from any language), but as a user you want to be able to declare the streams as part of your class. What
TopologyBuilder does when constructing the Thrift representation is call
declareOutputFields to get the declaration and convert it into the Thrift structure. The conversion happens at this portion of the
Specifying all the functionality via Java interfaces ensures that every feature of Storm is available via Java. Moreso, the focus on Java interfaces ensures that the user experience from Java-land is pleasant as well.
The implementation of Storm, on the other hand, is primarily in Clojure. While the codebase is about 50% Java and 50% Clojure in terms of LOC, most of the implementation logic is in Clojure. There are two notable exceptions to this, and that is the DRPC and transactional topologies implementations. These are implemented purely in Java. This was done to serve as an illustration for how to implement a higher level abstraction on Storm. The DRPC and transactional topologies implementations are in the org.apache.storm.coordination, org.apache.storm.drpc, and org.apache.storm.transactional packages.
Here's a summary of the purpose of the main Java packages and Clojure namespace:
org.apache.storm.coordination: Implements the pieces required to coordinate batch-processing on top of Storm, which both DRPC and transactional topologies use.
CoordinatedBolt is the most important class here.
org.apache.storm.drpc: Implementation of the DRPC higher level abstraction
org.apache.storm.generated: The generated Thrift code for Storm (generated using this fork of Thrift, which simply renames the packages to org.apache.thrift7 to avoid conflicts with other Thrift versions)
org.apache.storm.grouping: Contains interface for making custom stream groupings
org.apache.storm.spout: Definition of spout and associated interfaces (like the
SpoutOutputCollector). Also contains
ShellSpout which implements the protocol for defining spouts in non-JVM languages.
org.apache.storm.task: Definition of bolt and associated interfaces (like
OutputCollector). Also contains
ShellBolt which implements the protocol for defining bolts in non-JVM languages. Finally,
TopologyContext is defined here as well, which is provided to spouts and bolts so they can get data about the topology and its execution at runtime.
org.apache.storm.testing: Contains a variety of test bolts and utilities used in Storm's unit tests.
org.apache.storm.topology: Java layer over the underlying Thrift structure to provide a clean, pure-Java API to Storm (users don't have to know about Thrift).
TopologyBuilder is here as well as the helpful base classes for the different spouts and bolts. The slightly-higher level
IBasicBolt interface is here, which is a simpler way to write certain kinds of bolts.
org.apache.storm.transactional: Implementation of transactional topologies.
org.apache.storm.tuple: Implementation of Storm's tuple data model.
org.apache.storm.utils: Data structures and miscellaneous utilities used throughout the codebase.
org.apache.storm.command.*: These implement various commands for the
storm command line client. These implementations are very short.
org.apache.storm.cluster: This code manages how cluster state (like what tasks are running where, what spout/bolt each task runs as) is stored, typically in Zookeeper.
org.apache.storm.daemon.Acker: Implementation of the "acker" bolt, which is a key part of how Storm guarantees data processing.
org.apache.storm.daemon.DrpcServer: Implementation of the DRPC server for use with DRPC topologies.
org.apache.storm.event: Implements a simple asynchronous function executor. Used in various places in Nimbus and Supervisor to make functions execute in serial to avoid any race conditions.
org.apache.storm.messaging.*: Defines a higher level interface to implementing point to point messaging. In local mode Storm uses in-memory Java queues to do this; on a cluster, it uses Netty, but it is pluggable.
org.apache.storm.stats: Implementation of stats rollup routines used when sending stats to ZK for use by the UI. Does things like windowed and rolling aggregations at multiple granularities.
org.apache.storm.Thrift: Wrappers around the generated Thrift API to make working with Thrift structures more pleasant.
org.apache.storm.StormTimer: Implementation of a background timer to execute functions in the future or on a recurring interval. Storm couldn't use the Timer class because it needed integration with time simulation in order to be able to unit test Nimbus and the Supervisor.
org.apache.storm.clojure: Implementation of the Clojure DSL for Storm.
org.apache.storm.daemon.common: Implementation of common functions used in Storm daemons, like getting the id for a topology based on the name, mapping a user's topology into the one that actually executes (with implicit acking streams and acker bolt added - see
system-topology! function), and definitions for the various heartbeat and other structures persisted by Storm.
org.apache.storm.daemon.nimbus: Implementation of Nimbus.
org.apache.storm.daemon.supervisor: Implementation of Supervisor.
org.apache.storm.daemon.task: Implementation of an individual task for a spout or bolt. Handles message routing, serialization, stats collection for the UI, as well as the spout-specific and bolt-specific execution implementations.
org.apache.storm.daemon.worker: Implementation of a worker process (which will contain many tasks within). Implements message transferring and task launching.
org.apache.storm.log: Defines the functions used to log messages to log4j.
org.apache.storm.testing: Implementation of facilities used to test Storm topologies. Includes time simulation,
complete-topology for running a fixed set of tuples through a topology and capturing the output, tracker topologies for having fine grained control over detecting when a cluster is "idle", and other utilities.
org.apache.storm.ui.*: Implementation of Storm UI. Completely independent from rest of code base and uses the Nimbus Thrift API to get data.