Streaming Big Data: Storm, Spark and Samza | Blog T
In Storm, you design a graph of real-time computation called a topology, and feed it to the cluster where the master node will distribute the code among worker nodes to execute it. In a topology, data is passed around between spouts that emit data streams as immutable sets of key-value pairs called tuples, and bolts that transform those streams (count, filter etc.). Bolts themselves can optionally emit data to other bolts down the processing pipeline.
Read full article from Streaming Big Data: Storm, Spark and Samza | Blog T
No comments:
Post a Comment