1,分布式单词计数的流程
首先要有数据源,在SentenceSpout中定义了一个字符串数组sentences来模拟数据源。字符串数组中的每句话作为一个tuple发射。其实,SplitBolt接收SentenceSpout发射的tuple,它将每句话分割成每个单词,并将每个单词作为tuple发射。再次,WordCountBolt接收SplitBolt发送的tuple,它将接收到的每一个单词统计计数,并将 <单词:出现次数> 作为tuple发射。最后,ReportBolt接收WordCountBolt发送的tuple,将统计的结果存入HashMap中,并打印出结果。
流程图如下:
2,Topology的组成类
ISpout、IComponent、IBolt三个接口定义了一些最基本的方法,BaseRichSpout、BaseRichBolt是接口的实现类,自定义的Spout与Bolt通过继承实现类来完成工作。
详细解释参考代码里面的注释。参考《Storm分布式实时计算模式》第一章中的例子。
1, SentenceSpou.java分析:
private String[] sentences = { "my dog has fleas", "i like cold beverages", "the dog ate my homework", "don't have a cow man", "i don't think i like fleas" };
定义了待发射的数据源。Spout从该字符串数组一次取一个字符串生成tuple进行发射。
32 public void open(@SuppressWarnings("rawtypes") Map conf, 33 TopologyContext context, SpoutOutputCollector collector) { 34 // TODO Auto-generated method stub 35 this.collector = collector; 36 }
open函数,在ISpout接口中定义,所有的Spout组件在初始化时调用这个方法。在open()中初始化了发射器。
55 public void declareOutputFields(OutputFieldsDeclarer declarer) { 56 // TODO Auto-generated method stub 57 declarer.declare(new Fields("sentence"));// 标记SentenceSpout发送的tuple的键为 58 // sentence 59 }
declareOutputFields函数标记了该Spout发射的tuple的(字段值)键值。下游的Bolt可以通过该键值来接收它发出的tuple
41 public void nextTuple() { 42 // TODO Auto-generated method stub 43 // 以字符串数组sentences 中的每个字符串 作为参数 构造tuple 44 this.collector.emit(new Values(sentences[index]));// 通过emit方法将构造好的tuple发送出去 45 index++; 46 if (index >= sentences.length) { 47 index = 0; 48 } 49 Utils.sleep(100); 50 }
nextTuple()是所有Spout的核心方法。Storm通过调用这个方法向collector发射tuple。Values.java 继承了ArrayList,new Values(...)构造了一个List对象,并将之作为emit的参数通过collector发射出去。
这里的发射规则是:每次发射其中一个字符串,阻塞100ms。当发射完整个字符串数组时,将索引(index)重新置0。可以继续发射。除非显示终止Topology,否则它不会停止。
SentenceSpou.java代码如下:
1 package org.apache.storm.storm_core; 2 3 import java.util.Map; 4 5 import backtype.storm.spout.SpoutOutputCollector; 6 import backtype.storm.task.TopologyContext; 7 import backtype.storm.topology.OutputFieldsDeclarer; 8 import backtype.storm.topology.base.BaseRichSpout; 9 import backtype.storm.tuple.Fields; 10 import backtype.storm.tuple.Values; 11 import backtype.storm.utils.Utils; 12 13 public class SentenceSpout extends BaseRichSpout { 14 /** 15 * 16 */ 17 private static final long serialVersionUID = 3444934973982660864L; 18 private SpoutOutputCollector collector;// 用来向其他Spout发射tuple 19 private String[] sentences = { "my dog has fleas", "i like cold beverages", 20 "the dog ate my homework", "don't have a cow man", 21 "i don't think i like fleas" }; 22 23 private int index = 0; 24 25 /* 26 * open() 方法在所有的Spout组件初始化时被调用 27 * 28 * @param Map conf storm 配置信息 29 * 30 * @context TopologyContext topology 组件信息 31 */ 32 public void open(@SuppressWarnings("rawtypes") Map conf, 33 TopologyContext context, SpoutOutputCollector collector) { 34 // TODO Auto-generated method stub 35 this.collector = collector; 36 } 37 38 /* 39 * Values.java extends ArrayList Storm 调用该方法向输出的collector发射tuple 40 */ 41 public void nextTuple() { 42 // TODO Auto-generated method stub 43 // 以字符串数组sentences 中的每个字符串 作为参数 构造tuple 44 this.collector.emit(new Values(sentences[index]));// 通过emit方法将构造好的tuple发送出去 45 index++; 46 if (index >= sentences.length) { 47 index = 0; 48 } 49 Utils.sleep(100); 50 } 51 52 /* 53 * SentenceSpout 发送的tuple它是一个包含键值对的List,该方法声明了List中包含的键值对的键为 sentence 54 */ 55 public void declareOutputFields(OutputFieldsDeclarer declarer) { 56 // TODO Auto-generated method stub 57 declarer.declare(new Fields("sentence"));// 标记SentenceSpout发送的tuple的键为 58 // sentence 59 } 60 }
SplitBolt.java代码如下:
1 package org.apache.storm.storm_core; 2 3 import java.util.Map; 4 5 import backtype.storm.task.OutputCollector; 6 import backtype.storm.task.TopologyContext; 7 import backtype.storm.topology.OutputFieldsDeclarer; 8 import backtype.storm.topology.base.BaseRichBolt; 9 import backtype.storm.tuple.Fields; 10 import backtype.storm.tuple.Tuple; 11 import backtype.storm.tuple.Values; 12 13 public class SplitSentenceBolt extends BaseRichBolt { 14 /** 15 * 16 */ 17 private static final long serialVersionUID = -2107029392155190729L; 18 private OutputCollector collector;// 用来向其他Spout发射tuple的发射器 19 20 /* 21 * (non-Javadoc) prepare方法类似于open方法,prepare在bolt初始化时被调用 22 */ 23 public void prepare(Map stormConf, TopologyContext context, 24 OutputCollector collector) { 25 // TODO Auto-generated method stub 26 this.collector = collector;// 发射器初始化 27 28 } 29 30 public void execute(Tuple input) { 31 // TODO Auto-generated method stub 32 // 接收从SentenceSpout的发射器发射过来的tuple,因为SentenceSpout中声明的tuple字段为sentence,故getStringByField方法的参数为sentence 33 String sentence = input.getStringByField("sentence");// 该tuple是一个包含 34 // 键为sentence 35 // 值为字符串 36 // 的列表List<Map<sentence,String>> 37 String[] words = sentence.split(" ");// 将字符串分解成一个个的单词 38 for (String word : words) 39 this.collector.emit(new Values(word));// 将每个单词构造成tuple并发送给下一个Spout 40 } 41 42 public void declareOutputFields(OutputFieldsDeclarer declarer) { 43 // TODO Auto-generated method stub 44 declarer.declare(new Fields("word"));// 定义SplitSentenceBolt发送的tuple的字段("键值")为 word 45 } 46 }
WordCountBolt.java
1 package org.apache.storm.storm_core; 2 3 import java.util.HashMap; 4 import java.util.Map; 5 6 import backtype.storm.task.OutputCollector; 7 import backtype.storm.task.TopologyContext; 8 import backtype.storm.topology.OutputFieldsDeclarer; 9 import backtype.storm.topology.base.BaseRichBolt; 10 import backtype.storm.tuple.Fields; 11 import backtype.storm.tuple.Tuple; 12 import backtype.storm.tuple.Values; 13 14 public class WordCountBolt extends BaseRichBolt{ 15 16 private OutputCollector collector; 17 private HashMap<String, Long>counts = null;//统计每个单词出现的次数,放到HashMap中保存起来 18 19 public void prepare(Map stormConf, TopologyContext context, 20 OutputCollector collector) { 21 // TODO Auto-generated method stub 22 this.collector = collector; 23 this.counts = new HashMap<String, Long>();//初始化HashMap,因为prepare会被自动调用的 24 } 25 26 public void execute(Tuple input) { 27 // TODO Auto-generated method stub 28 String word = input.getStringByField("word"); 29 Long count = this.counts.get(word); 30 if(count == null)//HashMap中没有word这个单词 31 count = 0L; 32 count++; 33 this.counts.put(word, count);//更新该单词在HashMap中的统计次数 34 //此处发射的tuple包含了两个元素:单词和计数,它每次发送的是一个长度为2的List, 35 //可理解为:List.add(new HashMap("word",word)); List.add(new HashMap(("count",count)); 36 this.collector.emit(new Values(word, count));//第一个元素的键为 "word",值为该单词(a string),第二个键为 "count",值为单词的计数 37 } 38 39 public void declareOutputFields(OutputFieldsDeclarer declarer) { 40 // TODO Auto-generated method stub 41 declarer.declare(new Fields("word", "count")); 42 } 43 }
ReportBolt.java如下:
package org.apache.storm.storm_core; import java.util.ArrayList; import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Map; import backtype.storm.task.OutputCollector; import backtype.storm.task.TopologyContext; import backtype.storm.topology.OutputFieldsDeclarer; import backtype.storm.topology.base.BaseRichBolt; import backtype.storm.tuple.Tuple; public class ReportBolt extends BaseRichBolt{ /** * */ private static final long serialVersionUID = 4921144902730095910L; // private OutputCollector collector; ReportBolt不需要发射tuple了 private HashMap<String, Long> counts = null; public void prepare(Map stormConf, TopologyContext context, OutputCollector collector) { // TODO Auto-generated method stub this.counts = new HashMap<String, Long>(); } public void execute(Tuple input) { // TODO Auto-generated method stub String word = input.getStringByField("word"); Long count = input.getLongByField("count"); this.counts.put(word, count); } public void declareOutputFields(OutputFieldsDeclarer declarer) { // TODO Auto-generated method stub //不需要发出任何数据流 } //Topology在storm集群中运行时,cleanup方法是不可靠的,并不能保证它一定会执行 public void cleanup(){ System.out.println("------ print counts ------"); List<String> keys = new ArrayList<String>(); keys.addAll(counts.keySet());//将HashMap中所有的键都添加到一个集合中 Collections.sort(keys);//对键(单词)进行排序 for(String key : keys)//输出排好序的每个单词的出现次数 System.out.println(key + " : " + this.counts.get(key)); System.out.println("--------bye----------"); } }
WordCountTopology.java如下:
1 package org.apache.storm.storm_core; 2 3 import backtype.storm.Config; 4 import backtype.storm.LocalCluster; 5 import backtype.storm.topology.TopologyBuilder; 6 import backtype.storm.tuple.Fields; 7 import backtype.storm.utils.Utils; 8 9 public class WordCountTopology { 10 private static final String SENTENCE_SPOUT_ID = "sentence-spout"; 11 private static final String SPLIT_BOLT_ID = "split-bolt"; 12 private static final String COUNT_BOLT_ID = "count-bolt"; 13 private static final String REPORT_BOLT_ID = "report-bolt"; 14 private static final String TOPOLOGY_NAME = "word-count-topology"; 15 16 public static void main(String[] args) throws Exception{ 17 SentenceSpout spout = new SentenceSpout(); 18 SplitSentenceBolt splitBolt = new SplitSentenceBolt(); 19 WordCountBolt countBolt = new WordCountBolt(); 20 ReportBolt reportBolt = new ReportBolt(); 21 22 TopologyBuilder builder = new TopologyBuilder(); 23 builder.setSpout(SENTENCE_SPOUT_ID, spout); 24 builder.setBolt(SPLIT_BOLT_ID, splitBolt).shuffleGrouping(SENTENCE_SPOUT_ID); 25 builder.setBolt(COUNT_BOLT_ID, countBolt).fieldsGrouping(SPLIT_BOLT_ID, new Fields("word")); 26 builder.setBolt(REPORT_BOLT_ID, reportBolt).globalGrouping(COUNT_BOLT_ID); 27 28 Config config = new Config(); 29 LocalCluster cluster = new LocalCluster(); 30 31 cluster.submitTopology(TOPOLOGY_NAME, config, builder.createTopology()); 32 Utils.sleep(1000); 33 cluster.killTopology(TOPOLOGY_NAME); 34 cluster.shutdown(); 35 36 } 37 }