1 package com.xiangmu_bushu 2 3 import org.apache.spark.rdd.RDD 4 import org.apache.spark.{SparkConf, SparkContext} 5 6 object Stream_WordCount { 7 def main(args: Array[String]): Unit = { 8 var conf = new SparkConf().setAppName("Stream_WordCount").setMaster("local[2]") 9 //獲取SparkContext對象 10 var sc = new SparkContext(conf) 11 12 //1、讀取文件生成數據RDD 13 var dataRdd = sc.textFile("D:\\IDEA_Maven\\monthlycheck\\src\\main\\resources\\a.txt") 14 //切分 15 var flatRdd = dataRdd.flatMap(_.split(" ")) 16 //標記 17 //var mapRdd = flatRdd.map((_,1)) 18 val mapRdd: RDD[(String, Int)] = flatRdd.map((_, 1)) 19 //聚合 20 var reduceRdd = mapRdd.reduceByKey(_ + _) 21 //打印輸出 22 reduceRdd.saveAsTextFile("./result") 23 24 } 25 26 }