spark实例1---WordCount

import org.apache.spark.rdd.RDD
import org.apache.spark.{SparkConf, SparkContext}

object WordCount extends App {
  private val conf = new SparkConf()
  conf.setMaster("local").setAppName("wc")

  private val sc: SparkContext = new SparkContext(conf)
  private val line: RDD[String] = sc.textFile("data/wc.txt")
  private val word: RDD[String] = line.flatMap(x => x.split(" "))
  private val wordAnd1: RDD[(String, Int)] = word.map(x => (x, 1))
  private val result: RDD[(String, Int)] = wordAnd1.reduceByKey((x, y) => x + y)
  result.foreach(println)

  sc.stop()
}
posted @ 2022-04-09 21:15  jsqup  阅读(23)  评论(0编辑  收藏  举报