val file = sc.textFile("hdfs://datanode3:8020/test/word.txt")
val count = file.flatMap(line => line.split(",")).map(word => (word, 1)).reduceByKey(_+_)
count.collect()
本文介绍如何使用Apache Spark实现WordCount统计任务。通过读取HDFS上的文本文件,利用Spark的flatMap、map及reduceByKey等操作完成单词计数。
val file = sc.textFile("hdfs://datanode3:8020/test/word.txt")
val count = file.flatMap(line => line.split(",")).map(word => (word, 1)).reduceByKey(_+_)
count.collect()

被折叠的 条评论
为什么被折叠?