测试数据:
1 1603A 95
2 1603B 85
3 1603C 75
4 1603D 96
5 1604F 94
6 1604E 95
7 1604K 91
8 1604G 89
9 1501A 79
10 1502A 69
11 1503A 59
12 1504A 89
13 1701A 99
14 1702A 100
15 1703A 65
测试结果:
(1702A,100)
(1701A,99)
(1603D,96)
(1603A,95)
(1604E,95)
(1604F,94)
(1604K,91)
(1604G,89)
(1504A,89)
(1603B,85)
(1501A,79)
(1603C,75)
(1502A,69)
(1703A,65)
(1503A,59)
package com.spark.week3
import org.apache.spark.{SparkConf, SparkContext}
object Four {
System.setProperty("hadoop.home.dir","D:/soft/hadoop/hadoop-2.7.3")
def main(args: Array[String]): Unit = {
val conf = new SparkConf().setMaster("local").setAppName("Week4Class")
val df = new SparkContext(conf)
val fileRdd = df.textFile("G:/data/week3class.txt")
//方式一
// fileRdd.map(line => {
// val row= line.split(" ")
// (row(1),row(2).toInt)
//
// }).sortBy(line => line._2,false).foreach(println)
fileRdd.map(line=> (line.split(" ")(1),line.split(" ")(2).toInt)).sortBy(line=>line._2,false).foreach(println)
df.stop()
}
}