package org.Program1
import org.apache.spark.sql.{DataFrame, SparkSession}
/**
* @ @description:
* @ author: T-WHONG
* @ create: 2019-04-12 09:49:17
**/
object ObtainDataTest {
def main(args: Array[String]): Unit = {
val sparkSession = SparkSession.builder()
.appName("ObtainDataTest")
.master("local[2]")
.getOrCreate()
val localPath = "file:///E:/Programs/DProgram1\\actionlogtest"
val hdfsPath = "hdfs://hadoop03:8020//program1/actionlogtest"
val parquetDtat: DataFrame = sparkSession.read.parquet(localPath)
parquetDtat.show(10)
println("<------------------------------------------------------>")
parquetDtat.createOrReplaceTempView("view")
val ret = sparkSession.sql("select * from view limit 10")
ret.show()
sparkSession.stop()
}
}
- 一般读取本机电脑磁盘文件(若默认为hdfs)则需要加file协议
file:///E:/aa/bb\cc - 若读取hdfs文件
hdfs://hadoop03:8020/aa/bb\cc - 除了协议后面跟的符号以外,在正式路径表示中\\等同于/