-------------------------------------------------------------
前言:
前面配置看,Spark安装配置01:https://blog.youkuaiyun.com/With__Sunshine/article/details/88622291
-------------------------------------------------------------
1.修改/soft/spark/conf/spark-defaults.conf.templete 名称
cd /soft/spark/conf/
mv spark-defaults.conf.templete spark-defaults.conf
2.配置spark-defaults.conf文件
vim park-defaults.conf
内容:
spark.eventLog.enabled true
spark.eventLog.dir hdfs://s101:9000/directory
spark.eventLog.compress true
注意:HDFS上的目录需要提前存在。
3.修改spark-env.sh,添加如下配置:
export SPARK_HISTORY_OPTS="-Dspark.history.ui.port=4000
-Dspark.history.retainedApplications=3
-Dspark.history.fs.logDirectory=hdfs://master01:9000/directory"
4.分发文件到s102、s103、s10
cd /soft/spark/conf
scp -r spark-defaults.conf centos@102:/soft/spark/conf
scp -r spark-defaults.conf centos@103:/soft/spark/conf
scp -r spark-defaults.conf centos@104:/soft/spark/conf
scp -r spark-env.sh centos@102:/soft/spark/conf
scp -r spark-env.sh centos@103:/soft/spark/conf
scp -r spark-env.sh centos@104:/soft/spark/conf
5.启动历史服务
先启动spark hdfs:
/soft/spark/sbin/start-all.sh
再启动,start-history-server.sh
/soft/spark/sbin/sstart-history-server.sh
6.如果遇到Hadoop HDFS的写入权限问题:
org.apache.hadoop.security.AccessControlException
解决方案: 在hdfs-site.xml中添加如下配置,关闭权限验证
<property>
<name>dfs.permissions</name>
<value>false</value>
</property>