集群修改
1.修改配置文件conf/spark-env.sh
export JAVA_HOME=/root/apps/jdk1.8.0_201
#export SPARK_MASTER_HOST=hdp-1
#export SPARK_MASTER_PORT=7077
export SPARK_DAEMON_JAVA_OPTS="-Dspark.deploy.recoveryMode=ZOOKEEPER -Dspark.deploy.zookeeper.url=hdp-1,hdp-2,hdp-3 -Dspark.deploy.zookeeper.dir=/spark"
export SPARK_WORKER_CORES=8
export SPARK_WORKER_MEMORY=6g
export SPARK_LOCAL_IP="192.168.81.129"
conf/slaves
hdp-2
hdp-3
hdp-4
将配置好的spark路径整体分发到集群中其他几台机器:
scp -r spark-2.2.0-bin-hadoop2.7 -C hdp-2:$PWD
启动:
1、启动zk
2、启动hdfs start-dfs.sh
3、启动spark sbin/start-all.sh
验证:
http://hdp-1:8080
高可用集群
在worker机器上启动start-master