首先下载:
http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.7.4/hadoop-2.7.4.tar.gz
解压:
tar -zxvf hadoop-2.7.4.tar.gz
拷贝到工作目录 :
mv hadoop.xxxxx /usr/local/hadoop
安装java 1.8
下载
http://download.oracle.com/otn-pub/java/jdk/8u144-b01/090f390dda5b47b9b721c7dfaa008135/jdk-8u144-linux-x64.tar.gz
解压
tar -zxvf jdk-8u144-linux-x64.tar.gz
拷贝至工作目录
mv jdk-8u144-linux-x64 /usr/local/java1.8
修改配制文件
vim /etc/profile
添加变量,java和hadoop 环境变量
JAVA_HOME=/usr/local/java1.8
JRE_HOME=/usr/local/java1.8/jre
HADOOP_HOME=/usr/local/hadoop
PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin
PATH=$PATH:$HADOOP_HOME/bin
export PATH
创建软连接
ln -s /usr/local/hadoop/bin/hdfs /usr/bin/hdfs
ln -s /usr/local/hadoop/bin/hdfs /usr/bin/hdfs
查看是否成功
3 java -version
java version "1.8.0_144"
Java(TM) SE Runtime Environment (build 1.8.0_144-b01)
Java HotSpot(TM) 64-Bit Server VM (build 25.144-b01, mixed mode)
验证hadoop
#hadoop version
Hadoop 2.7.4
Subversion https://shv@git-wip-us.apache.org/repos/asf/hadoop.git -r cd915e1e8d9d0131462a0b7301586c175728a282
Compiled by kshvachk on 2017-08-01T00:29Z
Compiled with protoc 2.5.0
From source with checksum 50b0468318b4ce9bd24dc467b7ce1148
This command was run using /usr/local/hadoop/share/hadoop/common/hadoop-common-2.7.4.jar
出现以上结果表示安装成功
配制hadoop 配制文件:
hadoop/etc/hadoop/hadoop-env.sh
hadoop/etc/hadoop/yarn-env.sh
hadoop/etc/hadoop/core-site.xml
hadoop/etc/hadoop/hdfs-site.xml
hadoop/etc/hadoop/mapred-site.xml
hadoop/etc/hadoop/yarn-site.xml
首先
配置hadoop-env.sh
2)配置yarn-env.sh
3)配置core-site.xml
添加如下配置:
注意创建tmp目录 !!!!!!!!!!!!!!!!!!
--添加权限
#bin/hdfs dfs -chmod -R 777 /tmp
4),配置hdfs-site.xml
添加如下配置
5),配置mapred-site.xml
添加如下配置:
6),配置yarn-site.xml
添加如下配置:
Hadoop启动
1)格式化namenode
2)启动NameNode 和 DataNode 守护进程
3)启动ResourceManager 和 NodeManager 守护进程
启动验证
1)执行jps命令,有如下进程,说明Hadoop正常启动
如果想做成完全分布式需要做以下工作:
所有节点机器ssh免密登陆,请自己去百度。
在第一台机器slave 配制文件中。添加上所有节点名称
添加所有机器名至所有节点 的/etc/hosts文件。
把第一台机器 的hadoop目录 ,复制到所有节点。
完成。
二 spark 安装
下载:
https://www.apache.org/dyn/closer.lua/spark/spark-2.2.0/spark-2.2.0-bin-hadoop2.7.tgz
解压:
tar -zxvf spark-2.2.0-bin-hadoop2.7.tgz
拷贝
mv spark-2.2.0-bin-hadoop2.7 /usr/local/spark
修改配制文件:
cd /usr/local/spark/
cp
./conf/spark-env.sh.template
./conf/spark-env.sh
#vim spark-env.sh
添加一行
export
SPARK_DIST_CLASSPATH=$(/usr/local/hadoop/bin/hadoop classpath)
测试spark
./bin/run-example SparkPi 2>&1 | grep "Pi is roughly"
结果:
Pi is roughly 3.1423757118785596