安装Hadoop

本文详细介绍如何搭建Hadoop集群,包括SSH免密码登录配置、环境变量设置、各配置文件详解及节点间文件同步等步骤,并验证集群正确启动。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

1.设置ssh免密码登录。

2.设置Hadoop的环境变量

export HADOOP_HOME=/opt/op/hadoop-2.7.4
PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin:$HADOOP_HOME/bin
3.设置hadoop-env.sh

export JAVA_HOME=/opt/op/jdk1.8.0_144

4.设置core-site.xml

<configuration>
	<property>
         <name>hadoop.tmp.dir</name>
         <value>file:/data/hadoop/</value>
         <description>Abase for other temporary directories.</description>
    </property>
    <property>
         <name>fs.defaultFS</name>
         <value>hdfs://node5:9000</value>
    </property>
    <property>
			<name>dfs.permissions</name>
			<value>false</value>
			<description>
			If "true", enable permission checking in HDFS.
			If "false", permission checking is turned off,
			but all other behavior is unchanged.
			Switching from one parameter value to the other does not change the mode,
			owner or group of files or directories.
			</description>
		</property>
		<property>
		    <name>hadoop.proxyuser.root.hosts</name>
		    <value>*</value>
		</property>
		<property>
		    <name>hadoop.proxyuser.root.groups</name>
		    <value>*</value>
		</property>
</configuration>

5.设置hdfs-site.xml

<configuration>
	<property>    
      <name>dfs.namenode.secondary.http-address</name>    
      <value>node5:50090</value>    
	</property> 
	<property> 
		<name>dfs.webhdfs.enabled</name> 
		<value>true</value> 
		</property>
  <property>
       <name>dfs.replication</name>
       <value>1</value>
  </property>
  <property>
       <name>dfs.namenode.name.dir</name>
       <value>file:/data/hadoop/dfs/name</value>
  </property>
 
  <property>
       <name>dfs.datanode.data.dir</name>
       <value>file:/data/hadoop/dfs/data</value>
  </property>
  <property>
		<name>dfs.permissions</name>
		<value>false</value>
		<description>
		If "true", enable permission checking in HDFS.
		If "false", permission checking is turned off,
		but all other behavior is unchanged.
		Switching from one parameter value to the other does not change the mode,
		owner or group of files or directories.
		</description>
	</property>
</configuration>
6.设置mapred-site.xml

<configuration>
        <property>
             <name>mapreduce.framework.name</name>
             <value>yarn</value>
        </property>
</configuration>

7.设置slave

node5
node1
node2
node3
node4

8.我没有设置master

9.把node5(master)所有的文件复制到slave中

scp /opt/op/hadoop  node1:/opt/op/

10.设置各个slave的环境变量


export HADOOP_HOME=/opt/op/hadoop-2.7.4
PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin:$HADOOP_HOME/bin
运行命令

source /etc/profile

11.在master(node5)运行./start-all.sh命令

12.在master上运行jps命令查看是否包含以下进程

9236 ResourceManager
8788 NameNode
9732 Jps
8920 DataNode
9083 SecondaryNameNode
9341 NodeManager
13.在各个slave中jps查看是否包含以下进程


3190 DataNode
3432 Jps
3246 NodeManager


14.通过网页查看是否成功

http://node5:50070


评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值