hadoop 2.7.2使用maven编译安装

(1)依赖包:

libssl-dev   zlib1g-dev  openssl  protobuf2.5 maven  jdk1.7 cmake openssh-server

(2)

mvn  clean package -DskipTests -Pdist,native, docs

其中docs根据需要添加。


(3)配置文件:

   1.  core-site.xml:

<configuration>
       <configuration>
        <property>
                <name>fs.defaultFS</name>
                <value>hdfs://hostname:9000</value>
        </property>
        <property>
                <name>io.file.buffer.size</name>
                <value>131072</value>
        </property>
        <property>
                <name>hadoop.tmp.dir</name>
                <value>file:/home/hadoop/hadoop-2.5.1/temp</value>
                <description>Abase for other temporary directories.</description>
        </property>
        <property>
                <name>hadoop.proxyuser.hduser.hosts</name>
                <value>*</value>
        </property>
        <property>
                <name>hadoop.proxyuser.hduser.groups</name>
                <value>*</value>
        </property>
</configuration>

  2.hdfs-site.xml

 

<configuration>
        <property>
                <name>dfs.namenode.secondary.http-address</name>
                <value>hostname:9001</value>
        </property>
        <property>
                <name>dfs.namenode.name.dir</name>
                <value>file:/home/hadoop/hadoop-2.5.1/dfs/name</value>
        </property>
        <property>
                <name>dfs.datanode.data.dir</name>
                <value>file:/home/hadoop/hadoop-2.5.1/dfs/data</value>
        </property>
        <property>
                <name>dfs.replication</name>
                <value>1</value>
        </property>
        <property>
                <name>dfs.webhdfs.enabled</name>
                <value>true</value>
        </property>
</configuration>

  3.mapred-site.xml

<configuration>
        <property>
                <name>mapreduce.framework.name</name>
                <value>yarn</value>
        </property>
        <property>
                <name>mapreduce.jobhistory.address</name>
                <value>hostname:10020</value>
        </property>
        <property>
                <name>mapreduce.jobhistory.webapp.address</name>
                <value>hostname:19888</value>
        </property>
</configuration>

   4. yarn-site.xml

<configuration>
        <property>
                <name>yarn.nodemanager.aux-services</name>
                <value>mapreduce_shuffle</value>
        </property>
        <property>
                <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
                <value>org.apache.hadoop.mapred.ShuffleHandler</value>
        </property>
        <property>
                <name>yarn.resourcemanager.address</name>
                <value>hostname:8032</value>
        </property>
        <property>
                <name>yarn.resourcemanager.scheduler.address</name>
                <value>hostname:8030</value>
        </property>
        <property>
                <name>yarn.resourcemanager.resource-tracker.address</name>
                <value>hostname:8031</value>
        </property>
        <property>
                <name>yarn.resourcemanager.admin.address</name>
                <value>hostname:8033</value>
        </property>
        <property>
                <name>yarn.resourcemanager.webapp.address</name>
                <value>hostname:8088</value>
        </property>
</configuration>

  5. hadoop-env.sh

     export JAVA_HOME=/usr/java/jdk.1.7.0_67

  6 .yarn-env.sh

     export JAVA_HOME=/usr/java/jdk1.7.0_67

  7.修改完配置文件, 将修改后的所有文件都复制(覆盖)到各个节点 hadoop 目录下(必须保证.ssh权限为700)

  ssh-keygen -t rsa

  cat id_rsa.pub >> authorized_keys

或者 ssh-copy-id -i  username@serverIP

scp hadoop-env.sh hdfs-site.xml mapred-site.xml slaves yarn-env.sh yarn-site.xml 
   hadoop@hostname:/home/hadoop/hadoop-1.5.1/etc/hadoop/

(4)格式化namenode

     hdfs namenode -format

(5)启动hdfs

     hadoop/sbin/start-hdfs.sh

(6)hbase:

1.hbase-site.xml:

<configuration>

<property>

<name>hbase.rootdir</name>

<value>hdfs://master:9000/hbase</value>

</property>

<property>

<name>hbase.cluster.distributed</name>

<value>true</value>

</property>

<property>

<name>hbase.master</name>

<value>hdfs://master:6000</value>

</property>

<property>

<name>hbase.zookeeper.quorum</name>

<value>master,slave1</value>

</property>

<property>

<name>hbase.zookeeper.property.dataDir</name>

<value>/home/hadoop/opt/zookeeper-3.4.6</value>

</property>

</configuration>

     2.修改regionservers

master

slave1

     3.hbase-env.sh

         export HBASE_MANAGES_ZK=true//hbase自己维护zookeeper

     4.启动hbase

          ./zkServer.sh start

           ./start-hbase.sh

(7)spark编译:

         mvn -Dhadoop.version=2.6.0 -Phadoop-2.6 -DskipTests clean package

         mvn -Dhadoop.version=2.6.0 -Phadoop-2.6 -Dmaven.test.skip=true clean package -Dmaven.compile.fork=true

(8)spark/conf配置:

     spark-env.sh:

export JAVA_HOME=/usr/java/jdk1.8.0_101
export HADOOP_HOME=/home/templarzq/hadoop
export HADOOP_CONF_DIR=/home/templarzq/hadoop/etc/hadoop
export <span style="font-family: Arial, Helvetica, sans-serif;">SPARK_HOME=/home/templarzq/spark</span>

export <span style="font-family: Arial, Helvetica, sans-serif;">SPARK_CLASSPATH="$ES_HOME/lib/*":"$KAFKA_HOME/libs/*":"$SPARK_HOME/jars/*":$SPARK_CLASSPATH</span>

export SPARK_LOCAL_IP=192.168.8.157
export SPARK_EXECUTOR_INSTANCES=1
export SPARK_EXECUTOR_CORES=6
export SPARK_EXECUTOR_MEMORY=4G
export SPARK_UI_PORT=4040
export SPARK_MASTER_IP=192.168.8.157
#export SPARK_MASTER_IP=localhost
export SPARK_MASTER_PORT=7077
export SPARK_MASTER_WEBUI_PORT=8099
export SPARK_MASTER_HOST=192.168.8.157
export SPARK_WORKER_CORES=6
export SPARK_WORKER_INSTANCES=2
export SPARK_WORKER_MEMORY=8g
export SPARK_WORKER_WEBUI_PORT=8081
export LD_LIBRARY_PATH=${LD_LIBRARY_PATH}:$HADOOP_HOME/lib/native




    slaves:添加slave ip

   spark-deafults.conf:

 spark.master                     spark://192.168.8.157:7077
 spark.eventLog.enabled           true
 spark.eventLog.dir               /tmp/eventLog
# spark.eventLog.dir hdfs://192.168.8.157:9000/eventLog
# spark.serializer org.apache.spark.serializer.KryoSerializer 
 spark.driver.memory 512m 
 spark.executor.extraJavaOptions -XX:+UseG1GC
# spark.executor.extraJavaOptions -XX:+PrintGCDetails -Dkey=value -Dnumbers="one two three"
 spark.es.nodes 192.168.8.157
 spark.es.port 9200 
 spark.sql.hive.thriftServer.singleSession true


(8)启动spark

           $SPARK_HOME/sbin/start-all.sh

      (或者:

          bin/spark-sql –master spark://10.8.2.100:7077 –jars /usr/local/spark-1.5.2/lib/elasticsearch-hadoop-2.1.1.jar

         spark-shell --master spark://10.8.2.100:7077

       )

(9)启动thriftserver:

          编辑$SPARK_HOME/conf/hive-default.xml

<configuration>

<property>

  <name>javax.jdo.option.ConnectionURL</name>

  <value>jdbc:derby:;databaseName=metastore_db;create=true</value>

  <description>JDBC connect string for a JDBC metastore</description>

</property>

<property>

  <name>javax.jdo.option.ConnectionDriverName</name>

  <value>org.apache.derby.jdbc.EmbeddedDriver</value>

  <description>Driver class name for a JDBC metastore</description>

</property>

<property>

  <name>hive.metastore.warehouse.dir</name>

  <value>/home/hadoop/hive/hive-warehouse</value>

  <description>Where to store metastore data</description>

</property>

</configuration>

          重启spark:  $SPARK_HOME/sbin/start-all.sh

          启动thriftserver:$SPARK_HOME/sbin/start-thriftserver.sh --master spark://localhost:7077



评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值