val sparkConf = new SparkConf().setAppName("sparkRealTimeServer")
.setMaster("yarn").setSparkHome("/app/data1/spark").set("spark.submit.deployMode", "client").
set("spark.testing.memory", "2147480000").
set("spark.sql.hive.verifyPartitionPath", "true").
set("spark.yarn.executor.memoryOverhead", "2048m").
set("spark.dynamicAllocation.enabled", "true").
set("spark.shuffle.service.enabled", "true").
set("spark.dynamicAllocation.executorIdleTimeout", "60").
set("spark.dynamicAllocation.cachedExecutorIdleTimeout", "18000").
set("spark.dynamicAllocation.initialExecutors", "15").
set("spark.dynamicAllocation.maxExecutors", "15").
set("spark.dynamicAllocation.minExecutors", "15").
set("spark.dynamicAllocation.schedulerBacklogTimeout", "10").
set("spark.eventLog.enabled", "true").
set("spark.serializer", "org.apache.spark.serializer.KryoSerializer").
set("spark.hadoop.dfs.nameservices","iptvcluster").
set("spark.hadoop.dfs.client.failover.proxy.provider.iptvcluster","org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider").
set("spark.hadoop.dfs.ha.namenodes.iptvcluster", "nn1,nn2").
set("spark.hadoop.dfs.namenode.rpc-address.iptvcluster.nn1", "namenode1:8020").
set("spark.hadoop.dfs.namenode.rpc-address.iptvcluster.nn2", "namenode2:8020").
set("spark.hadoop.yarn.resourcemanager.hostname","namenode1").
set("spark.hadoop.yarn.resourcemanager.address","namenode1:8032").
set("spark.yarn.access.namenodes", "hdfs://namenode1:8020,hdfs://namenode2:8020").
set("spark.yarn.stagingDir", "hdfs://iptvcluster/user/hadoop/").
set("spark.driver.allowMultipleContexts", "true").
set("spark.scheduler.listenerbus.eventqueue.size", "100000").
set("spark.rdd.compress","true")
Spark项目连接Hadoop HA配置文件编写方案
最新推荐文章于 2025-01-06 16:10:11 发布