2011-06-17

Words

charity

regime

sanction

asset freezes

dynasty

terrorist

envoy

charisma

predecessor

deputy

alleged

influence

excessive

odds

cover up

extent

lead poisoning

insert overwrite table case_data_sample select * from case_data_sample_tmp; 2025-06-18 16:37:06,500 INFO [main] conf.HiveConf: Using the default value passed in for log id: 531f6207-2ea7-471a-9eac-9ce1e6a79910 2025-06-18 16:37:06,500 INFO [main] session.SessionState: Updating thread name to 531f6207-2ea7-471a-9eac-9ce1e6a79910 main 2025-06-18 16:37:06,503 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Compiling command(queryId=root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990): insert overwrite table case_data_sample select * from case_data_sample_tmp 2025-06-18 16:37:06,546 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Concurrency mode is disabled, not creating a lock manager 2025-06-18 16:37:06,547 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Starting Semantic Analysis 2025-06-18 16:37:06,559 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Completed phase 1 of Semantic Analysis 2025-06-18 16:37:06,560 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for source tables 2025-06-18 16:37:06,588 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for subqueries 2025-06-18 16:37:06,588 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for destination tables 2025-06-18 16:37:06,627 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Completed getting MetaData in Semantic Analysis 2025-06-18 16:37:08,746 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for source tables 2025-06-18 16:37:08,784 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for subqueries 2025-06-18 16:37:08,784 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for destination tables 2025-06-18 16:37:08,884 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] common.FileUtils: Creating directory if it doesn&#39;t exist: hdfs://master:8020/user/hive/warehouse/ad_traffic.db/case_data_sample/.hive-staging_hive_2025-06-18_16-37-06_538_2993870488593298816-1 2025-06-18 16:37:09,012 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Generate an operator pipeline to autogather column stats for table ad_traffic.case_data_sample in query insert overwrite table case_data_sample select * from case_data_sample_tmp 2025-06-18 16:37:09,069 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for source tables 2025-06-18 16:37:09,098 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for subqueries 2025-06-18 16:37:09,098 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for destination tables 2025-06-18 16:37:09,155 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Context: New scratch dir is hdfs://master:8020/user/hive/tmp/root/531f6207-2ea7-471a-9eac-9ce1e6a79910/hive_2025-06-18_16-37-09_012_5684077500801740374-1 2025-06-18 16:37:09,221 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] common.FileUtils: Creating directory if it doesn&#39;t exist: hdfs://master:8020/user/hive/tmp/root/531f6207-2ea7-471a-9eac-9ce1e6a79910/hive_2025-06-18_16-37-09_012_5684077500801740374-1/-mr-10000/.hive-staging_hive_2025-06-18_16-37-09_012_5684077500801740374-1 2025-06-18 16:37:09,234 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: CBO Succeeded; optimized logical plan. 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for FS(2) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for FS(9) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for SEL(8) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for GBY(7) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for RS(6) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for GBY(5) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for SEL(4) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for SEL(1) 2025-06-18 16:37:09,330 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for TS(0) 2025-06-18 16:37:09,385 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] optimizer.ColumnPrunerProcFactory: RS 6 oldColExprMap: {VALUE._col20=Column[_col20], VALUE._col10=Column[_col10], VALUE._col21=Column[_col21], VALUE._col11=Column[_col11], VALUE._col12=Column[_col12], VALUE._col2=Column[_col2], VALUE._col3=Column[_col3], VALUE._col4=Column[_col4], VALUE._col5=Column[_col5], VALUE._col0=Column[_col0], VALUE._col1=Column[_col1], VALUE._col13=Column[_col13], VALUE._col14=Column[_col14], VALUE._col15=Column[_col15], VALUE._col16=Column[_col16], VALUE._col6=Column[_col6], VALUE._col17=Column[_col17], VALUE._col7=Column[_col7], VALUE._col18=Column[_col18], VALUE._col8=Column[_col8], VALUE._col19=Column[_col19], VALUE._col9=Column[_col9]} 2025-06-18 16:37:09,386 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] optimizer.ColumnPrunerProcFactory: RS 6 newColExprMap: {VALUE._col20=Column[_col20], VALUE._col10=Column[_col10], VALUE._col21=Column[_col21], VALUE._col11=Column[_col11], VALUE._col12=Column[_col12], VALUE._col2=Column[_col2], VALUE._col3=Column[_col3], VALUE._col4=Column[_col4], VALUE._col5=Column[_col5], VALUE._col0=Column[_col0], VALUE._col1=Column[_col1], VALUE._col13=Column[_col13], VALUE._col14=Column[_col14], VALUE._col15=Column[_col15], VALUE._col16=Column[_col16], VALUE._col6=Column[_col6], VALUE._col17=Column[_col17], VALUE._col7=Column[_col7], VALUE._col18=Column[_col18], VALUE._col8=Column[_col8], VALUE._col19=Column[_col19], VALUE._col9=Column[_col9]} 2025-06-18 16:37:09,500 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SetSparkReducerParallelism: Number of reducers for sink RS[6] was already determined to be: 1 2025-06-18 16:37:09,646 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Examining input format to see if vectorization is enabled. 2025-06-18 16:37:09,655 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Vectorization is enabled for input format(s) [org.apache.hadoop.mapred.TextInputFormat] 2025-06-18 16:37:09,655 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Validating and vectorizing MapWork... (vectorizedVertexNum 0) 2025-06-18 16:37:09,706 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map vectorization enabled: true 2025-06-18 16:37:09,706 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map vectorized: false 2025-06-18 16:37:09,707 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map notVectorizedReason: Aggregation Function expression for GROUPBY operator: UDF compute_stats not supported 2025-06-18 16:37:09,707 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map vectorizedVertexNum: 0 2025-06-18 16:37:09,707 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map enabledConditionsMet: [hive.vectorized.use.vector.serde.deserialize IS true] 2025-06-18 16:37:09,707 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map inputFileFormatClassNameSet: [org.apache.hadoop.mapred.TextInputFormat] 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Validating and vectorizing ReduceWork... (vectorizedVertexNum 1) 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reduce vectorization enabled: true 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reduce vectorized: false 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reduce notVectorizedReason: Aggregation Function expression for GROUPBY operator: UDF compute_stats not supported 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reduce vectorizedVertexNum: 1 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reducer hive.vectorized.execution.reduce.enabled: true 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reducer engine: spark 2025-06-18 16:37:09,784 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Completed plan generation 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Semantic Analysis Completed (retrial = false) 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: , FieldSchema(name:case_data_sample_tmp.timestamps, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.camp, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.creativeid, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.mobile_os, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.mobile_type, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.app_key_md5, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.app_name_md5, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.placementid, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.useragent, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.mediaid, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.os_type, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.born_time, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.label, type:int, comment:null)], properties:null) 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Completed compiling command(queryId=root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990); Time taken: 3.282 seconds 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] reexec.ReExecDriver: Execution #1 of query 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Concurrency mode is disabled, not creating a lock manager 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Executing command(queryId=root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990): insert overwrite table case_data_sample select * from case_data_sample_tmp 2025-06-18 16:37:09,786 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Query ID = root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990 2025-06-18 16:37:09,786 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Total jobs = 1 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Launching Job 1 out of 1 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Starting task [Stage-1:MAPRED] in serial mode 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: In order to change the average load for a reducer (in bytes): 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: set hive.exec.reducers.bytes.per.reducer=<number> 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: In order to limit the maximum number of reducers: 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: set hive.exec.reducers.max=<number> 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: In order to set a constant number of reducers: 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: set mapreduce.job.reduces=<number> 2025-06-18 16:37:09,834 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] session.SparkSessionManagerImpl: Setting up the session manager. 2025-06-18 16:37:10,327 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] session.SparkSession: Trying to open Spark session 4201eecc-977e-4e69-89ec-50403379b3d2 Failed to execute spark task, with exception &#39;org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2)&#39; 2025-06-18 16:37:10,372 ERROR [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: .lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.util.RunJar.run(RunJar.java:318) at org.apache.hadoop.util.RunJar.main(RunJar.java:232) Caused by: java.lang.NoClassDefFoundError: org/apache/spark/SparkConf at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.generateSparkConf(HiveSparkClientFactory.java:263) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:98) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:76) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:87) ... 24 more Caused by: java.lang.ClassNotFoundException: org.apache.spark.SparkConf at java.net.URLClassLoader.findClass(URLClassLoader.java:387) at java.lang.ClassLoader.loadClass(ClassLoader.java:418) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:359) at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ... 28 more 2025-06-18 16:37:10,378 ERROR [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: Failed to execute spark task, with exception &#39;org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2)&#39; org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2 at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.getHiveException(SparkSessionImpl.java:221) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:92) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:115) at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:136) at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:115) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.util.RunJar.run(RunJar.java:318) at org.apache.hadoop.util.RunJar.main(RunJar.java:232) Caused by: java.lang.NoClassDefFoundError: org/apache/spark/SparkConf at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.generateSparkConf(HiveSparkClientFactory.java:263) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:98) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:76) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:87) ... 24 more Caused by: java.lang.ClassNotFoundException: org.apache.spark.SparkConf at java.net.URLClassLoader.findClass(URLClassLoader.java:387) at java.lang.ClassLoader.loadClass(ClassLoader.java:418) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:359) at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ... 28 more 2025-06-18 16:37:10,391 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] reexec.ReOptimizePlugin: ReOptimization: retryPossible: false FAILED: Execution Error, return code 30041 from org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2 2025-06-18 16:37:10,391 ERROR [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: FAILED: Execution Error, return code 30041 from org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2 2025-06-18 16:37:10,392 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Completed executing command(queryId=root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990); Time taken: 0.607 seconds 2025-06-18 16:37:10,392 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Concurrency mode is disabled, not creating a lock manager 2025-06-18 16:37:10,433 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] conf.HiveConf: Using the default value passed in for log id: 531f6207-2ea7-471a-9eac-9ce1e6a79910 2025-06-18 16:37:10,433 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] session.SessionState: Resetting thread name to main
06-19
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值