java.lang.NoSuchMethodError: org.apache.hadoop.mapred.TaskID.<init>

本文详细解析了在使用Spark进行大数据处理时遇到的Hadoop Jar包冲突问题,特别是java.lang.NoSuchMethodError异常的根源。文章深入探讨了不同版本的Hadoop和Spark之间的兼容性问题,并提供了解决方案,帮助读者避免此类错误,确保大数据任务的顺利执行。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

java.lang.NoSuchMethodError: org.apache.hadoop.mapred.TaskID.(Lorg/apache/hadoop/mapreduce/JobID;Lorg/apache/hadoop/mapreduce/TaskType;I)V
at org.apache.spark.rdd.HadoopRDD.addLocalConfiguration(HadoopRDD.scala:384)atorg.apache.spark.rdd.HadoopRDD.addLocalConfiguration(HadoopRDD.scala:384) at org.apache.spark.rdd.HadoopRDD.addLocalConfiguration(HadoopRDD.scala:384)atorg.apache.spark.rdd.HadoopRDD$anon1.&lt;init&gt;(HadoopRDD.scala:246)atorg.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:211)atorg.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:102)atorg.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)atorg.apache.spark.rdd.RDD.iterator(RDD.scala:287)atorg.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)atorg.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)atorg.apache.spark.rdd.RDD.iterator(RDD.scala:287)atorg.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)atorg.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)atorg.apache.spark.rdd.RDD1.&lt;init&gt;(HadoopRDD.scala:246) at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:211) at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:102) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323) at org.apache.spark.rdd.RDD.iterator(RDD.scala:287) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323) at org.apache.spark.rdd.RDD.iterator(RDD.scala:287) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323) at org.apache.spark.rdd.RDD1.<init>(HadoopRDD.scala:246)atorg.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:211)atorg.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:102)atorg.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)atorg.apache.spark.rdd.RDD.iterator(RDD.scala:287)atorg.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)atorg.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)atorg.apache.spark.rdd.RDD.iterator(RDD.scala:287)atorg.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)atorg.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)atorg.apache.spark.rdd.RDD$anonfun8.apply(RDD.scala:336)atorg.apache.spark.rdd.RDD8.apply(RDD.scala:336) at org.apache.spark.rdd.RDD8.apply(RDD.scala:336)atorg.apache.spark.rdd.RDD$anonfun8.apply(RDD.scala:334)atorg.apache.spark.storage.BlockManager8.apply(RDD.scala:334) at org.apache.spark.storage.BlockManager8.apply(RDD.scala:334)atorg.apache.spark.storage.BlockManageranonfunanonfunanonfundoPutIterator1.apply(BlockManager.scala:957)atorg.apache.spark.storage.BlockManager1.apply(BlockManager.scala:957) at org.apache.spark.storage.BlockManager1.apply(BlockManager.scala:957)atorg.apache.spark.storage.BlockManageranonfunanonfunanonfundoPutIterator1.apply(BlockManager.scala:948)atorg.apache.spark.storage.BlockManager.doPut(BlockManager.scala:888)atorg.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:948)atorg.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:694)atorg.apache.spark.rdd.RDD.getOrCompute(RDD.scala:334)atorg.apache.spark.rdd.RDD.iterator(RDD.scala:285)atorg.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)atorg.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)atorg.apache.spark.rdd.RDD1.apply(BlockManager.scala:948) at org.apache.spark.storage.BlockManager.doPut(BlockManager.scala:888) at org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:948) at org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:694) at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:334) at org.apache.spark.rdd.RDD.iterator(RDD.scala:285) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323) at org.apache.spark.rdd.RDD1.apply(BlockManager.scala:948)atorg.apache.spark.storage.BlockManager.doPut(BlockManager.scala:888)atorg.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:948)atorg.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:694)atorg.apache.spark.rdd.RDD.getOrCompute(RDD.scala:334)atorg.apache.spark.rdd.RDD.iterator(RDD.scala:285)atorg.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)atorg.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)atorg.apache.spark.rdd.RDD$anonfun8.apply(RDD.scala:336)atorg.apache.spark.rdd.RDD8.apply(RDD.scala:336) at org.apache.spark.rdd.RDD8.apply(RDD.scala:336)atorg.apache.spark.rdd.RDD$anonfun8.apply(RDD.scala:334)atorg.apache.spark.storage.BlockManager8.apply(RDD.scala:334) at org.apache.spark.storage.BlockManager8.apply(RDD.scala:334)atorg.apache.spark.storage.BlockManageranonfunanonfunanonfundoPutIterator1.apply(BlockManager.scala:957)atorg.apache.spark.storage.BlockManager1.apply(BlockManager.scala:957) at org.apache.spark.storage.BlockManager1.apply(BlockManager.scala:957)atorg.apache.spark.storage.BlockManageranonfunanonfunanonfundoPutIterator1.apply(BlockManager.scala:948)atorg.apache.spark.storage.BlockManager.doPut(BlockManager.scala:888)atorg.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:948)atorg.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:694)atorg.apache.spark.rdd.RDD.getOrCompute(RDD.scala:334)atorg.apache.spark.rdd.RDD.iterator(RDD.scala:285)atorg.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)atorg.apache.spark.scheduler.Task.run(Task.scala:99)atorg.apache.spark.executor.Executor1.apply(BlockManager.scala:948) at org.apache.spark.storage.BlockManager.doPut(BlockManager.scala:888) at org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:948) at org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:694) at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:334) at org.apache.spark.rdd.RDD.iterator(RDD.scala:285) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87) at org.apache.spark.scheduler.Task.run(Task.scala:99) at org.apache.spark.executor.Executor1.apply(BlockManager.scala:948)atorg.apache.spark.storage.BlockManager.doPut(BlockManager.scala:888)atorg.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:948)atorg.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:694)atorg.apache.spark.rdd.RDD.getOrCompute(RDD.scala:334)atorg.apache.spark.rdd.RDD.iterator(RDD.scala:285)atorg.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)atorg.apache.spark.scheduler.Task.run(Task.scala:99)atorg.apache.spark.executor.ExecutorTaskRunner.run(Executor.scala:282)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutorWorker.run(ThreadPoolExecutor.java:624)atjava.lang.Thread.run(Thread.java:748)19/01/2410:09:24WARNscheduler.TaskSetManager:Losttask0.0instage0.0(TID0,localhost,executordriver):java.lang.NoSuchMethodError:org.apache.hadoop.mapred.TaskID.&lt;init&gt;(Lorg/apache/hadoop/mapreduce/JobID;Lorg/apache/hadoop/mapreduce/TaskType;I)Vatorg.apache.spark.rdd.HadoopRDDWorker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 19/01/24 10:09:24 WARN scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, localhost, executor driver): java.lang.NoSuchMethodError: org.apache.hadoop.mapred.TaskID.&lt;init&gt;(Lorg/apache/hadoop/mapreduce/JobID;Lorg/apache/hadoop/mapreduce/TaskType;I)V at org.apache.spark.rdd.HadoopRDDWorker.run(ThreadPoolExecutor.java:624)atjava.lang.Thread.run(Thread.java:748)19/01/2410:09:24WARNscheduler.TaskSetManager:Losttask0.0instage0.0(TID0,localhost,executordriver):java.lang.NoSuchMethodError:org.apache.hadoop.mapred.TaskID.<init>(Lorg/apache/hadoop/mapreduce/JobID;Lorg/apache/hadoop/mapreduce/TaskType;I)Vatorg.apache.spark.rdd.HadoopRDD.addLocalConfiguration(HadoopRDD.scala:384)
at org.apache.spark.rdd.HadoopRDDKaTeX parse error: Can't use function '$' in math mode at position 5: anon$̲1.<init>(Hadoop…anonfun$8.apply(RDD.scala:336)
问题解决:
Hadoop的jar包冲突

评论 2
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值