1.spark报错信息:
org.apache.spark.SparkException: Job aborted due to stage failure: Task 5 in stage 2.0 failed 4 times, most recent failure: Lost task 5.3 in stage 2.0 (TID 227, dhpnm236-192-064, executor 25): ExecutorLostFailure (executor 25 exited caused by one of the running tasks) Reason: Container from a bad node: container_e98_1651843250015_1453361_01_000028 on host: dhpnm236-192-064. Exit status: 143.
Diagnostics: Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143
Killed by external signal
问题:资源不足;
解决方案:增大job可用内存或调整代码;
本文针对Spark运行中出现的任务失败问题进行分析,特别是当任务因为资源不足导致反复失败的情况。介绍了具体的错误信息及其含义,并提出了相应的解决策略,包括调整作业内存配置和优化代码等。
242

被折叠的 条评论
为什么被折叠?



