首页
学习
活动
专区
圈层
工具
发布
社区首页 >问答首页 >spark-submit在作业完成后继续挂起

spark-submit在作业完成后继续挂起
EN

Stack Overflow用户
提问于 2016-05-25 03:03:56
回答 2查看 2.2K关注 0票数 10

我正在尝试在AWS中使用hdfs测试spark 1.6。我使用的是examples文件夹中的wordcount python示例。我使用spark-submit提交作业,作业成功完成,并在控制台上打印结果。web-UI还说它已经完成了。但是,spark-submit永远不会终止。我已经验证了上下文在单词计数示例代码中也被停止了。

会出什么问题呢?

这就是我在控制台上看到的。

代码语言:javascript
复制
6-05-24 14:58:04,749 INFO  [Thread-3] handler.ContextHandler (ContextHandler.java:doStop(843)) - stopped o.s.j.s.ServletContextHandler{/stages/stage,null}
2016-05-24 14:58:04,749 INFO  [Thread-3] handler.ContextHandler (ContextHandler.java:doStop(843)) - stopped o.s.j.s.ServletContextHandler{/stages/json,null}
2016-05-24 14:58:04,749 INFO  [Thread-3] handler.ContextHandler (ContextHandler.java:doStop(843)) - stopped o.s.j.s.ServletContextHandler{/stages,null}
2016-05-24 14:58:04,749 INFO  [Thread-3] handler.ContextHandler (ContextHandler.java:doStop(843)) - stopped o.s.j.s.ServletContextHandler{/jobs/job/json,null}
2016-05-24 14:58:04,750 INFO  [Thread-3] handler.ContextHandler (ContextHandler.java:doStop(843)) - stopped o.s.j.s.ServletContextHandler{/jobs/job,null}
2016-05-24 14:58:04,750 INFO  [Thread-3] handler.ContextHandler (ContextHandler.java:doStop(843)) - stopped o.s.j.s.ServletContextHandler{/jobs/json,null}
2016-05-24 14:58:04,750 INFO  [Thread-3] handler.ContextHandler (ContextHandler.java:doStop(843)) - stopped o.s.j.s.ServletContextHandler{/jobs,null}
2016-05-24 14:58:04,802 INFO  [Thread-3] ui.SparkUI (Logging.scala:logInfo(58)) - Stopped Spark web UI at http://172.30.2.239:4040
2016-05-24 14:58:04,805 INFO  [Thread-3] cluster.SparkDeploySchedulerBackend (Logging.scala:logInfo(58)) - Shutting down all executors
2016-05-24 14:58:04,805 INFO  [dispatcher-event-loop-2] cluster.SparkDeploySchedulerBackend (Logging.scala:logInfo(58)) - Asking each executor to shut down
2016-05-24 14:58:04,814 INFO  [dispatcher-event-loop-5] spark.MapOutputTrackerMasterEndpoint (Logging.scala:logInfo(58)) - MapOutputTrackerMasterEndpoint stopped!
2016-05-24 14:58:04,818 INFO  [Thread-3] storage.MemoryStore (Logging.scala:logInfo(58)) - MemoryStore cleared
2016-05-24 14:58:04,818 INFO  [Thread-3] storage.BlockManager (Logging.scala:logInfo(58)) - BlockManager stopped
2016-05-24 14:58:04,820 INFO  [Thread-3] storage.BlockManagerMaster (Logging.scala:logInfo(58)) - BlockManagerMaster stopped
2016-05-24 14:58:04,821 INFO  [dispatcher-event-loop-3] scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint (Logging.scala:logInfo(58)) - OutputCommitCoordinator stopped!
2016-05-24 14:58:04,824 INFO  [Thread-3] spark.SparkContext (Logging.scala:logInfo(58)) - Successfully stopped SparkContext
2016-05-24 14:58:04,827 INFO  [sparkDriverActorSystem-akka.actor.default-dispatcher-2] remote.RemoteActorRefProvider$RemotingTerminator (Slf4jLogger.scala:apply$mcV$sp(74)) - Shutting down remote daemon.
2016-05-24 14:58:04,828 INFO  [sparkDriverActorSystem-akka.actor.default-dispatcher-2] remote.RemoteActorRefProvider$RemotingTerminator (Slf4jLogger.scala:apply$mcV$sp(74)) - Remote daemon shut down; proceeding with flushing remote transports.
2016-05-24 14:58:04,843 INFO  [sparkDriverActorSystem-akka.actor.default-dispatcher-2] remote.RemoteActorRefProvider$RemotingTerminator (Slf4jLogger.scala:apply$mcV$sp(74)) - Remoting shut down.

我必须执行ctrl-c来终止spark-submit进程。这真的是一个奇怪的问题,我不知道如何解决这个问题。请让我知道是否有任何我应该查看的日志,或者在这里做一些不同的事情。

下面是spark-submit进程的jstack输出的粘贴库链接:http://pastebin.com/Nfnt4XmT

EN

回答 2

Stack Overflow用户

发布于 2019-01-17 22:13:21

我的spark作业代码中的自定义线程池也有同样的问题。我发现spark-submit在代码中使用定制的守护线程池时会挂起。您可以查看ThreadUtils.newDaemonCachedThreadPool()以了解spark开发人员如何创建线程池,或者您可以使用此实用程序,但要小心,因为它们是包私有的。

票数 1
EN

Stack Overflow用户

发布于 2021-01-10 10:15:11

对我来说:在本地,它会停止,但在集群上执行时会继续发送日志。我认为这是意料之中的行为。这是从另一个答案中得到的,对我来说很有效。

代码语言:javascript
复制
val sc=new SparkContext(conf)
try{
      //code goes here
}
finally{
      sc.stop()
}

这项技术在Spark和PySpark中都有效。sample shot

票数 -1
EN
页面原文内容由Stack Overflow提供。腾讯云小微IT领域专用引擎提供翻译支持
原文链接:

https://stackoverflow.com/questions/37421852

复制
相关文章

相似问题

领券
问题归档专栏文章快讯文章归档关键词归档开发者手册归档开发者手册 Section 归档