You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sean Owen (JIRA)" <ji...@apache.org> on 2016/03/11 12:45:52 UTC

[jira] [Updated] (SPARK-13818) the spark streaming job will be always processing status when restart elasticsearch

     [ https://issues.apache.org/jira/browse/SPARK-13818?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Sean Owen updated SPARK-13818:
------------------------------
         Priority: Major  (was: Blocker)
    Fix Version/s:     (was: 1.5.3)
                       (was: 1.4.2)

@yuemeng Please don't open a JIRA until you read https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark  You should not set blocker, and it does not make sense to set fix versions. Further, this is an Elasticsearch issue, not Spark (at this stage at least). I'm going to close it.

> the spark streaming job will be always processing status when restart elasticsearch 
> ------------------------------------------------------------------------------------
>
>                 Key: SPARK-13818
>                 URL: https://issues.apache.org/jira/browse/SPARK-13818
>             Project: Spark
>          Issue Type: Bug
>          Components: Streaming
>    Affects Versions: 1.3.0, 1.4.0, 1.5.0
>            Reporter: yuemeng
>
> Using spark streaming to write data into elasticsearch-hadoop system ,when we restart  elasticsearch system,tasks in some job at this time will be get follow error:
> Job aborted due to stage failure: Task 0 in stage 4.0 failed 4 times, most recent failure: Lost task 0.3 in stage 4.0 (TID 75, CIS-store02): org.elasticsearch.hadoop.EsHadoopIllegalStateException: Cluster state volatile; cannot find node backing shards - please check whether your cluster is stable
> at org.elasticsearch.hadoop.rest.RestRepository.getWriteTargetPrimaryShards(RestRepository.java:370)
> at org.elasticsearch.hadoop.rest.RestService.initSingleIndex(RestService.java:425)
> at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:393)
> at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:40)
> at org.elasticsearch.spark.rdd.EsSpark$$anonfun$saveToEs$1.apply(EsSpark.scala:67)
> at org.elasticsearch.spark.rdd.EsSpark$$anonfun$saveToEs$1.apply(EsSpark.scala:67)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:63)
> at org.apache.spark.scheduler.Task.run(Task.scala:70)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> Driver stacktrace:
> and this batch will be always in the status of processing,Never failed or finished,it maybe cause resources for this batch never release.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org