You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@mahout.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2016/06/11 06:37:55 UTC

Build failed in Jenkins: Mahout-Quality #3368

See <https://builds.apache.org/job/Mahout-Quality/3368/changes>

Changes:

[smarthi] Rolling back 0.12.2 Release candidate 2

------------------------------------------
[...truncated 136038 lines...]
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
281226 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (3/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (3/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281233 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (8/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (8/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281246 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (6/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (6/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281246 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (7/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (7/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281248 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (11/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (11/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281248 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (1/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (1/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281250 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (14/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (14/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281248 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (4/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (4/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281248 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (5/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (5/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281251 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (13/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (13/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281250 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (2/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (2/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281249 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (15/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (15/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281249 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (10/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (10/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281249 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (9/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (9/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
281248 [CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (12/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:93)) (12/16)
java.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:91)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:342)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:137)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [10000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at scala.concurrent.Await.result(package.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:71)
	... 4 more
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 86.9% relative to #3363
Recording test results
Publishing Javadoc

Jenkins build is back to normal : Mahout-Quality #3380

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3380/changes>


Build failed in Jenkins: Mahout-Quality #3379

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3379/changes>

Changes:

[apalumbo] MAHOUT-1837: Sparse/Dense Matrix analysis for Matrix Multiplication.

------------------------------------------
[...truncated 23602 lines...]
37273 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37273 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-57]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37273 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37274 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-58]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37274 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37275 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-59]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37275 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37275 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-60]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37276 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37276 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-61]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37276 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
37277 [flink-akka.actor.default-dispatcher-32] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-62]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.06 MB of artifacts by 86.8% relative to #3363
Recording test results
Publishing Javadoc
Updating MAHOUT-1837

Build failed in Jenkins: Mahout-Quality #3378

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3378/changes>

Changes:

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

------------------------------------------
[...truncated 62034 lines...]
	... 4 more
428080 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428080 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-73]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428081 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428081 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-74]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428081 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428082 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-75]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428082 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428082 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-76]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428083 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428083 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-77]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428083 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
428084 [flink-akka.actor.default-dispatcher-22] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-78]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 86.9% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3377

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3377/changes>

Changes:

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

------------------------------------------
[...truncated 55968 lines...]
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(13/16) switched to RUNNING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(12/16) switched to RUNNING 
06/12/2016 23:10:37	Job execution switched to status FAILING.
java.lang.Exception: Failed to deploy the task to slot SimpleSlot (0)(1) - df67b19955b334462482d095752b6ddb @ localhost - 16 slots - URL: akka://flink/user/taskmanager_1 - ALLOCATED/ALIVE: Response was not of type Acknowledge
	at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:395)
	at akka.dispatch.OnComplete.internal(Future.scala:247)
	at akka.dispatch.OnComplete.internal(Future.scala:244)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171)
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/12/2016 23:10:37	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(5/16) switched to CANCELING 
06/12/2016 23:10:37	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(6/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(1/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(2/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(3/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(4/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(5/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(6/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(7/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(8/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(9/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(10/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(11/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(11/16) switched to RUNNING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(12/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(13/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(14/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(15/16) switched to CANCELING 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(10/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(7/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(9/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(12/16) switched to CANCELED 
06/12/2016 23:10:37	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(5/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(13/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(8/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(1/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(2/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(4/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(3/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(15/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(5/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(11/16) switched to CANCELED 
06/12/2016 23:10:37	DataSink (org.apache.flink.api.java.Utils$CollectHelper@4846fae7)(14/16) switched to CANCELED 
06/12/2016 23:10:37	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(6/16) switched to CANCELED 
06/12/2016 23:10:37	Job execution switched to status FAILED.
- Model DFS Serialization *** FAILED ***
  org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
  at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:717)
  at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663)
  at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663)
  at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
  at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
  at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41)
  at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401)
  at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
  at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
  at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
  ...
  Cause: java.lang.Exception: Failed to deploy the task to slot SimpleSlot (0)(1) - df67b19955b334462482d095752b6ddb @ localhost - 16 slots - URL: akka://flink/user/taskmanager_1 - ALLOCATED/ALIVE: Response was not of type Acknowledge
  at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:395)
  at akka.dispatch.OnComplete.internal(Future.scala:247)
  at akka.dispatch.OnComplete.internal(Future.scala:244)
  at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174)
  at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171)
  at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
  at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
  at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
  at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
  at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
  ...
06/12/2016 23:10:38	Job execution switched to status RUNNING.
06/12/2016 23:10:38	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to SCHEDULED 
06/12/2016 23:10:38	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to DEPLOYING 
06/12/2016 23:10:38	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to RUNNING 
06/12/2016 23:10:38	RangePartition: LocalSample(1/1) switched to SCHEDULED 
06/12/2016 23:10:38	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to FINISHED 
06/12/2016 23:10:38	RangePartition: LocalSample(1/1) switched to DEPLOYING 
06/12/2016 23:10:38	RangePartition: PreparePartition(1/1) switched to SCHEDULED 
06/12/2016 23:10:38	RangePartition: PreparePartition(1/1) switched to DEPLOYING 
06/12/2016 23:10:38	RangePartition: LocalSample(1/1) switched to RUNNING 
06/12/2016 23:10:38	RangePartition: PreparePartition(1/1) switched to RUNNING 
06/12/2016 23:10:38	RangePartition: GlobalSample(1/1) switched to SCHEDULED 
06/12/2016 23:10:38	RangePartition: GlobalSample(1/1) switched to DEPLOYING 
06/12/2016 23:10:38	RangePartition: LocalSample(1/1) switched to FINISHED 
06/12/2016 23:10:38	RangePartition: GlobalSample(1/1) switched to RUNNING 
06/12/2016 23:10:38	RangePartition: Histogram(1/1) switched to SCHEDULED 
06/12/2016 23:10:38	RangePartition: Histogram(1/1) switched to DEPLOYING 
06/12/2016 23:10:38	RangePartition: GlobalSample(1/1) switched to FINISHED 
06/12/2016 23:10:38	RangePartition: Histogram(1/1) switched to RUNNING 
06/12/2016 23:10:38	RangePartition: Histogram(1/1) switched to FINISHED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(2/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(3/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(4/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(5/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(6/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(7/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(4/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(8/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(5/16) switched to DEPLOYING 
06/12/2016 23:10:38	RangePartition: PreparePartition(1/1) switched to FINISHED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(9/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(8/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(7/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(12/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(14/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(16/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(15/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(11/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(2/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(14/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(13/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(11/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(10/16) switched to SCHEDULED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(12/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(9/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(6/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(3/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(10/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(13/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(15/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(16/16) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(4/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/16) switched to RUNNING 
06/12/2016 23:10:38	Reduce (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/1) switched to SCHEDULED 
06/12/2016 23:10:38	Reduce (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/1) switched to DEPLOYING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(5/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(14/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(2/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(11/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(12/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(9/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(8/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(6/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(3/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(7/16) switched to RUNNING 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/16) switched to FINISHED 
06/12/2016 23:10:38	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(6/16) switched to FINISHED 
261715 [CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194)) (12/16)] ERROR org.apache.flink.runtime.taskmanager.Task  - FATAL - exception in task resource cleanup
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at scala.concurrent.forkjoin.ForkJoinPool.tryAddWorker(ForkJoinPool.java:1672)
	at scala.concurrent.forkjoin.ForkJoinPool.signalWork(ForkJoinPool.java:1966)
	at scala.concurrent.forkjoin.ForkJoinPool.fullExternalPush(ForkJoinPool.java:1905)
	at scala.concurrent.forkjoin.ForkJoinPool.externalPush(ForkJoinPool.java:1834)
	at scala.concurrent.forkjoin.ForkJoinPool.execute(ForkJoinPool.java:2955)
	at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinPool.execute(AbstractDispatcher.scala:387)
	at akka.dispatch.ExecutorServiceDelegate$class.execute(ThreadPoolBuilder.scala:212)
	at akka.dispatch.Dispatcher$LazyExecutorServiceDelegate.execute(Dispatcher.scala:43)
	at akka.dispatch.Dispatcher.registerForExecution(Dispatcher.scala:118)
	at akka.dispatch.Dispatcher.dispatch(Dispatcher.scala:59)
	at akka.actor.dungeon.Dispatch$class.sendMessage(Dispatch.scala:123)
	at akka.actor.ActorCell.sendMessage(ActorCell.scala:369)
	at akka.actor.Cell$class.sendMessage(ActorCell.scala:290)
	at akka.actor.ActorCell.sendMessage(ActorCell.scala:369)
	at akka.actor.RepointableActorRef.$bang(RepointableActorRef.scala:166)
	at akka.actor.ActorRef.tell(ActorRef.scala:123)
	at org.apache.flink.runtime.instance.AkkaActorGateway.tell(AkkaActorGateway.java:79)
	at org.apache.flink.runtime.taskmanager.Task.notifyFinalState(Task.java:735)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:670)
	at java.lang.Thread.run(Thread.java:745)
261715 [flink-akka.actor.default-dispatcher-19] ERROR org.apache.flink.runtime.executiongraph.ExecutionGraph  - Error while notifying execution graph of execution state transition.
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at scala.concurrent.forkjoin.ForkJoinPool.tryAddWorker(ForkJoinPool.java:1672)
	at scala.concurrent.forkjoin.ForkJoinPool.signalWork(ForkJoinPool.java:1966)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.push(ForkJoinPool.java:1072)
	at scala.concurrent.forkjoin.ForkJoinTask.fork(ForkJoinTask.java:654)
	at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinPool.execute(AbstractDispatcher.scala:386)
	at akka.dispatch.ExecutorServiceDelegate$class.execute(ThreadPoolBuilder.scala:212)
	at akka.dispatch.Dispatcher$LazyExecutorServiceDelegate.execute(Dispatcher.scala:43)
	at akka.dispatch.Dispatcher.registerForExecution(Dispatcher.scala:118)
	at akka.dispatch.Dispatcher.dispatch(Dispatcher.scala:59)
	at akka.actor.dungeon.Dispatch$class.sendMessage(Dispatch.scala:123)
	at akka.actor.ActorCell.sendMessage(ActorCell.scala:369)
	at akka.actor.Cell$class.sendMessage(ActorCell.scala:290)
	at akka.actor.ActorCell.sendMessage(ActorCell.scala:369)
	at akka.actor.RepointableActorRef.$bang(RepointableActorRef.scala:166)
	at akka.actor.ActorRef.tell(ActorRef.scala:123)
	at org.apache.flink.runtime.instance.AkkaActorGateway.tell(AkkaActorGateway.java:79)
	at org.apache.flink.runtime.executiongraph.ExecutionGraph.notifyExecutionChange(ExecutionGraph.java:1217)
	at org.apache.flink.runtime.executiongraph.ExecutionVertex.notifyStateTransition(ExecutionVertex.java:627)
	at org.apache.flink.runtime.executiongraph.Execution.transitionState(Execution.java:984)
	at org.apache.flink.runtime.executiongraph.Execution.transitionState(Execution.java:966)
	at org.apache.flink.runtime.executiongraph.Execution.markFinished(Execution.java:658)
	at org.apache.flink.runtime.executiongraph.ExecutionGraph.updateState(ExecutionGraph.java:1091)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$4.apply$mcV$sp(JobManager.scala:518)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$4.apply(JobManager.scala:517)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$4.apply(JobManager.scala:517)
	at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
	at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
	at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41)
	at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
261718 [flink-akka.actor.default-dispatcher-4] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - 
==============================================================
======================      FATAL      =======================
==============================================================

A fatal error occurred, forcing the TaskManager to shut down: FATAL - exception in task resource cleanup
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at scala.concurrent.forkjoin.ForkJoinPool.tryAddWorker(ForkJoinPool.java:1672)
	at scala.concurrent.forkjoin.ForkJoinPool.signalWork(ForkJoinPool.java:1966)
	at scala.concurrent.forkjoin.ForkJoinPool.fullExternalPush(ForkJoinPool.java:1905)
	at scala.concurrent.forkjoin.ForkJoinPool.externalPush(ForkJoinPool.java:1834)
	at scala.concurrent.forkjoin.ForkJoinPool.execute(ForkJoinPool.java:2955)
	at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinPool.execute(AbstractDispatcher.scala:387)
	at akka.dispatch.ExecutorServiceDelegate$class.execute(ThreadPoolBuilder.scala:212)
	at akka.dispatch.Dispatcher$LazyExecutorServiceDelegate.execute(Dispatcher.scala:43)
	at akka.dispatch.Dispatcher.registerForExecution(Dispatcher.scala:118)
	at akka.dispatch.Dispatcher.dispatch(Dispatcher.scala:59)
	at akka.actor.dungeon.Dispatch$class.sendMessage(Dispatch.scala:123)
	at akka.actor.ActorCell.sendMessage(ActorCell.scala:369)
	at akka.actor.Cell$class.sendMessage(ActorCell.scala:290)
	at akka.actor.ActorCell.sendMessage(ActorCell.scala:369)
	at akka.actor.RepointableActorRef.$bang(RepointableActorRef.scala:166)
	at akka.actor.ActorRef.tell(ActorRef.scala:123)
	at org.apache.flink.runtime.instance.AkkaActorGateway.tell(AkkaActorGateway.java:79)
	at org.apache.flink.runtime.taskmanager.Task.notifyFinalState(Task.java:735)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:670)
	at java.lang.Thread.run(Thread.java:745)
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 88.9% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3376

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3376/changes>

Changes:

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

------------------------------------------
[...truncated 58952 lines...]
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/12/2016 19:29:17	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(11/16) switched to CANCELING 
06/12/2016 19:29:17	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(15/16) switched to CANCELING 
06/12/2016 19:29:17	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(16/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(1/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(2/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(3/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(4/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(5/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(6/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(9/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(7/16) switched to FAILED 
java.lang.Exception: Failed to deploy the task to slot SimpleSlot (13)(1) - ac5064bff43db4138bcc5494eda323a4 @ localhost - 16 slots - URL: akka://flink/user/taskmanager_1 - ALLOCATED/ALIVE: Response was not of type Acknowledge
	at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:395)
	at akka.dispatch.OnComplete.internal(Future.scala:247)
	at akka.dispatch.OnComplete.internal(Future.scala:244)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171)
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(10/16) switched to CANCELING 
357474 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(11/16) switched to CANCELED 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(12/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(13/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(12/16) switched to CANCELED 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(14/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(15/16) switched to CANCELING 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(16/16) switched to CANCELED 
357475 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(13/16) switched to CANCELED 
357475 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(6/16) switched to CANCELED 
357476 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(1/16) switched to CANCELED 
357476 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(9/16) switched to CANCELED 
357477 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(15/16) switched to CANCELED 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(3/16) switched to CANCELED 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(10/16) switched to CANCELED 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(14/16) switched to CANCELED 
06/12/2016 19:29:17	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(11/16) switched to CANCELED 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(4/16) switched to CANCELED 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(2/16) switched to CANCELED 
06/12/2016 19:29:17	DataSink (org.apache.flink.api.java.Utils$CollectHelper@5ac9c802)(5/16) switched to CANCELED 
06/12/2016 19:29:17	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(16/16) switched to CANCELED 
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 87.9% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3375

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3375/changes>

Changes:

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

------------------------------------------
[...truncated 58167 lines...]
	... 4 more
353136 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353136 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-51]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353136 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353137 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-52]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353137 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353137 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-53]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353138 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353138 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-54]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353138 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353138 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-55]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353139 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
353139 [flink-akka.actor.default-dispatcher-26] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-56]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 88.9% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3374

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3374/changes>

Changes:

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

------------------------------------------
[...truncated 58214 lines...]
	... 4 more
348219 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348228 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-52]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348229 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348230 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-53]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348230 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348231 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-54]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348232 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348232 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-55]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348232 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348233 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-56]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348233 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - exception on LARS? timer thread
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
348234 [flink-akka.actor.default-dispatcher-2] ERROR akka.actor.ActorSystemImpl  - Uncaught error from thread [flink-scheduler-57]
java.lang.IllegalStateException: problem in scala.concurrent internal callback
	at scala.concurrent.Future$InternalCallbackExecutor$.reportFailure(Future.scala:592)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:471)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:133)
	at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$4.apply(JobManager.scala:132)
	at scala.concurrent.impl.ExecutionContextImpl.reportFailure(ExecutionContextImpl.scala:125)
	at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
	at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	... 4 more
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 87.9% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3373

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3373/changes>

Changes:

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

------------------------------------------
[...truncated 55412 lines...]
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(1/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(4/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(4/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(13/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(9/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(13/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(5/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(9/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(11/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(5/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(11/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(7/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(14/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(2/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(14/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(2/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(6/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(3/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(10/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(3/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(10/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(8/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(8/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(6/16) switched to DEPLOYING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(7/16) switched to DEPLOYING 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(14/16) switched to FINISHED 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(1/16) switched to FINISHED 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(10/16) switched to FINISHED 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(3/16) switched to FINISHED 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(4/16) switched to FINISHED 
Exception in thread "ForkJoinPool-898-worker-17" java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at scala.concurrent.forkjoin.ForkJoinPool.tryAddWorker(ForkJoinPool.java:1672)
	at scala.concurrent.forkjoin.ForkJoinPool.signalWork(ForkJoinPool.java:1966)
	at scala.concurrent.forkjoin.ForkJoinPool.scan(ForkJoinPool.java:2037)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(9/16) switched to FINISHED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(15/16) switched to SCHEDULED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(15/16) switched to DEPLOYING 
Exception in thread "ForkJoinPool-898-worker-7" java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at scala.concurrent.forkjoin.ForkJoinPool.tryAddWorker(ForkJoinPool.java:1672)
	at scala.concurrent.forkjoin.ForkJoinPool.signalWork(ForkJoinPool.java:1966)
	at scala.concurrent.forkjoin.ForkJoinPool.scan(ForkJoinPool.java:2037)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(15/16) switched to FINISHED 
Exception in thread "ForkJoinPool-898-worker-29" java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at scala.concurrent.forkjoin.ForkJoinPool.tryAddWorker(ForkJoinPool.java:1672)
	at scala.concurrent.forkjoin.ForkJoinPool.signalWork(ForkJoinPool.java:1966)
	at scala.concurrent.forkjoin.ForkJoinPool.scan(ForkJoinPool.java:2037)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Exception in thread "ForkJoinPool-898-worker-21" java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at scala.concurrent.forkjoin.ForkJoinPool.tryAddWorker(ForkJoinPool.java:1672)
	at scala.concurrent.forkjoin.ForkJoinPool.signalWork(ForkJoinPool.java:1966)
	at scala.concurrent.forkjoin.ForkJoinPool.scan(ForkJoinPool.java:2037)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(6/16) switched to FINISHED 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(2/16) switched to FINISHED 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(11/16) switched to FINISHED 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(5/16) switched to FINISHED 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(7/16) switched to FINISHED 
259395 [flink-akka.actor.default-dispatcher-8] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(3/16) switched to FAILED 
java.lang.Exception: Failed to deploy the task to slot SimpleSlot (5)(1) - a2c5d8333c3e7bdd7dd75e2211837f1c @ localhost - 16 slots - URL: akka://flink/user/taskmanager_1 - ALLOCATED/ALIVE: Response was not of type Acknowledge
	at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:395)
	at akka.dispatch.OnComplete.internal(Future.scala:247)
	at akka.dispatch.OnComplete.internal(Future.scala:244)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171)
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

06/11/2016 23:03:16	Job execution switched to status FAILING.
java.lang.Exception: Failed to deploy the task to slot SimpleSlot (5)(1) - a2c5d8333c3e7bdd7dd75e2211837f1c @ localhost - 16 slots - URL: akka://flink/user/taskmanager_1 - ALLOCATED/ALIVE: Response was not of type Acknowledge
	at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:395)
	at akka.dispatch.OnComplete.internal(Future.scala:247)
	at akka.dispatch.OnComplete.internal(Future.scala:244)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171)
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(8/16) switched to CANCELING 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(12/16) switched to CANCELING 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(13/16) switched to CANCELING 
06/11/2016 23:03:16	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(16/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(1/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(2/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(4/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(5/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(6/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(7/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(8/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(9/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(10/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(11/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(12/16) switched to CANCELED 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(13/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(14/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(15/16) switched to CANCELING 
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(16/16) switched to CANCELED 
259400 [flink-akka.actor.default-dispatcher-8] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(7/16) switched to CANCELED 
259400 [flink-akka.actor.default-dispatcher-8] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(2/16) switched to CANCELED 
259401 [flink-akka.actor.default-dispatcher-8] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 23:03:16	DataSink (org.apache.flink.api.java.Utils$CollectHelper@66501fcd)(15/16) switched to CANCELED 
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 86.9% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3372

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3372/changes>

Changes:

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

------------------------------------------
[...truncated 55780 lines...]
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 19:20:02	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(4/16) switched to CANCELING 
06/11/2016 19:20:02	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(10/16) switched to CANCELING 
06/11/2016 19:20:02	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(12/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(1/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(2/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(3/16) switched to CANCELING 
257738 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(4/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(5/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(6/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(7/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(8/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(9/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(10/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(12/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(13/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(14/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(15/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(16/16) switched to CANCELING 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(6/16) switched to CANCELED 
257739 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(9/16) switched to CANCELED 
257740 [flink-akka.actor.default-dispatcher-11] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(8/16) switched to CANCELED 
06/11/2016 19:20:02	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(12/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(1/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(2/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(16/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(5/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(13/16) switched to CANCELED 
06/11/2016 19:20:02	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(10/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(7/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(3/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(14/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(15/16) switched to CANCELED 
06/11/2016 19:20:02	DataSink (org.apache.flink.api.java.Utils$CollectHelper@3cebfc69)(12/16) switched to CANCELED 
06/11/2016 19:20:02	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(4/16) switched to CANCELED 
06/11/2016 19:20:02	Job execution switched to status FAILED.
- Model DFS Serialization *** FAILED ***
  org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
  at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:717)
  at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663)
  at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663)
  at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
  at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
  at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41)
  at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401)
  at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
  at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
  at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
  ...
  Cause: java.lang.Exception: Failed to deploy the task to slot SimpleSlot (10)(1) - 33fa336ad0e9aedc7c4bce2e6990b9ee @ localhost - 16 slots - URL: akka://flink/user/taskmanager_1 - ALLOCATED/ALIVE: Response was not of type Acknowledge
  at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:395)
  at akka.dispatch.OnComplete.internal(Future.scala:247)
  at akka.dispatch.OnComplete.internal(Future.scala:244)
  at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174)
  at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171)
  at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
  at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
  at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
  at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
  at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
  ...
06/11/2016 19:20:02	Job execution switched to status RUNNING.
06/11/2016 19:20:02	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to SCHEDULED 
06/11/2016 19:20:02	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to DEPLOYING 
06/11/2016 19:20:02	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to RUNNING 
06/11/2016 19:20:02	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to FINISHED 
06/11/2016 19:20:02	RangePartition: LocalSample(1/1) switched to SCHEDULED 
06/11/2016 19:20:02	RangePartition: LocalSample(1/1) switched to DEPLOYING 
06/11/2016 19:20:02	RangePartition: PreparePartition(1/1) switched to SCHEDULED 
06/11/2016 19:20:02	RangePartition: PreparePartition(1/1) switched to DEPLOYING 
06/11/2016 19:20:02	RangePartition: LocalSample(1/1) switched to RUNNING 
06/11/2016 19:20:02	RangePartition: PreparePartition(1/1) switched to RUNNING 
06/11/2016 19:20:02	RangePartition: GlobalSample(1/1) switched to SCHEDULED 
06/11/2016 19:20:02	RangePartition: LocalSample(1/1) switched to FINISHED 
06/11/2016 19:20:02	RangePartition: GlobalSample(1/1) switched to DEPLOYING 
06/11/2016 19:20:02	RangePartition: GlobalSample(1/1) switched to RUNNING 
06/11/2016 19:20:02	RangePartition: Histogram(1/1) switched to SCHEDULED 
06/11/2016 19:20:02	RangePartition: Histogram(1/1) switched to DEPLOYING 
06/11/2016 19:20:02	RangePartition: GlobalSample(1/1) switched to FINISHED 
06/11/2016 19:20:02	RangePartition: Histogram(1/1) switched to RUNNING 
06/11/2016 19:20:02	RangePartition: Histogram(1/1) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(2/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(4/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(3/16) switched to SCHEDULED 
06/11/2016 19:20:02	RangePartition: PreparePartition(1/1) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(5/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(7/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(2/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(8/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(9/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(10/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(5/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(11/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(7/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(12/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(13/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(8/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(15/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(11/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(16/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(14/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(10/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(13/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(6/16) switched to SCHEDULED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(15/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(6/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(9/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(3/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(4/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(14/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(16/16) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(12/16) switched to DEPLOYING 
06/11/2016 19:20:02	Reduce (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/1) switched to SCHEDULED 
06/11/2016 19:20:02	Reduce (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/1) switched to DEPLOYING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(2/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(5/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(7/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(10/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(6/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(3/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(8/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(4/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(9/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(15/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(13/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(11/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(14/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(16/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(12/16) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(11/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(5/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(7/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(13/16) switched to FINISHED 
06/11/2016 19:20:02	Reduce (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(1/1) switched to RUNNING 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(15/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(6/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(9/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(3/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(14/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(16/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(8/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(4/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(2/16) switched to FINISHED 
06/11/2016 19:20:02	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:192)) -> Combine (Reduce at org.apache.mahout.flinkbindings.FlinkEngine$.colSums(FlinkEngine.scala:194))(12/16) switched to FINISHED 
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 86.9% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3371

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3371/changes>

Changes:

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

------------------------------------------
[...truncated 63686 lines...]
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(7/16) switched to RUNNING 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(4/16) switched to SCHEDULED 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(4/16) switched to RUNNING 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(4/16) switched to DEPLOYING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(1/16) switched to RUNNING 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(2/16) switched to FINISHED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(2/16) switched to SCHEDULED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(2/16) switched to DEPLOYING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(16/16) switched to FAILED 
java.lang.Exception: Failed to deploy the task to slot SimpleSlot (0)(1) - 09bf0280e80b04bbb5f206c11daef115 @ localhost - 16 slots - URL: akka://flink/user/taskmanager_1 - ALLOCATED/ALIVE: Response was not of type Acknowledge
	at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:395)
	at akka.dispatch.OnComplete.internal(Future.scala:247)
	at akka.dispatch.OnComplete.internal(Future.scala:244)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171)
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(2/16) switched to RUNNING 
06/11/2016 15:40:02	Job execution switched to status FAILING.
java.lang.Exception: Failed to deploy the task to slot SimpleSlot (0)(1) - 09bf0280e80b04bbb5f206c11daef115 @ localhost - 16 slots - URL: akka://flink/user/taskmanager_1 - ALLOCATED/ALIVE: Response was not of type Acknowledge
	at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:395)
	at akka.dispatch.OnComplete.internal(Future.scala:247)
	at akka.dispatch.OnComplete.internal(Future.scala:244)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174)
	at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171)
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(8/16) switched to FINISHED 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(11/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(12/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(13/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(14/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(15/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(16/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(8/16) switched to RUNNING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(1/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(2/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(3/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(4/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(5/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(6/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(7/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(8/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(9/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(10/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(11/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(12/16) switched to CANCELING 
436346 [flink-akka.actor.default-dispatcher-15] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(13/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(14/16) switched to CANCELING 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(11/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(15/16) switched to CANCELING 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(1/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(2/16) switched to CANCELING 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(3/16) switched to CANCELING 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(4/16) switched to CANCELING 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(5/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(6/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(7/16) switched to CANCELING 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(8/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(9/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(10/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(11/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(12/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(13/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(14/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(15/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(16/16) switched to CANCELED 
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(7/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(13/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(16/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(14/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(3/16) switched to CANCELED 
436348 [flink-akka.actor.default-dispatcher-7] ERROR org.apache.flink.runtime.taskmanager.TaskManager  - SubmitTask failed
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:714)
	at org.apache.flink.runtime.taskmanager.Task.startTaskThread(Task.java:401)
	at org.apache.flink.runtime.taskmanager.TaskManager.submitTask(TaskManager.scala:1043)
	at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala:411)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
	at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
	at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
	at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
	at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
	at akka.actor.ActorCell.invoke(ActorCell.scala:487)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
	at akka.dispatch.Mailbox.run(Mailbox.scala:221)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
06/11/2016 15:40:02	DataSink (org.apache.flink.api.java.io.TypeSerializerOutputFormat@38784680)(4/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(15/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN RangePartition: Partition -> Partition(12/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(7/16) switched to CANCELED 
06/11/2016 15:40:02	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(4/16) switched to CANCELED 
442314 [Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (15/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (15/16)
org.apache.flink.runtime.io.network.partition.PartitionNotFoundException: Partition d474fccf5ce3d836e060d1ee90d512c8@4b58fed97961f6963e2fadeaf97ae34e not found.
	at org.apache.flink.runtime.io.network.partition.ResultPartitionManager.createSubpartitionView(ResultPartitionManager.java:76)
	at org.apache.flink.runtime.io.network.partition.consumer.LocalInputChannel.requestSubpartition(LocalInputChannel.java:103)
	at org.apache.flink.runtime.io.network.partition.consumer.LocalInputChannel$1.run(LocalInputChannel.java:136)
	at java.util.TimerThread.mainLoop(Timer.java:555)
	at java.util.TimerThread.run(Timer.java:505)
442315 [Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (15/16)] ERROR org.apache.flink.runtime.taskmanager.Task  - FATAL - exception in task resource cleanup
java.lang.IllegalStateException: Memory manager has been shut down.
	at org.apache.flink.runtime.memory.MemoryManager.releaseAll(MemoryManager.java:468)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:661)
	at java.lang.Thread.run(Thread.java:745)
442317 [Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (6/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (6/16)
org.apache.flink.runtime.io.network.partition.PartitionNotFoundException: Partition ac59c361e85adb032d116a6aa5e8e65c@5e4270569522bc026a0525ada9101c46 not found.
	at org.apache.flink.runtime.io.network.partition.ResultPartitionManager.createSubpartitionView(ResultPartitionManager.java:76)
	at org.apache.flink.runtime.io.network.partition.consumer.LocalInputChannel.requestSubpartition(LocalInputChannel.java:103)
	at org.apache.flink.runtime.io.network.partition.consumer.LocalInputChannel$1.run(LocalInputChannel.java:136)
	at java.util.TimerThread.mainLoop(Timer.java:555)
	at java.util.TimerThread.run(Timer.java:505)
442318 [Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (6/16)] ERROR org.apache.flink.runtime.taskmanager.Task  - FATAL - exception in task resource cleanup
java.lang.IllegalStateException: Memory manager has been shut down.
	at org.apache.flink.runtime.memory.MemoryManager.releaseAll(MemoryManager.java:468)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:661)
	at java.lang.Thread.run(Thread.java:745)
442329 [Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (11/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (11/16)
org.apache.flink.runtime.io.network.partition.PartitionNotFoundException: Partition 2047d64a60ccdfe974ab92510e31a4bd@404e1b3c0f37f3b015ed3827da6163ba not found.
	at org.apache.flink.runtime.io.network.partition.ResultPartitionManager.createSubpartitionView(ResultPartitionManager.java:76)
	at org.apache.flink.runtime.io.network.partition.consumer.LocalInputChannel.requestSubpartition(LocalInputChannel.java:103)
	at org.apache.flink.runtime.io.network.partition.consumer.LocalInputChannel$1.run(LocalInputChannel.java:136)
	at java.util.TimerThread.mainLoop(Timer.java:555)
	at java.util.TimerThread.run(Timer.java:505)
442330 [Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (11/16)] ERROR org.apache.flink.runtime.taskmanager.Task  - FATAL - exception in task resource cleanup
java.lang.IllegalStateException: Memory manager has been shut down.
	at org.apache.flink.runtime.memory.MemoryManager.releaseAll(MemoryManager.java:468)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:661)
	at java.lang.Thread.run(Thread.java:745)
442332 [Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (5/16)] ERROR org.apache.flink.runtime.operators.BatchTask  - Error in task code:  Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (5/16)
org.apache.flink.runtime.io.network.partition.PartitionNotFoundException: Partition 27b902bc3f6bc85e348eacba15e202d1@6854e73192d275d789fa43d97afd9b23 not found.
	at org.apache.flink.runtime.io.network.partition.ResultPartitionManager.createSubpartitionView(ResultPartitionManager.java:76)
	at org.apache.flink.runtime.io.network.partition.consumer.LocalInputChannel.requestSubpartition(LocalInputChannel.java:103)
	at org.apache.flink.runtime.io.network.partition.consumer.LocalInputChannel$1.run(LocalInputChannel.java:136)
	at java.util.TimerThread.mainLoop(Timer.java:555)
	at java.util.TimerThread.run(Timer.java:505)
442333 [Combine (Reduce at org.apache.mahout.flinkbindings.blas.FlinkOpAt$.sparseTrick(FlinkOpAt.scala:61)) (5/16)] ERROR org.apache.flink.runtime.taskmanager.Task  - FATAL - exception in task resource cleanup
java.lang.IllegalStateException: Memory manager has been shut down.
	at org.apache.flink.runtime.memory.MemoryManager.releaseAll(MemoryManager.java:468)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:661)
	at java.lang.Thread.run(Thread.java:745)
06/11/2016 15:40:42	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(14/16) switched to CANCELED 
06/11/2016 15:40:42	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(12/16) switched to CANCELED 
06/11/2016 15:40:42	CHAIN MapPartition (MapPartition at org.apache.mahout.flinkbindings.drm.RowsFlinkDrm.asBlockified(FlinkDrm.scala:52)) -> Map (Map at org.apache.mahout.flinkbindings.blas.FlinkOpMapBlock$.apply(FlinkOpMapBlock.scala:37)) -> FlatMap (FlatMap at org.apache.mahout.flinkbindings.drm.BlockifiedFlinkDrm.asRowWise(FlinkDrm.scala:93))(15/16) switched to CANCELED 
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Apache Mahout
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Mahout Build Tools
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Mahout Build Tools ................................. SUCCESS [  2.424 s]
[INFO] Apache Mahout ...................................... SUCCESS [  0.144 s]
[INFO] Mahout Math ........................................ SUCCESS [01:19 min]
[INFO] Mahout HDFS ........................................ SUCCESS [  4.194 s]
[INFO] Mahout Map-Reduce .................................. SUCCESS [13:05 min]
[INFO] Mahout Integration ................................. SUCCESS [ 46.705 s]
[INFO] Mahout Examples .................................... SUCCESS [ 24.518 s]
[INFO] Mahout Math Scala bindings ......................... SUCCESS [05:17 min]
[INFO] Mahout H2O backend ................................. SUCCESS [03:32 min]
[INFO] Mahout Spark bindings .............................. SUCCESS [02:26 min]
[INFO] Mahout Flink bindings .............................. FAILURE [  03:15 h]
[INFO] Mahout Spark bindings shell ........................ SKIPPED
[INFO] Mahout Release Package ............................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 03:42 h
[INFO] Finished at: 2016-06-11T18:47:34+00:00
[INFO] Final Memory: 62M/787M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.scalatest:scalatest-maven-plugin:1.0:test (test) on project mahout-flink_2.10: There are test failures -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :mahout-flink_2.10
Compressed 171.05 MB of artifacts by 86.9% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3370

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3370/changes>

Changes:

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

------------------------------------------
[...truncated 59078 lines...]
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(4/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(14/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(9/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(13/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(7/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(8/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(16/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(11/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(15/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(10/16) switched to RUNNING 
06/11/2016 10:56:55	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(12/16) switched to RUNNING 
06/11/2016 10:56:56	Reduce (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/1) switched to SCHEDULED 
06/11/2016 10:56:56	Reduce (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/1) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(3/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(5/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(14/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(7/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(4/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(12/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(13/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(9/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(16/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(15/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(8/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(11/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(6/16) switched to FINISHED 
06/11/2016 10:56:56	Reduce (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/1) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(2/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@754a88b9)(1/1) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@754a88b9)(1/1) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(10/16) switched to FINISHED 
06/11/2016 10:56:56	Reduce (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/1) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@754a88b9)(1/1) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@754a88b9)(1/1) switched to FINISHED 
06/11/2016 10:56:56	Job execution switched to status FINISHED.
(1,1)
06/11/2016 10:56:56	Job execution switched to status RUNNING.
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(1/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(1/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(2/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(2/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(3/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(3/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(4/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(4/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(5/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(5/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(6/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(6/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(7/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(7/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(8/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(8/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(9/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(9/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(10/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(10/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(11/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(11/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(12/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(12/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(13/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(13/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(14/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(14/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(15/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(15/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(16/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(16/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(1/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(2/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(3/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(4/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(5/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(6/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(7/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(10/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(9/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(8/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(12/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(13/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(11/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(14/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(15/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(16/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(9/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(9/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(7/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(7/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(1/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(11/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(1/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(11/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(2/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(13/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(12/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(6/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(4/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(10/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(13/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(12/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(10/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(2/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(4/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(6/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(14/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(8/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(8/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(9/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(7/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(1/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(11/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(12/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(14/16) switched to DEPLOYING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(16/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(2/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(16/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(6/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(15/16) switched to SCHEDULED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(13/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(15/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(10/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(14/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(8/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(9/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(7/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(3/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(3/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(16/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(9/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(4/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(1/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(11/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(13/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(12/16) switched to RUNNING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(15/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(10/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(7/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(2/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(4/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(6/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(1/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(8/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(16/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(13/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(14/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(11/16) switched to FINISHED 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(3/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(15/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(12/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(10/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(4/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(2/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(3/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(6/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(14/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(16/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(15/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(3/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(8/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(5/16) switched to SCHEDULED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(5/16) switched to DEPLOYING 
06/11/2016 10:56:56	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75))(5/16) switched to FINISHED 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(5/16) switched to RUNNING 
06/11/2016 10:56:56	DataSink (org.apache.flink.api.java.Utils$CollectHelper@a36d42c)(5/16) switched to FINISHED 
06/11/2016 10:56:56	Job execution switched to status FINISHED.
06/11/2016 10:56:57	Job execution switched to status RUNNING.
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(2/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(2/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(3/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(3/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(4/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(4/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(5/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(5/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(6/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(6/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(7/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(7/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(8/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(8/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(9/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(9/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(10/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(10/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(11/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(11/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(12/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(12/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(13/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(13/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(14/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(14/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(15/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(15/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(16/16) switched to SCHEDULED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(16/16) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(2/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(3/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(4/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(5/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(6/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(10/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(12/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(13/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(14/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(15/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(7/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(16/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(9/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(8/16) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(11/16) switched to RUNNING 
06/11/2016 10:56:57	Reduce (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/1) switched to SCHEDULED 
06/11/2016 10:56:57	Reduce (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/1) switched to DEPLOYING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(2/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(12/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(5/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(11/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(8/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(13/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(4/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(14/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(6/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(15/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(7/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(10/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(3/16) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/16) switched to FINISHED 
06/11/2016 10:56:57	Reduce (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/1) switched to RUNNING 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(16/16) switched to FINISHED 
06/11/2016 10:56:57	DataSink (org.apache.flink.api.java.Utils$CollectHelper@2a4343ab)(1/1) switched to SCHEDULED 
06/11/2016 10:56:57	DataSink (org.apache.flink.api.java.Utils$CollectHelper@2a4343ab)(1/1) switched to DEPLOYING 
06/11/2016 10:56:57	Reduce (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(1/1) switched to FINISHED 
06/11/2016 10:56:57	CHAIN DataSource (at org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:396) (org.apache.flink.api.scala.hadoop.mapred.HadoopInputFo) -> Map (Map at org.apache.mahout.flinkbindings.FlinkEngine$.drmDfsRead(FlinkEngine.scala:75)) -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:79)) -> Combine (Reduce at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dim$lzycompute(CheckpointedFlinkDrm.scala:83))(9/16) switched to FINISHED 
06/11/2016 10:56:57	DataSink (org.apache.flink.api.java.Utils$CollectHelper@2a4343ab)(1/1) switched to RUNNING 
06/11/2016 10:56:57	DataSink (org.apache.flink.api.java.Utils$CollectHelper@2a4343ab)(1/1) switched to FINISHED 
06/11/2016 10:56:57	Job execution switched to status FINISHED.
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 87.7% relative to #3363
Recording test results
Publishing Javadoc

Build failed in Jenkins: Mahout-Quality #3369

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Mahout-Quality/3369/changes>

Changes:

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] Rolling back Mahout 0.12.2 Release candidate, thanks github connectivity

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

[smarthi] [maven-release-plugin] prepare release mahout-0.12.2

[smarthi] [maven-release-plugin] prepare for next development iteration

[smarthi] [maven-release-plugin] rollback the release of mahout-0.12.2

------------------------------------------
[...truncated 58262 lines...]
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(15/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(5/16) switched to RUNNING 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(11/16) switched to RUNNING 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(13/16) switched to RUNNING 
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(12/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(14/16) switched to RUNNING 
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(7/16) switched to FINISHED 
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(9/16) switched to FINISHED 
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(6/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(7/16) switched to RUNNING 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(9/16) switched to RUNNING 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(6/16) switched to RUNNING 
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(3/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(12/16) switched to RUNNING 
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(8/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(16/16) switched to RUNNING 
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(14/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(15/16) switched to RUNNING 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(8/16) switched to RUNNING 
06/11/2016 07:12:28	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(16/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(3/16) switched to RUNNING 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(14/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(3/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(11/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(5/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(9/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(4/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(10/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(6/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(8/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(12/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(13/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(2/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(15/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(7/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(16/16) switched to FINISHED 
06/11/2016 07:12:28	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@36356e30)(1/16) switched to FINISHED 
06/11/2016 07:12:28	Job execution switched to status FINISHED.
06/11/2016 07:12:29	Job execution switched to status RUNNING.
06/11/2016 07:12:29	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to SCHEDULED 
06/11/2016 07:12:29	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to DEPLOYING 
06/11/2016 07:12:29	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to RUNNING 
06/11/2016 07:12:29	RangePartition: LocalSample(1/1) switched to SCHEDULED 
06/11/2016 07:12:29	RangePartition: LocalSample(1/1) switched to DEPLOYING 
06/11/2016 07:12:29	DataSource (at org.apache.mahout.flinkbindings.FlinkEngine$.parallelize(FlinkEngine.scala:273) (org.apache.flink.api.java.io.CollectionInputFormat))(1/1) switched to FINISHED 
06/11/2016 07:12:29	RangePartition: PreparePartition(1/1) switched to SCHEDULED 
06/11/2016 07:12:29	RangePartition: PreparePartition(1/1) switched to DEPLOYING 
06/11/2016 07:12:29	RangePartition: LocalSample(1/1) switched to RUNNING 
06/11/2016 07:12:29	RangePartition: PreparePartition(1/1) switched to RUNNING 
06/11/2016 07:12:29	RangePartition: GlobalSample(1/1) switched to SCHEDULED 
06/11/2016 07:12:29	RangePartition: GlobalSample(1/1) switched to DEPLOYING 
06/11/2016 07:12:29	RangePartition: LocalSample(1/1) switched to FINISHED 
06/11/2016 07:12:29	RangePartition: GlobalSample(1/1) switched to RUNNING 
06/11/2016 07:12:29	RangePartition: Histogram(1/1) switched to SCHEDULED 
06/11/2016 07:12:29	RangePartition: Histogram(1/1) switched to DEPLOYING 
06/11/2016 07:12:29	RangePartition: GlobalSample(1/1) switched to FINISHED 
06/11/2016 07:12:29	RangePartition: Histogram(1/1) switched to RUNNING 
06/11/2016 07:12:29	RangePartition: Histogram(1/1) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(1/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(2/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(3/16) switched to SCHEDULED 
06/11/2016 07:12:29	RangePartition: PreparePartition(1/1) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(1/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(4/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(5/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(6/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(7/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(8/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(3/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(6/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(9/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(10/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(12/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(11/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(8/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(13/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(14/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(10/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(16/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(2/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(16/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(15/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(14/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(13/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(11/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(12/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(9/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(7/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(5/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(4/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(15/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(1/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(1/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(1/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(10/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(10/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(16/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(16/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(2/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(2/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(6/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(6/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(14/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(14/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(3/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(3/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(11/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(13/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(9/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(11/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(7/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(5/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(9/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(5/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(13/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(7/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(15/16) switched to SCHEDULED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(6/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(8/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(16/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(13/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(11/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(12/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(9/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(14/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(7/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(10/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(2/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(3/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(15/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(12/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(8/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(12/16) switched to DEPLOYING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(8/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(5/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(4/16) switched to SCHEDULED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(4/16) switched to DEPLOYING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(4/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(15/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(1/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(1/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(2/16) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(14/16) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(10/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(10/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(16/16) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(3/16) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(11/16) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(6/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(16/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(9/16) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(13/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(2/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(7/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(12/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(15/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(8/16) switched to FINISHED 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(12/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(8/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(4/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(4/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(6/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(11/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(15/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(3/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(13/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(7/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(5/16) switched to RUNNING 
06/11/2016 07:12:29	CHAIN RangePartition: Partition -> Partition -> Map (Map at org.apache.mahout.flinkbindings.drm.CheckpointedFlinkDrm.dfsWrite(CheckpointedFlinkDrm.scala:231))(5/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(14/16) switched to RUNNING 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(2/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(1/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(16/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(6/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(3/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(12/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(8/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(15/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(14/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(5/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(4/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(10/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(7/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(13/16) switched to FINISHED 
06/11/2016 07:12:29	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(11/16) switched to FINISHED 
06/11/2016 07:12:39	DataSink (org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat@1210c01c)(9/16) switched to FAILED 
java.lang.Exception: Failed to send ExecutionStateChange notification to JobManager
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage$3$$anonfun$apply$2.apply(TaskManager.scala:398)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage$3$$anonfun$apply$2.apply(TaskManager.scala:382)
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.processBatch$1(BatchingExecutor.scala:67)
	at akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:82)
	at akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
	at akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
	at scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
	at akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:58)
	at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41)
	at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: akka.pattern.AskTimeoutException: Ask timed out on [Actor[akka://flink/user/jobmanager_1#667045242]] after [10000 ms]
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)

06/11/2016 07:12:39	Job execution switched to status FAILING.
java.lang.Exception: Failed to send ExecutionStateChange notification to JobManager
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage$3$$anonfun$apply$2.apply(TaskManager.scala:398)
	at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$org$apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage$3$$anonfun$apply$2.apply(TaskManager.scala:382)
	at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
	at akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.processBatch$1(BatchingExecutor.scala:67)
	at akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:82)
	at akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
	at akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
	at scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
	at akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:58)
	at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41)
	at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: akka.pattern.AskTimeoutException: Ask timed out on [Actor[akka://flink/user/jobmanager_1#667045242]] after [10000 ms]
	at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
	at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
	at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
	at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
	at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
	at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
	at java.lang.Thread.run(Thread.java:745)
06/11/2016 07:12:39	Job execution switched to status FAILED.
Build timed out (after 223 minutes). Marking the build as failed.
Build was aborted
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Compressed 171.05 MB of artifacts by 86.9% relative to #3363
Recording test results
Publishing Javadoc