You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-dev@hadoop.apache.org by Steve Loughran <st...@hortonworks.com> on 2015/04/02 12:51:14 UTC

All the jenkins tests are failing

Can I draw attention to the fact that within the last week, every single Hadoop jenkins run has started failing.

Ports in use
https://issues.apache.org/jira/browse/YARN-3433
https://issues.apache.org/jira/browse/HADOOP-11788

NPEs
https://issues.apache.org/jira/browse/HADOOP-11789 NPE in TestCryptoStreamsWithOpensslAesCtrCryptoCodec
https://issues.apache.org/jira/browse/HDFS-8043 MiniDFSCluster

MR tests are showing NPEs during FS creation, with ~190 tests down
https://builds.apache.org/view/H-L/view/Hadoop/job/Hadoop-Mapreduce-trunk-Java8/

We are trying to get a release out the door, trying to get jenkins to test every single patch to see its not a regression -yet with this failure rate its impossible to do so.

The whole purpose of a Continuous Integration system is to keep on top of changes, to catch regressions on the individual commits, to preflight check new code, to help identify  exactly when and where something broke. It's been one of the most tangible changes to Software Engineering in recent years. I remember when Apache Gump used to be able to build the entire ASF source tree every night -a full stack regression test of everything from Ant up. I remember reading Fowler's blog posts on it, in 2001, (see http://www.martinfowler.com/articles/continuousIntegration.html ), installing CruiseControl and loving it so much we gladly dedicated 20 pages to in our Ant book; using it in projects ourselves to implement some of the first continuous deployment pipelines.

For all this to work, people have to care about the build, to care about the build passing its tests, and to keep atop of it when it breaks.

Please look at the test reports and help fix those that are breaking. And then start paying attention to those Jenkins emails

-Steve