You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2018/03/15 12:12:15 UTC

Build failed in Jenkins: beam_PerformanceTests_TFRecordIOIT_HDFS #2

See <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/2/display/redirect?page=changes>

Changes:

[joey.baruch] Add javadoc to ConsoleIO

[herohde] Initial sketches of a Go SDK

[herohde] Initial version of the direct style w/ direct runner. Incomplete.

[herohde] Add Data as UserFn context w/ immediate value.

[herohde] Added no-I/O wordcount for profiling.

[herohde] Fleshed out possible approach to generic transformations.

[herohde] Add “dag” example that use multiplexing and side input.

[herohde] Added a more complex DAG example.

[herohde] Add yatzy example with more complex construction-time setup

[herohde] Add proto for Fn API

[herohde] Add beam.Composite helper for the most common pattern to align with java

[herohde] Move pipeline-construction time errors into an accumulator

[herohde] Add Dataflow job and Fn API clients. Incomplete.

[herohde] Add binary cross-compile and upload to Dataflow runner. Incomplete.

[herohde] Add tentative runner indirection (default: local).

[herohde] Made data flow runner detect user main for cross-compilation.

[herohde] Remove error accumulation in favor of panic.

[herohde] Improve Dataflow translation of coders, side input and composite names.

[herohde] Fix name for AsView.

[herohde] Use 2 grpc endpoints in harness

[herohde] Add gRPC harness logging

[herohde] Flesh out harness and serialization further.

[herohde] Made the dataflow runner wait for job termination by default

[herohde] beam:

[herohde] beam:

[herohde] combinefn.go: fix compilation issues

[herohde] Improve dataflow serialization and execution. Incomplete.

[herohde] Sleep 30 sec in wordcap to allow logs to propagate to Cloud Logging.

[herohde] Move the 30s sleep for logging to the harness instead of in WordCap.

[herohde] Post-review updates.

[herohde] Doc updates.

[herohde] Flesh out coders. Incomplete.

[herohde] Added prototype implementation of more coders and the runner source.

[herohde] dofn: illustrates how dofns are written.

[herohde] beam: add viewfn and windowfn to side inputs match support Beam 1.0

[herohde] dofn: timers

[herohde] Complete revamp: coders, graph and execution use element-wise

[herohde] Fix coder encoding for Dataflow side input. Otherwise, the job is

[herohde] Added more godoc comments to graph types.

[herohde] Added more comments plus made local GBK use coder equality.

[herohde] Added Flatten support and “forest” example that uses it.

[herohde] Move bigqueryio to defunct

[herohde] Make forest example print less

[herohde] Add external struct registry and serialization.

[herohde] Updated comments in node.go.

[herohde] Replace real type with 'full type' since that's the current term.

[herohde] Refactor Fn API dependency.

[herohde] Added more comments to the runner/dataflow and runner/beamexec packages

[herohde] Fix most go vet issues

[herohde] Make core operations panic to cut down on the error propagation

[herohde] Add more comments to the graph package.

[herohde] Add DoFn wrapper to handle either function or (ptr to) struct

[herohde] Fix remaining go vet warnings.

[herohde] Code review for beam/graph/coder package.

[herohde] Code review of the runtime/graphx package.

[herohde] Remove Data options in favor of using a Fn struct

[herohde] Code review of the beam/graph/userfn package.

[herohde] Code review for beam/graph package.

[herohde] godoc for runtime/graphx

[herohde] Add support for []T and Combine functions

[herohde] Add adapted documentation from the Java SDK to the beam package

[herohde] Update snapshot of Fn API.

[herohde] Add experiments flag to the Dataflow runner

[herohde] Remove context arg from beamexec.Init

[herohde] Migration to Runner API.

[herohde] Add support for creating DOT graphs.

[herohde] Make pretty printing of types and coders more concise

[herohde] Add flexible Signature to aid type checking

[herohde] Adding unit testability to harness translation.

[herohde] Fix crash due to initialization order

[herohde] Add CreateValues and Impulse

[herohde] Add Runner API support for WindowingStrategy.

[herohde] Run goimports on baseline.

[herohde] Fix encoding of global window strategy.

[herohde] Ensure the windowed value is atomically encoded.

[herohde] Limit gRPC messages to max size.

[herohde] Developer conveniences for running jobs.

[herohde] Fix sends to not close the network channel.

[herohde] Add re-iterable side input

[herohde] Add per-key Combine

[herohde] Add Min

[herohde] Reorganize non-user-facing code into core

[herohde] Make type register reject unnamed or predeclared types

[herohde] Add type specialization tool

[herohde] Don't run grpc plugin in generate phase.

[herohde] Fix import reference path for runner API proto.

[herohde] Revamp runner registration as _ imports

[herohde] Add stats.Max and Mean

[herohde] Add global pipeline options

[herohde] Unify global and per-key combiners

[herohde] Add beam convenience wrapper for imports and runner selection

[herohde] Add session recording and CPU profiling to harness.

[herohde] Add ptest and passert for testing pipelines

[herohde] Add GCS and glob support to textio

[herohde] Add BigQuery IO and examples

[herohde] Adds a session runner for testing.

[herohde] Add Partition and dynamic functions

[herohde] Adding example that returns 10 words that contain provided search

[herohde] Remove duplicate LOG line

[herohde] Enable Combine Fns in Dataflow runner by modifying translation.

[herohde] Fixing type bug by dropping T and using underlying type of value in

[herohde] Adding graph validation at build time.

[herohde] Import the Fn API changes.

[herohde] Simple changes to support new Fn API coder changes.

[herohde] Update translator to work with new Fn API changes.

[herohde] Use appropriate equality tests.

[herohde] Fix test to not use path of package.

[herohde] Renaming directory to match package name.

[herohde] Fixing random nits in comments.

[herohde] Modify build command to avoid bash.

[herohde] Fixing selected golint issues.

[herohde] Addressing import review comments.

[herohde] Add coder specialization for bytes/strings.

[herohde] Adding unit tests to stats.

[herohde] Fixing typo.

[herohde] Add beam.External

[herohde] Fix grpc.Dial calls to block properly.

[herohde] Creates a symtab verifier by running Sym2Addr and Addr2Sym in a binary.

[herohde] Add spec field to help interpretation of payload.

[herohde] Use beam.T alias for typex.T etc with Go 1.9

[herohde] Move shared GCP options to a separate package

[herohde] Update portability protos

[herohde] Remove old source/sink from beam package

[herohde] Add context-aware logging for both pipeline-submission time and runtime

[herohde] Fix coder inference for strings.

[herohde] Improve tornadoes example

[herohde] Fix beam.External to map arguments correctly.

[herohde] Added comments to yatzy and forest

[herohde] Add comments to tornadoes from the java counterpart

[herohde] Rename Pipeline Composite to Scope

[herohde] Add 3 progressive wordcount examples

[herohde] Clarify comments in wordcount pipelines

[herohde] Add apache 2.0 license to files

[herohde] Updates to examples.

[herohde] Adding more godoc for the main beam package.

[herohde] Update to new proto structure

[herohde] Split Combine and fields in to global and per-key variants

[herohde] Refactor Flatten of a single []T into Explode

[herohde] Rename local runner to direct runner

[herohde] Fix argument index error in ParDo execution

[herohde] Add Apache copyright header to files that need it.

[herohde] Made debug.Head not just work per bundle

[herohde] Impose a total ordering on Fn parameters.

[herohde] Rename Dedup to Distinct for consistency with other SDKs

[herohde] Add coder to model coder translation

[herohde] Simplify harness coder translation

[herohde] Split Pipeline into Pipeline and Scope

[herohde] Relocate Go SDK code

[herohde] Fix Go SDK maven build

[herohde] Move Go SKD to latest version of bigquery

[herohde] Add Go SDK container image

[herohde] Add Go SDK README

[herohde] Update version for Go Dataflow pipelines

[herohde] Make Scope a value type

[herohde] Add Go graph/pipeline translation

[herohde] Stage Go model pipeline for Dataflow

[herohde] Use pipeline unmarhaller in runtime harness

[herohde] CR: [BEAM-3287] Use model pipelines in Go SDK

[herohde] CR: [BEAM-3287] Use model pipelines in Go SDK

[herohde] Fix name of syscallx ErrUnsupported

[herohde] Allow any named type to be registered and serialized as external

[herohde] Add more package comments for core packages

[herohde] Make Go SDK External a graph primitive

[herohde] Cache Go runtime symbol lookups

[github] Fix code comment to match code

[wcn] Fix storagePath variable.

[wcn] BEAM-3368 fix translation for external

[robertwb] [BEAM-3356] Add Go SDK int and varint custom coders (#4276)

[lcwik] BEAM-3361 Increase Go gRPC message size

[herohde] Go SDK runtime revamp

[lcwik] Add a few function call overhead benchmarks

[lcwik] Add type-specialized emitters

[lcwik] BEAM-3324 improve symtab memory usage

[lcwik] BEAM-3324 improve symtab memory usage

[lcwik] BEAM-3324 improve symtab memory usage

[lcwik] Store objects in pool so they can be reused.

[lcwik] Add builtin varint coder

[herohde] Type-specialize custom decoders and encoders in Go SDK runtime

[herohde] Type-specialize iterators and side input in the Go SDK

[herohde] Add warnings if Go runtime registrations are overwritten

[herohde] Add reusable element coders in Go SDK runtime

[wcn] Updated translater to preserve payload and its URN.

[github] NotImplementedErrror --> NotImplementedError

[herohde] Initial version of type-specialized general reflective calls

[herohde] Add general-purpose untyped callers in Go SDK runtime

[herohde] Use fast caller for filter transform predicate

[abhyankar] Add support for ValueProvider in JdbcIO.DataSourceConfiguration methods

[herohde] CR: Clarified comment on caller template

[robertwb] Remove legacy windowfn translation.

[herohde] Fix value encoding for Create

[github] Update BUILD.md

[lcwik] BEAM-3473: Fix GroupByKey iterators to be initialized.

[lcwik] BEAM-3474 Include stacks in panic messages.

[shashank] fix serialization error in BigQueryIO's DynamicDestinations

[shashank] correct side input check in BigQueryIO DynamicDestination for pull

[lcwik] BEAM-3299: Add source reporting support.

[lcwik] Remove GetId() call from under lock.

[lcwik] Add additional comments about concurrency invariants.

[lcwik] Add initialization of active plans map.

[zoy] Increasing BatchElements's max_batch_size to 10K

[robertwb] Curry CombineFn arguments into runner API protos.

[batbat] Added an example pipeline that uses stateful processing to output team

[markliu] [BEAM-2762] Python code coverage report in Postcommit

[markliu] fixup! Exclude auto-generated files in portability/api

[markliu] fixup! Clean up configuration and use better way to exclude cover in mvn

[markliu] fixup! Add more comments

[wcn] Update generated version of Fn API code.

[Pablo] Tracking of time spent reading side inputs, and bytes read in Dataflow.

[Pablo] Fix lint issues

[Pablo] Fixing counter names

[Pablo] Using comments to improve changes

[Pablo] Placing feature behind an experiment flag.

[Pablo] Fixing lint issue

[Pablo] Rebasing

[Pablo] Added cythonization and test

[Pablo] Fixing lint issue

[Pablo] Modifying the experiment flag check to reduce performance impact.

[Pablo] Addressing comments. Profile pending.

[Pablo] Reducing impact - Tracking only Bytes if experiment is active.

[tgroh] Add a BundleProcessor to SdkHarnessClient

[lcwik] Renamed Go runtime Caller to Func and added name

[lcwik] Use reflectx.Func as the fundamental function representation

[lcwik] CR: fix DynFn comments

[lcwik] CR: fix comments

[herohde] Avoid reflect.Value conversions in Go runtime

[robertwb] [BEAM-3490] Picklable side inputs for FnApi Workers.

[robertwb] Manually specify direct runner for global-state modifying tests.

[robertwb] FakeSource requires direct runner.

[robertwb] Explicitly use DirectRunner in DirectRunner tests.

[markliu] fixup! Improve comments

[altay] Disable combiner lifting optimization in DataflowRunner for fnapi

[wcn] Allow grpcx.Dial to support overrides.

[jb] [BEAM-3507] Add a way to specify the batch size in JdbcIO Write

[jbonofre] [BEAM-3507] Add DEFAULT_BATCH_SIZE and use batchSize in default

[iemejia] [BEAM-3492] Fix spark 2.10 dependency leak on hadoop tests

[iemejia] [BEAM-3492] Force netty version when running with the spark runner

[lcwik] [BEAM-3427] Update build to Java 8 (before clean out) (#4424)

[sduskis] [BEAM-3412] Upgrade Cloud Bigtable to 1.0.0 The 1.0.0 Cloud Bigtable

[sduskis] BigtableServiceImplTest now uses a List of ranges. The PR did not

[sduskis] Updating Cloud Bigtable dependnecy to 1.0.0 in build.gradle.

[ekirpichov] Introduces the Wait transform

[github] [BEAM-2963] Remove layer of indirection in output name mapping in

[ehudm] Pass PipelineOptions to FileSystem constructor.

[altay] Disable combiner lifting when only the streaming flag is set.

[Pablo] Addressing comments

[Pablo] Improved IO Target documentation

[tgroh] Register Environments in SdkComponents

[Pablo] Improving documentation

[tgroh] Move BeamFnDataInboundObserver to java-fn-execution

[tgroh] Add an implementation of the Fn API Data Service

[robertwb] Document DirectRunnerOnly tests.

[rober] Use a typeswitch instead of reflect.Convert when encoding strings or

[robert] Add CONTRIBUTING.md

[lcwik] [BEAM-3008] Adds parameters templatization for Bigtable (#4357)

[lcwik] Migrate Flink ValidatesRunner to Gradle

[lcwik] Increment the Dataflow runner major version to 7.

[zhouhai02] Update copyright date in NOTICE

[sduskis] Using 1.0.0-pre3 for bigtable-proto. This should allow Cloud Bigtable's

[sduskis] Updating build.gradle with a bigtable_proto_version of 1.0.0-pre3

[github] Update coder.go

[robertwb] [BEAM-3126] Adding a new Flatten test to Python SDK. (#4463)

[rober] Update printed gcloud cancel commands to include the job's region.

[sduskis] Fixing a bad merge in BigtableServiceImpl.

[kirpichov] Code compiles after auto-transition to lambda

[kirpichov] google-java-format

[kirpichov] Removes unnecessary explicit type arguments

[kirpichov] google-java-format

[kirpichov] checkstyle fixups

[kirpichov] More removal of explicit type arguments

[kirpichov] google-java-format

[kirpichov] Manually fixed a couple cases of bad comment formatting

[kirpichov] Manual fixup of some call sites where lambdas mess up coder inference

[kirpichov] A couple of final example fixups

[lcwik] Upgrade Jenkins jobs to use Maven 3.5.2

[iemejia] [BEAM-3432] Remove hadoop/jdk1.8-tests module

[lcwik] [BEAM-2728] Add Count-Min Sketch in sketching extension

[lcwik] Small fix in SketchCoder

[lcwik] Make Sketch AutoValue + Javadoc update

[lcwik] Optimize coder memory use

[lcwik] [BEAM-3160] Fix issue where we would choose which coder to use

[lcwik] Initial post-release snapshot test

[lcwik] Make the snapshot and url parameters to the job to allow installing a

[lcwik] Make the snapshot and url parameters to the job to allow installing a

[lcwik] Rename TestHelper to TestScripts, it() to intent()

[chamikara] [BEAM-3060] Support for Perfkit execution of file-based-io-tests on HDFS

[Pablo] Addressing comments.

[jbonofre] [BEAM-3428] Merge Java8 examples in "main" Java examples

[jbonofre] [BEAM-3428] Replace MinimalWordCount with Java8 one

[echauchot] [BEAM-3534] Add a spark validates runner test for metrics sink in

[coheigea] BEAM-3533 - Replace hard-coded UTF-8 Strings

[tgroh] Retrieve Environments from PTransforms

[jbonofre] [BEAM-3466] Remove JDK 7 references in Jenkins

[kirpichov] [BEAM-3083] Do not call getSchema() and getTable() on every element

[altay] Use non-deprecated version of Futures.transform

[aromanenko.dev] [BEAM-3539] BigtableIO.Write javadoc of some methods is incorrect

[iemejia] Remove unneeded profile for javadoc on Java 8

[iemejia] Remove unneeded explicit Java 8 references on maven-compiler-plugin

[iemejia] Fix doc error on hadoop-input-format ITs after move to Java 8 only tests

[iemejia] Remove references to non-existent examples:java8 module in gradle

[iemejia] Remove references to java 7/8 only examples from the README

[iemejia] Remove some comments on Java 7/8 only stuff that don't make sense

[kenn] Add a test for an event time timer loop in ParDo

[lcwik] [BEAM-2273] Cleanup examples Maven Archetype to copy in a clean state

[XuMingmin] [BEAM-3526] KakfaIO support for finalizeCheckpoint() (#4481)

[kedin] Add Avro dependency to KafkaIO

[rangadi] Remove an unused test dependency in KafkaIO.

[lcwik] [Beam-2500] Add S3FileSystem to SDKs/Java/IO

[lcwik] implement serializing AWS credentials provider

[lcwik] fixup! Clarify error message is received from SDK harness

[iemejia] Fix modules that were activated only on Java 8 profile

[robert] Replace reflective convert to direct convert.

[iemejia] [BEAM-3275] Fix ValidatesRunner Spark runner after the Kafka update

[iemejia] Refactor code into idiomatic Java 8 style

[iemejia] Fix missing gearpump module activated only on Java 8 profile

[iemejia] Add missing amazon-web-services module from javadoc

[jbonofre] [BEAM-2271] Exclude files not intended to be in the releases source

[jbonofre] Typo fix in MinimalWordCount example

[Mottyg1] [BEAM-675] Introduce message mapper in JmsIO

[iemejia] Update maven-shade-plugin version to 3.1.0

[iemejia] Update maven-compiler-plugin version to 3.7.0

[iemejia] Update maven-dependency-plugin version to 3.0.2

[iemejia] Update maven-surefire-plugin version to 2.20.1

[iemejia] Update maven-failsafe-plugin version to 2.20.1

[iemejia] Update maven-assembly-plugin version to 3.1.0

[iemejia] Update versions-maven-plugin version to 2.5

[iemejia] Update findbugs-maven-plugin version to 3.0.5

[iemejia] Update license-maven-plugin version to 1.14

[iemejia] Update jacoco-maven-plugin version to 0.7.9

[iemejia] Update dockerfile-maven-plugin version to 1.3.6

[iemejia] Add maven-enforcer-plugin version to pluginManagement

[iemejia] Fix warning on using directly parent.version on Nexmark

[iemejia] Remove warnings on repeated maven-jar-plugin invocation / deps

[iemejia] Remove warning on defining <prerequisites> for non-maven plugin project

[iemejia] Update parent POM to version 19

[iemejia] Remove repeated dependency in hadoop-input-format module

[mariagh] Support argparse-style choices for ValueProvider

[jbonofre] [BEAM-3551] Add -parameters flag to the compiler

[jbonofre] [BEAM-3551] Align gradle java compile task options with the

[coheigea] BEAM-3560 - Switch to use BigInteger/BigDecimal.ZERO/ONE/TEN

[lcwik] [BEAM-2500, BEAM-3249] Add amazon-web-services gradle build rules.

[jbonofre] [maven-release-plugin] prepare branch release-2.3.0

[jbonofre] [maven-release-plugin] prepare for next development iteration

[jbonofre] Update Python SDK version post release

[iemejia] Fix dependency order error, harness must be built before container

[robertwb] [BEAM-3537] Allow more general eager in-process pipeline execution

[lcwik] [BEAM-3550] Add --awsServiceEndpoint option and use with S3 filesystem.

[lcwik] Fixes for sdks/java/core for the eclipse compiler

[lcwik] Add some m2e lifecycle things to get the various maven plugins running

[lcwik] Set the proper package for the Snippets example so eclipse won't raise

[lcwik] Add some template args and direct casts to help the eclipse compiler

[lcwik] Incorporate reviews and rebase on latest master

[lcwik] Return no environment more often

[Pablo] Logging deviation from sampling expectation. This will allow to track

[kedin] [SQL] Refactor Variance

[robertwb] [BEAM-3490] Make runtime type checking code runner agnostic.

[kedin] [Nexmark][SQL] Implement sql query 3

[github] [BEAM-3557] Sets parent pointer of AppliedPTransform objects correctly

[pawel.pk.kaczmarczyk] [BEAM-2469] Handling Kinesis shards splits and merges

[Pablo] Adding a static getter for RuntimeValueProvider.

[tgroh] Add CoderTranslatorRegistrar

[tgroh] Add slf4j_simple to the top level Gradle build

[tgroh] Implement FnService in FnApiControlClientPoolService

[tgroh] Add a Timeout to GrpcDataService#send

[tgroh] Use a Data Service in SdkHarnessClient

[github] get the query from configuration not options

[XuMingmin] [BEAM-3525] Fix KafkaIO metric (#4524)

[chamikara] Updates PTransform overriding to create a new AppliedPTransform object

[dariusz.aniszewski] use build $WORKSPACE as pkb temp_dir and update pip and setuptools in

[iemejia] [BEAM-3578] SQL module build breaks because of missing dependency

[Pablo] Renaming the ZERO element of DistributionResult to be IDENTITY_ELEMENT.

[kenn] google-java-format

[kenn] Fix Distinct null pointer error with speculative triggers

[kenn] Move TestCountingSource to appropriate location

[robertwb] Direct runner fixes.

[lcwik] [BEAM-2926] Add support for side inputs to the runner harness.

[kenn] Sickbay ApexRunner gradle WordCountIT

[kenn] Sickbay flakey KinesisReaderTest

[Pablo] Addressing comments.

[rober] Fix beam.Combine to combine globally

[ehudm] Split out buffered read and write code from gcsio.

[github] Removing unnecessary code.

[lcwik] [BEAM-3249] Make sure that all java projects package tests. Also package

[lcwik] [BEAM-3249] Do not assume build directory is within build/, use the

[github] Fix undefined names: exc_info --> self.exc_info

[github] import logging for line 1163

[iemejia] [BEAM-3592] Fix spark-runner profile for Nexmark after move to Spark 2.x

[dkulp] [BEAM-3562] Update to Checkstyle 8.7

[klk] Encourage a good description in a good spot on a PR description.

[lcwik] Change info to debug statement

[cclauss] global INT64_MAX, INT64_MIN to placate linters

[tgroh] Add QueryablePipeline

[gene] Changing FileNaming to public to allow for usage in lambdas/inheritance

[robertwb] [BEAM-3207] Create a standard location to enumerate and document URNs.

[cclauss] xrange() was removed in Python 3 (en masse)

[robertwb] Reduce the flakiness of the state sampler progress metrics.

[robertwb] Revert URNs that are currently hard-coded in the Dataflow worker.

[herohde] Add optional function registration to Go SDK runtime

[kedin] [SQL] Inherit windowing strategy from the input in Aggregate operation

[jbonofre] [BEAM-3551] Define compiler -parameters flag in the default options

[tgroh] Add SdkHarnessClientControlService

[tgroh] Update Synchronization in FnApiControlClient

[coheigea] BEAM-3593 - Remove methods that just call super()

[lcwik] Move off of deprecated method in Guava.

[tgroh] Add a LocalArtifactStagingLocation

[tgroh] Add LocalArtifactStagingLocation#forExisting

[tgroh] Add an ArtifactRetrievalService interface

[tgroh] Implement a Local ArtifactRetrievalService

[chamikara] Adds a ReadAll transform to tfrecordio.

[rangadi] KafkaIO : move source and sink implemenations into own files.

[rangadi] minor

[kedin] [SQL] Add SqlTypeCoder, replace java.sql.Types

[Pablo] Moving User metrics to be in the PTransform proto for Fn API.

[mairbek] Update cloud spanner library to 0.29.0

[mairbek] Fix test

[mairbek] More google-cloud-platform whitelisting

[mairbek] pom updates to make maven happy

[mairbek] Update netty deps

[rober] fixup! Remove reflection from varint codecs

[ccy] [BEAM-3566] Replace apply_* hooks in DirectRunner with

[ccy] Address reviewer comments

[davidyan] Correct typo in SpannerIO.Write.withHost

[klk] google-java-format

[klk] Fix empty window assignments in Nexmark

[klk] Fix empty window assignment in FlattenEvaluatorFactoryTest

[klk] Switch DataflowRunner to its own private ValueInEmptyWindows

[klk] Remove deprecated valueInEmptyWindows

[jiangkai] Covariance Functions

[aljoscha.krettek] Remove erroneous cast in FlinkStreamingTransformTranslators

[aljoscha.krettek] [BEAM-3186] Correctly use deserialized timerService in Flink Runner

[lcwik] Adjust gradle build dirs and hints to help IntelliJ (#4583)

[coheigea] BEAM-3618 - Remove extraneous "return" statement

[robertwb] [BEAM-3183] Allow a callable as input to runner.run().

[sidhom] Fix gradle java sdk image build

[kenn] Add MoreFutures utility

[kenn] Switch runners/java-fn-execution from Guava futures to Java 8 futures

[kenn] Switch DataflowRunner from Guava futures to Java 8 futures

[kenn] Switch gcp-core from Guava futures to Java 8 futures

[kenn] Switch runners/core-construction-java from Guava futures to Java 8

[kenn] Switch AWS IO from Guava futures to Java 8 futures

[kenn] Switch BigTableIO from Guava futures to Java 8 futures

[kenn] Deprecate DoFnTester

[herohde] Changed core GBK to CoGBK

[herohde] Add CoGBK support to direct runner and Join example

[herohde] [BEAM-3316] Translate bundle descriptors directly to execution plans in

[herohde] Translate CoGBK into GBK for Dataflow and model pipeline runners

[mairbek] Fixed broken test

[herohde] CR: [BEAM-3302] Support CoGBK in the Go SDK

[rangadi] Remove older Kafka versions from build time support.

[ekirpichov] Adds PositionT and claim callback to RestrictionTracker

[ekirpichov] Changes OutputAndTimeBounded invoker to start checkpoint timer after

[ekirpichov] Compresses encoded GrowthState with Snappy - about 2x-3x more compact

[ekirpichov] InMemoryStateInternals.copy clones the values using the coder

[ekirpichov] Final fixups

[ekirpichov] Bump worker to 20180205

[klk] Sickbay flaky KinesisIO tests

[klk] Remove DoFnTester from core SDK tests

[cclauss] from six import integer_types (en masse)

[aljoscha.krettek] [BEAM-2806] Fix pipeline translation mode recognition in Flink Runner

[github] Consistantly show Python and pip versions in tests

[jbonofre] Revert "Reinstate proto round trip in Java DirectRunner"

[tgroh] Update Assign Window URN Constant Name

[coheigea] BEAM-3624 - Remove collapsible if statements

[kenn] Sickbay ApexRunner ParDoTranslatorTest.testAssertionFailure

[kenn] Switch FullWindowedValueCoder to bypass validation

[kedin] Refactor BeamRecordType and BeamRecord

[sidhom] Allow docker tag root to be specified as in Maven image build

[herohde] [BEAM-3457] Upgrade gogradle and fix thrift resolution issue

[herohde] [BEAM-3457] Add Go Gradle precommit

[ccy] [BEAM-3635] Infer type hints on PTransformOverrides

[robertwb] [BEAM-3625] Enable DoFn params in Map, Filter, etc.

[herohde] [BEAM-3579] Fix textio.Write

[kedin] Rename BeamRecord -> Row, BeamRecordType -> RowType

[mariagh] Add test for processing-time timer

[iemejia] Add missing gradle build config for sdks/java/extensions/sketching

[iemejia] Fix type on shadowTest when it should be testShadow

[ccy] Update snippets to fix pickling and clarify encoding issues

[wcn] Modify BufferedElementCountingOutputStream to use a buffer pool for its

[klk] Fix stable name errors in HBaseIOTest

[arnaudfournier021] Minor Javadoc corrections for SketchFrequencies

[arnaudfournier021] [BEAM-2728] Add Quantiles finding transforms to sketching extension

[arnaudfournier021] Change coder serialization + improve Javadoc comments + minor fixes

[aromanenko.dev] [BEAM-3291] Add Kinesis write transform

[coheigea] Remove unused private variables.

[ehudm] Add and migrate to HdfsCLI library for Python SDK.

[klk] Fix typo in gradle idea hints

[dkulp] [BEAM-3639] Update to gradle 4.5.1

[herohde] Also ignore alternative path for gogradle thrift location

[herohde] Remove gogradle manual dependency ordering

[herohde] Lock Go dependency versions

[herohde] Ignore gogradle.lock in rat check

[herohde] Ignore gogradle.lock in rat check for maven

[github] Revert "Update cloud spanner library to 0.29.0"

[herohde] CR: fix Go SDK textio.Write

[tgroh] Add Javadoc on how Teardown is best-effort

[kedin] [Schema Generation] Generate BeamRecordTypes based on pojos.

[alan] [BEAM-3524] Automate testing using python sdk container built at head

[coheigea] Replace boolean ternary operator + simplify some Boolean expressions

[apilloud] [BEAM-410] Sort PriorityQueue<QuantileBuffer> with explicit comparator

[robertwb] Disable verbose typecoder warnings.

[fjetumale] [BEAM-2817] BigQuery queries are allowed to run in either BATCH or

[jb] [BEAM-793] Add backoff support in JdbcIO Write

[kenn] Increase gradle logging to --info

[tgroh] Add a Primitive Impulse PTransform

[ccy] Add switchable DirectRunner which uses the fast FnApiRunner when

[rangadi] Move kafka-clients dependency to provided scope.

[daniel.o.programmer] [BEAM-419] Modifying FindBug comment.

[tgroh] Add a single-stage fusion implementation

[robertwb] [BEAM-3074] Serialize DoFns by portable id in Dataflow runner.

[lcwik] [BEAM-3629] Send the windowing strategy and whether its a merging window

[jb] [BEAM-3668] Quick workaround fix for netty conflict waiting better fix

[wcn] Fixing filename.

[iemejia] Fix warning on jenkins on non-existent profile 'validates-runner-tests'

[iemejia] Remove unneeded overwrites of maven-compiler-plugin

[iemejia] Change tests execution order from filesystem (default) to random

[iemejia] Remove repeated dependencies on runners/java-fn-execution module

[iemejia] Add missing modules to javadoc generation: TikaIO, RedisIO, Jackson, Xml

[iemejia] [BEAM-2530] Make final fixes to ensure code and tests compile with Java

[aromanenko.dev] [BEAM-3637] HBaseIOTest - random table names for every test

[jbonofre] [BEAM-3692] Remove maven deploy plugin configuration with skip in the

[ehudm] Integration test for Python HDFS implementation.

[herohde] Remove bad gogradle.lock files

[pawel.pk.kaczmarczyk] [BEAM-3605] Use verification with timeout instead of Thread.sleep

[XuMingmin] [BEAM-3176] support drop table (#4184)

[kirpichov] Two fixes to common URN handling

[arnaudfournier921] Improve Javadoc ° minor fixes

[dkulp] [BEAM-3581] Make sure calcite gets an appropriate charset PRIOR to any

[ehudm] Print correct line numbers for warnings.

[Pablo] Adding Gauge metric to Python SDK.

[Pablo] Fix lint issue

[wcn] Improve rendering of DOT diagrams.

[lcwik] [BEAM-3626] Add a handler capable of executing a window mapping fn on a

[Pablo] Addressing comments

[herohde] Update Go SDK coder constants

[lcwik] [BEAM-3339] Fix failing post-release test by running groovy from gradle,

[batbat] Fixed a bug that timer ID was not used for comparing timer data. Added

[ccy] Use TestClock when TestStream is present in pipeline

[klk] Function interface for Fn API instructions

[cclauss] Exception.message was removed in Python 3

[iemejia] [BEAM-3697] Add Maven profile to run error-prone static analysis

[iemejia] [BEAM-3697] Fix MoreFutures errorprone

[alan] [BEAM-3695] Fix failing validates container test

[aromanenko.dev] [BEAM-3228] Fix flaky Kinesis tests

[tgroh] Add a multi-stage fuser

[XuMingmin] [BEAM-3345][SQL] Reject unsupported inputs into JOIN (#4642)

[kedin] Update 'PCollectionTuple.apply()' generic definition

[kedin] [SQL] Refactor BeamSql

[tgroh] fixup! Add a multi-stage fuser

[Pablo] Addressing comments.

[luke_zhu] Support Python 3 in the metrics, internal, typehints, and utils modules.

[apilloud] [Nexmark][SQL] Use Timestamp type for timestamps

[dcavazos] Added snippets for BigQueryIO, serializable functions, Dynamic

[apilloud] [Nexmark][SQL] Implement Query5

[dcavazos] Use beam.io.WriteToBigQuery()

[rangadi] Fix unbounded reader leak in direct-runner. Also close the reader at end

[rangadi] `extractOutput()` ended up resetting underlying aggregation. This is due

[rangadi] review comments.

[kirpichov] Updates BigQuery dependency version

[XuMingmin] Bump calcite version to 1.15.0 (#4692)

[ccy] Use the switching DirectRunner implementation

[cclauss] long was renamed to int in Python 3 (en masse)

[dkulp] [BEAM-3640] Part1: Update Checkstyle to enforce blank lines for imports

[tgroh] fixup! Add a multi-stage fuser

[kenn] Ignore IntelliJ Gradle build outputs

[Pablo] Renaming MetricAggregator.zero to MetricAggregator.identity_element

[apilloud] [Nexmark][SQL] Implement Query7

[kenn] Add hints for IntelliJ owned output dirs

[iemejia] Update byte-buddy to version 1.7.10 (adds support for Java 9)

[iemejia] Update google-auto-value to version 1.5.3

[iemejia] Pin maven-gpg-plugin to version 1.6

[iemejia] Pin missing version for license-maven-plugin to version 1.14

[iemejia] [BEAM-2530] Fix dependencies for XmlIO on Java 9

[iemejia] [BEAM-2530] Add a java 9 profile to parent pom

[iemejia] Add extra-enforcer-rules maven plugin to version 1.0-beta-7

[iemejia] Add ignore rule for multi-release jars compatible with Java 8

[sidhom] Use maven-publish plugin to publish java artifacts

[1028332163] replace mockito-all

[1028332163] replace mockito

[1028332163] replace mockito-all and harcrest-all

[1028332163] replace mockito-all and harcrest-all

[1028332163] replace mockito-all and harcrest-all

[1028332163] replace mockito-all and harcrest-all

[1028332163] replace mockito-all and harcrest-all

[1028332163] replace mockito-all and harcrest-all

[1028332163] replace mockito-all and harcrest-all

[1028332163] replace mockito-all and harcrest-all

[1028332163] ignoredUnusedDeclaredDependencies

[lcwik] [BEAM-3690] swapping to use mockito-core, hamcrest-core and

[github] Updates javadocs of Setup and Teardown

[1028332163] ban hamcrest-all and mockito-all

[1028332163] ban mockito-all and hamcrest-all

[willy] [BEAM-3662] Port MongoDbIOTest off DoFnTester

[lukasz.gajowy] [BEAM-3456] Re-enable JDBC performance test

[lukasz.gajowy] fixup! [BEAM-3456] Re-enable JDBC performance test

[lukasz.gajowy] fixup! fixup! [BEAM-3456] Re-enable JDBC performance test

[kirpichov] Updates BigQuery documentation

[luke_zhu] Revert invalid use of io.StringIO in utils/profiler.py

[Pablo] Plumbing Gauge metrics through the Fn API.

[dkulp] [BEAM-3640] part 2 - add

[kenn] Spotless gradle: remove extraneous globbing of all java everywhere

[rmannibucau] ensure pipeline options setup is using contextual classloader and not

[kenn] Explicitly exclude some troublesome optional deps from

[Pablo] Fixing nanosecond translation issue in Gauge Fn API translation.

[lcwik] Break fusion for a ParDo which has State or Timers

[boyuanz] Add distribution counter implementation

[kirpichov] A relative directory should be applied (if specified) even when using a

[kenn] Explicitly exclude further optional deps from elasticsearch-hadoop

[lcwik] [BEAM-2573] Don't force importing filesystems, if they fail then give up

[tgroh] Use Conccurrent Constructs in InMemoryArtifactStagerService

[kirpichov] Adds more logging of BigQuery jobs and makes load job naming more

[tgroh] Add Environment Manager Interfaces

[aromanenko.dev] [BEAM-3538] Remove (or merge) Java 8 specific tests module into the main

[iemejia] [BEAM-3632] Add missing partitioning parameter in WriteTables

[iemejia] [BEAM-3632] Add TableDestination.withTableReference and fix WriteTables

[rmannibucau] [BEAM-3728][BEAM-3729] fixing the classloader lookup for pipeline

[pawel.pk.kaczmarczyk] [BEAM-3317] Use fixed system time for testing

[lukasz.gajowy] [BEAM-3732] Fix broken maven profiles

[rangadi] Use TreeSet in place of PriorityQueue.

[coheigea] Make sure there is a space between closing round bracket and opening

[ehudm] Don't cache pubsub subscription prematurely.

[robertwb] Add MultiMap side inputs to Python SDK.

[altay] Fixing minor bugs:

[ankurgoenka] Making default thread count 12

[rangadi] update checksWithMultipleMerges() to check for multiple merges by

[holden] First pass at fixing all of E999 (invalid parsing) errors in Py3 found

[aljoscha.krettek] Make parameter of DoFnRunners.lateDataDroppingRunner() more specific

[aljoscha.krettek] Allow overriding DoFnRunners in subclasses of Flink DoFnOperator

[aljoscha.krettek] Invoke finishBundle() before teardown() in DoFnOperator

[aljoscha.krettek] [BEAM-2140] Ignore event-time timers in SplittableDoFnOperator

[aljoscha.krettek] [BEAM-2140] Block DoFnOperator.close() if we have pending timers

[aljoscha.krettek] [BEAM-2140] Don't use StatefulDoFnRunner when running SDF in FlinkRunner

[aljoscha.krettek] Make ProcessFnRunner constructor public

[aljoscha.krettek] [BEAM-2140] Use ProcessFnRunner in DoFnRunner for executing SDF

[aljoscha.krettek] [BEAM-2140] Enable SDF tests for Flink Streaming Runner

[aljoscha.krettek] [BEAM-2140] Enable SDF tests in gradle for Flink Streaming Runner

[aljoscha.krettek] [BEAM-3727] Never shutdown sources in Flink Streaming execution mode

[mairbek] Changes multipe structuredValue implementations. Most notably, updates

[tgroh] Make Impulse#create() visible

[XuMingmin] [BEAM-3634] Refactor BeamRelNode to return a PTransform (#4705)

[XuMingmin] [BEAM-591] KafkaIO : Improve watermarks and support server side

[holden] Fix some raise_from to reraise.

[holden] vcfio somehow has some sort issues. It's not overly important and

[wtanaka] [BEAM-3034] Utilize some java 5 features

[xumingmingv] add setup/teardown for BeamSqlSeekableTable.

[xumingmingv] rename method as suggested and declare as default methods.

[yifanzou] [BEAM-3339] add python RC validation automation

[github] Fixes typo in BigQueryIO javadoc

[markliu] [BEAM-3750] Remove TestPipeline.covnertToArgs from integration tests

[XuMingmin] [BEAM-591]: Update KafkaIO JavaDoc to reflect new timestamp API. (#4749)

[tgroh] Add To/From Proto Round Trip for ExecutableStage

[tgroh] Make GreedyStageFuser a Factory Class

[kirpichov] Makes it possible to use Wait with default windowing, eg. in batch

[xumingmingv] @Parameter annotation does not work for UDFs in Beam SQL

[xumingmingv] fix type error

[grzegorz.kolakowski] [BEAM-3043] Set user-specified PTransform names on Flink operators

[coheigea] Avoid unnecessary autoboxing by replacing Integer/Long.valueOf with

[tgroh] Update QueryablePipeline Factory Method Name

[tgroh] Scan Core Construction NeedsRunner Tests

[klk] Add MetricsTranslation

[rmannibucau] extracting the scheduled executor service in a factory variable in SDF

[echauchot] [BEAM-3681] Make S3FileSystem copy atomic for smaller than 5GB objects

[echauchot] [BEAM-3681] Add a comment for the extra check of objectSize in

[grzegorz.kolakowski] [BEAM-3753] Fix failing integration tests

[grzegorz.kolakowski] [BEAM-3753] Rename *ITCase.java tests files to *Test.java

[lcwik] [BEAM-3762] Update Dataflow worker container image to support unlimited

[github] Bump container image tag to fix incompatibility between the SDK and the

[github] Update dependency.py

[sidhom] Run NeedsRunner tests from direct runner gradle build

[ccy] Fix issue from incomplete removal of has_cache

[yifanzou] [BEAM-3735] copy mobile-gaming sources in to archetypes

[mairbek] Update SpannerIO to use Batch API.

[sidhom] Address review comments

[sidhom] Remove old sourceSets.test.output references

[rangadi] Upate Readme and JavaDoc for KafkaIO.

[robertwb] Avoid warning in our default runner.

[github] [BEAM-3719] Adds support for reading side-inputs from SDFs

[github] print() is a function in Python 3

[ehudm] Add Python lint check for calls to unittest.main.

[tgroh] Add JavaReadViaImpulse to core-construction

[robertwb] [maven-release-plugin] prepare branch release-2.4.0

[robertwb] [maven-release-plugin] prepare for next development iteration

[robertwb] Bump Python dev version.

[tgroh] Revert "extracting the scheduled executor service in a factory variable

[rangadi] Mention support for older versions.

[amyrvold] Beam runner inventory, run as a cron job, not on each CL

[github] Fixing formatting bug in filebasedsink.py.

[github] Fix lint issue.

[ehudm] Add support for Pub/Sub messages with attributes.

[amyrvold] [BEAM-3775] Increase timeout in

[ehudm] Address review comment from PR #4744.

[mariagh] Add TestClock to test

[wcn] Update generated code to match import from 5e6db92

[XuMingmin] [BEAM-3754]: Fix readBytes() initialization (#4792)

[daniel.o.programmer] [BEAM-3126] Fixing incorrect function call in bundle processor.

[cclauss] Change unicode --> six.text_type for Python 3

[alan.myrvold] [BEAM-3621] Add Spring repo to build_rules to allow downloading pentaho

[samuel.waggoner] [BEAM-3777] allow UDAFs to be indirect subclasses of CombineFn

[ehudm] Improve FileBasedSink rename safety.

[ehudm] Add missing import statements.

[lcwik] [BEAM-3690] ban mockito-all and hamcrest-all

[kenn] Bump sdks/go/container.pom.xml to 2.5.0-SNAPSHOT

[amyrvold] [BEAM-3791] Update version number in build_rules.gradle

[rmannibucau] Make StateInternals short state method defaulting to the implementation

[apilloud] [Nexmark][SQL] Convert SQL Rows to Java models

[apilloud] [Nexmark][SQL] Output java pojos

[apilloud] [Nexmark][SQL] Add model tests for sql

[tgroh] Fallback to byte equality in MutaitonDetection

[herohde] [BEAM-3457] Drop Go Maven PreCommit

[markliu] Fix integration tests that use NumberedSharedFiles

[kedin] [SQL] Add support for ARRAY expression

[kedin] [SQL] Support ARRAY in projections

[kedin] [SQL] Implement array elements access expression

[kedin] [SQL] Add support for ELEMENT(collection) function

[kedin] [SQL] Add support for CARDINALITY(collection)

[tgroh] Add an InProcess SdkHarness Rule

[tgroh] Close Outstanding Control Clients on Service close

[tgroh] Don't try to close a cancelled Multiplexer

[herohde] [BEAM-3804] Build Go SDK container with gradle

[herohde] [BEAM-3793] Validate provision response and add beamctl support

[cclauss] [BEAM-1251] Fix basestring, file(), reduce(), and xrange() for Python 3

[apilloud] [Nexmark] Use queue for Query 6 model

[apilloud] [Nexmark] Ensure enough data to produce output

[altay] Fix topic URIs.

[apilloud] [BEAM-3802] Remove broken cachedMetricResults

[tgroh] fixup! Close Outstanding Control Clients on Service close

[tgroh] fixup! Don't try to close a cancelled Multiplexer

[tgroh] fixup! Add an InProcess SdkHarness Rule

[cclauss] [BEAM-1251] Fix basestring for Python 3 - again

[cclauss] [BEAM-1251] Change unicode --> six.text_type for Python 3 - again

[herohde] Fix package name in Go container gradle build

[apilloud] [Nexmark] Sickbay query 6

[mariand] Improve implementation of nexmark.StringsGenerator.nextExactString()

[ehudm] Update Dataflow Beam container version.

[chamikara] [BEAM-3734] Performance tests for XmlIO  (#4747)

[apilloud] Dataflow runner must supply attempted metrics

[apilloud] Throw UnsupportedOperationException instead of returning null

[robertwb] Remove obsolete MapTaskRunner.

[ccy] Revert #4666 "Use beam.io.WriteToBigQuery()"

[github] [BEAM-3806] Fix direct-runner hang (#4829)

[kedin] [SQL] Support nested Rows

[klk] Add GAUGE_DATA case to metricUpdatesFromProto

[klk] Add Gauge metric tests to ensure value persists

[klk] Add test for a trigger with windowed SQL query

[klk] Increase whitelist of false detections in SdkCoreApiSurfaceTest

[klk] Eliminate beam-model-fn-execution test-jar deps

[klk] Eliminate beam-sdks-java-fn-execution test-jar deps

[klk] Eliminate incorrect sdks-java-core test-jar deps

[klk] Notate uses of beam-runners-core-java test-jar

[chamikara] [BEAM-3217] Jenkins job for HadoopInputFormatIOIT (#4758)

[ekirpichov] [BEAM-3741] Proto changes for splitting over Fn API

[ekirpichov] Address comments

[aljoscha.krettek] Annotate ParDoTest.duplicateTimerSetting with UsesTestStream

[lcwik] [BEAM-3842] Allow static methods to be defined inside PipelineOptions.

[lcwik] [BEAM-3843] Add a convenience method in ExperimentalOptions to check to

[lcwik] [BEAM-3843] Migrate to using ExperimentalOptions.hasExperiment

[tgroh] fixup! Don't try to close a cancelled Multiplexer

[ehudm] Use Python 3 compatible string.

[XuMingmin] [SQL] Add support for arrays of rows (#4857)

[XuMingmin] [BEAM-2281][Sql] Use SqlFunctions.toBigDecimal not toString (#4865)

------------------------------------------
[...truncated 193.11 KB...]
[INFO] Running org.apache.beam.sdk.io.tfrecord.TFRecordIOIT
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 468.073 s <<< FAILURE! - in org.apache.beam.sdk.io.tfrecord.TFRecordIOIT
[ERROR] writeThenReadAll(org.apache.beam.sdk.io.tfrecord.TFRecordIOIT)  Time elapsed: 468.073 s  <<< ERROR!
java.lang.RuntimeException: 
java.io.IOException: Failed to advance reader of source: hdfs://146.148.62.9:9000/TEXTIO_IT__1521115464780-00000-of-00003.tfrecord range [0, 19222204)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:605)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation$SynchronizedReaderIterator.advance(ReadOperation.java:379)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.runReadLoop(ReadOperation.java:185)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.start(ReadOperation.java:150)
	at com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:74)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:381)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:353)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:284)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Not a valid TFRecord. Fewer than 12 bytes.
	at org.apache.beam.sdk.repackaged.com.google.common.base.Preconditions.checkState(Preconditions.java:444)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordCodec.read(TFRecordIO.java:631)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordSource$TFRecordReader.readNextRecord(TFRecordIO.java:526)
	at org.apache.beam.sdk.io.CompressedSource$CompressedReader.readNextRecord(CompressedSource.java:426)
	at org.apache.beam.sdk.io.FileBasedSource$FileBasedReader.advanceImpl(FileBasedSource.java:473)
	at org.apache.beam.sdk.io.OffsetBasedSource$OffsetBasedReader.advance(OffsetBasedSource.java:267)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:602)
	... 14 more
(402142cc52d644e0): java.io.IOException: Failed to advance reader of source: hdfs://146.148.62.9:9000/TEXTIO_IT__1521115464780-00000-of-00003.tfrecord range [0, 19222204)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:605)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation$SynchronizedReaderIterator.advance(ReadOperation.java:379)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.runReadLoop(ReadOperation.java:185)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.start(ReadOperation.java:150)
	at com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:74)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:381)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:353)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:284)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Not a valid TFRecord. Fewer than 12 bytes.
	at org.apache.beam.sdk.repackaged.com.google.common.base.Preconditions.checkState(Preconditions.java:444)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordCodec.read(TFRecordIO.java:631)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordSource$TFRecordReader.readNextRecord(TFRecordIO.java:526)
	at org.apache.beam.sdk.io.CompressedSource$CompressedReader.readNextRecord(CompressedSource.java:426)
	at org.apache.beam.sdk.io.FileBasedSource$FileBasedReader.advanceImpl(FileBasedSource.java:473)
	at org.apache.beam.sdk.io.OffsetBasedSource$OffsetBasedReader.advance(OffsetBasedSource.java:267)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:602)
	... 14 more
(402142cc52d64c95): java.io.IOException: Failed to advance reader of source: hdfs://146.148.62.9:9000/TEXTIO_IT__1521115464780-00000-of-00003.tfrecord range [0, 19222204)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:605)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation$SynchronizedReaderIterator.advance(ReadOperation.java:379)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.runReadLoop(ReadOperation.java:185)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.start(ReadOperation.java:150)
	at com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:74)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:381)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:353)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:284)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Not a valid TFRecord. Fewer than 12 bytes.
	at org.apache.beam.sdk.repackaged.com.google.common.base.Preconditions.checkState(Preconditions.java:444)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordCodec.read(TFRecordIO.java:631)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordSource$TFRecordReader.readNextRecord(TFRecordIO.java:526)
	at org.apache.beam.sdk.io.CompressedSource$CompressedReader.readNextRecord(CompressedSource.java:426)
	at org.apache.beam.sdk.io.FileBasedSource$FileBasedReader.advanceImpl(FileBasedSource.java:473)
	at org.apache.beam.sdk.io.OffsetBasedSource$OffsetBasedReader.advance(OffsetBasedSource.java:267)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:602)
	... 14 more
(402142cc52d6444a): java.io.IOException: Failed to advance reader of source: hdfs://146.148.62.9:9000/TEXTIO_IT__1521115464780-00000-of-00003.tfrecord range [0, 19222204)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:605)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation$SynchronizedReaderIterator.advance(ReadOperation.java:379)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.runReadLoop(ReadOperation.java:185)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.start(ReadOperation.java:150)
	at com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:74)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:381)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:353)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:284)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Not a valid TFRecord. Fewer than 12 bytes.
	at org.apache.beam.sdk.repackaged.com.google.common.base.Preconditions.checkState(Preconditions.java:444)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordCodec.read(TFRecordIO.java:631)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordSource$TFRecordReader.readNextRecord(TFRecordIO.java:526)
	at org.apache.beam.sdk.io.CompressedSource$CompressedReader.readNextRecord(CompressedSource.java:426)
	at org.apache.beam.sdk.io.FileBasedSource$FileBasedReader.advanceImpl(FileBasedSource.java:473)
	at org.apache.beam.sdk.io.OffsetBasedSource$OffsetBasedReader.advance(OffsetBasedSource.java:267)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:602)
	... 14 more
(ca56e8b0d6266738): Workflow failed. Causes: (d03bc8b58b44abcf): S02:TFRecordIO.Read/Read+Transform bytes to strings/Map+Calculate hashcode/WithKeys/AddKeys/Map+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write failed., (40549a95d3ed11a8): A work item was attempted 4 times without success. Each time the worker eventually lost contact with the service. The work item was attempted on: 
  tfrecordioit0writethenrea-03150508-fd86-harness-v056,
  tfrecordioit0writethenrea-03150508-fd86-harness-v056,
  tfrecordioit0writethenrea-03150508-fd86-harness-v056,
  tfrecordioit0writethenrea-03150508-fd86-harness-v056
	at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:133)
	at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:89)
	at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:54)
	at org.apache.beam.sdk.Pipeline.run(Pipeline.java:311)
	at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:346)
	at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:328)
	at org.apache.beam.sdk.io.tfrecord.TFRecordIOIT.writeThenReadAll(TFRecordIOIT.java:127)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
	at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:317)
	at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:317)
	at org.junit.rules.RunRules.evaluate(RunRules.java:20)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
	at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)

[INFO] 
[INFO] Results:
[INFO] 
[ERROR] Errors: 
[ERROR]   TFRecordIOIT.writeThenReadAll:127  Runtime java.io.IOException: Failed to adv...
[INFO] 
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0
[INFO] 
[INFO] 
[INFO] --- maven-dependency-plugin:3.0.2:analyze-only (default) @ beam-sdks-java-io-file-based-io-tests ---
[WARNING] Used undeclared dependencies found:
[WARNING]    javax.xml.bind:jaxb-api:jar:2.2.2:runtime
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 08:27 min
[INFO] Finished at: 2018-03-15T12:12:14Z
[INFO] Final Memory: 104M/1393M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-dependency-plugin:3.0.2:analyze-only (default) on project beam-sdks-java-io-file-based-io-tests: Dependency problems found -> [Help 1]
org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.apache.maven.plugins:maven-dependency-plugin:3.0.2:analyze-only (default) on project beam-sdks-java-io-file-based-io-tests: Dependency problems found
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:213)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146)
    at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)
    at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81)
    at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:51)
    at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128)
    at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:309)
    at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:194)
    at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:107)
    at org.apache.maven.cli.MavenCli.execute (MavenCli.java:955)
    at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:290)
    at org.apache.maven.cli.MavenCli.main (MavenCli.java:194)
    at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke (Method.java:498)
    at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289)
    at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229)
    at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415)
    at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356)
Caused by: org.apache.maven.plugin.MojoExecutionException: Dependency problems found
    at org.apache.maven.plugins.dependency.analyze.AbstractAnalyzeMojo.execute (AbstractAnalyzeMojo.java:254)
    at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:134)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:208)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146)
    at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)
    at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81)
    at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:51)
    at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128)
    at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:309)
    at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:194)
    at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:107)
    at org.apache.maven.cli.MavenCli.execute (MavenCli.java:955)
    at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:290)
    at org.apache.maven.cli.MavenCli.main (MavenCli.java:194)
    at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke (Method.java:498)
    at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289)
    at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229)
    at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415)
    at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356)
[ERROR] 
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException

STDERR: 
2018-03-15 12:12:14,593 240db462 MainThread beam_integration_benchmark(1/1) ERROR    Error during benchmark beam_integration_benchmark
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 623, in RunBenchmark
    DoRunPhase(spec, collector, detailed_timer)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 526, in DoRunPhase
    samples = spec.BenchmarkRun(spec)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",> line 159, in Run
    job_type=job_type)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",> line 90, in SubmitJob
    assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2018-03-15 12:12:14,595 240db462 MainThread beam_integration_benchmark(1/1) INFO     Cleaning up benchmark beam_integration_benchmark
2018-03-15 12:12:14,596 240db462 MainThread beam_integration_benchmark(1/1) INFO     Running: kubectl --kubeconfig=<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/config-filebasedioithdfs-1521114081611> delete -f <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/src/.test-infra/kubernetes/hadoop/SmallITCluster/hdfs-single-datanode-cluster.yml>
2018-03-15 12:12:15,007 240db462 MainThread beam_integration_benchmark(1/1) INFO     Running: kubectl --kubeconfig=<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/config-filebasedioithdfs-1521114081611> delete -f <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/src/.test-infra/kubernetes/hadoop/SmallITCluster/hdfs-single-datanode-cluster-for-local-dev.yml>
2018-03-15 12:12:15,153 240db462 MainThread beam_integration_benchmark(1/1) ERROR    Exception running benchmark
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 733, in RunBenchmarkTask
    RunBenchmark(spec, collector)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 623, in RunBenchmark
    DoRunPhase(spec, collector, detailed_timer)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 526, in DoRunPhase
    samples = spec.BenchmarkRun(spec)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",> line 159, in Run
    job_type=job_type)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",> line 90, in SubmitJob
    assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2018-03-15 12:12:15,154 240db462 MainThread beam_integration_benchmark(1/1) ERROR    Benchmark 1/1 beam_integration_benchmark (UID: beam_integration_benchmark0) failed. Execution will continue.
2018-03-15 12:12:15,155 240db462 MainThread beam_integration_benchmark(1/1) INFO     Benchmark run statuses:
---------------------------------------------------------------------------------
Name                        UID                          Status  Failed Substatus
---------------------------------------------------------------------------------
beam_integration_benchmark  beam_integration_benchmark0  FAILED                  
---------------------------------------------------------------------------------
Success rate: 0.00% (0/1)
2018-03-15 12:12:15,155 240db462 MainThread beam_integration_benchmark(1/1) INFO     Complete logs can be found at: <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/runs/240db462/pkb.log>
2018-03-15 12:12:15,155 240db462 MainThread beam_integration_benchmark(1/1) INFO     Completion statuses can be found at: <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/runs/240db462/completion_statuses.json>
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: beam_PerformanceTests_TFRecordIOIT_HDFS #4

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/4/display/redirect?page=changes>

Changes:

[alex] Correct BigQuery.write JavaDoc example

[rober] Correct EventTime encoding. time.(Time).Unix() returns seconds not

[tgroh] Enable Proto Round Trips in the DirectRunner Tests

[lcwik] [BEAM-3249, BEAM-3457] Make Apache Rat a dependency of the Java and

[sidhom] Add Java bounded read overrides

[herohde] [BEAM-3817] Switch BQ write to not use side input

[lcwik] [BEAM-3457] Exclude generated and data files part of the Go SDK.

[sidhom] Add ExecutableStagePayload to make aid runner stage reconstruction

[sidhom] Fix typo

[coheigea] Replacing size() == 0 with isEmpty()

[markliu] [BEAM-3841] Fix TestDataflowRunner.run to run_pipeline

[coheigea] Put the String literal first when comparing it to an Object

[ankurgoenka] Send error response for process bundle progress instead of throwing and

[tgroh] Use InstructionRequestHandler in RemoteEnvironment

[ekirpichov] BigQueryIO.read().fromQuery() improvements

[szewinho] [BEAM-3060] Fixing mvn dependency issue when runnning filebasedIOIT

[lukasz.gajowy] [BEAM-3798] Remove error check on dataflow when getting batch job state

[tgroh] Add ExecutableStage -> ProcessBundleDescriptor

[tgroh] Clean up GrpcFnServer Constructor Signature

[tgroh] Add a RemoteExecutionTest

[tgroh] Clean up a couple of SdkHarnessClient Signatures

[tgroh] Separate RemoteInputDestinations from SdkHarnessClient

[XuMingmin] [SQL] Add support for DOT expression (#4863)

[herohde] Add TODO to revert Go IO to use side input

[jb] [BEAM-3500] "Attach" JDBC connection to the bundle (improve the pooling)

[jb] [BEAM-3500] Test if the user provides both withDataSourceConfiguration()

[jb] [BEAM-3500] Wrap the datasource as a poolable datasource and expose

[jb] [BEAM-3500] Add commons-pool2 dependency

[jb] [BEAM-3500] Only expose max number of connections in the pool to the

[jb] [BEAM-3500] Cleanup pool configuration parameters

[jb] [BEAM-3500] Remove dataSourceFactory

[jb] [BEAM-3500] Remove unecessary check on dataSourceConfiguration

[szewinho] Jax-api added to dependencyManagement.

[szewinho] Jaxb-api added to beam root pom.xml

[szewinho] Removed empty line

[Pablo] Fixing counter string name.

[tgroh] Add Network Utility Methods

[lcwik] Fix missing license header in test.

[aaltay] Enable input subscription in Python streaming wordcount (#4864)

[tgroh] fixup! Add Network Utility Methods

[tgroh] Remove Future#set(null) in fn-execution

[herohde] [BEAM-3869] Fix Go Dataflow break due to zero timestamps

[axelmagn] Add a generic interface for the state service.

[herohde] Remove WindowedValue on PCollections for Go SDK

[cclauss] [BEAM-3872] Change Exception.message --> Exception.args[0]

[alan.myrvold] [BEAM-3249] Publish java artifacts using gradle

[herohde] CR: Fix comments to remove old windowing

[lcwik] Fix minor checkstyle/findbugs errors

[mariagh] Add support for streaming side inputs in the DirectRunner

[andreas.ehrencrona] [BEAM-2264] Credentials were not being reused between GCS calls

[ccy] Add support for PaneInfo in WindowedValues

[axelmagn] Write unit tests for GrpcStateService.

[XuMingmin] Bump calcite and avatica versions (#4887)

[axelmagn] Fix StateRequestHandler interface to be idiomatic

[herohde] Add Go support for universal runners, incl Flink

[herohde] CR: Fixed comments for job service helper functions

[iemejia] Add missing ASF license to ExecutableStageTranslation file

[szewinho] Updated gradle build to use jaxb-api. Jaxb-api version set to 2.2.12

[mariand] Switched AvroIO default codec to snappyCodec().

[yifanzou] [BEAM-3840] Get python mobile-gaming automating on core runners

[coheigea] Remove "i == numSplits" condition, which can never be reached as i is

[sidhom] [BEAM-3565] Clean up ExecutableStage

[wcn] Fix incorrect read of atomic counter.

[tgroh] Add a suborder parameter to Networks#topologicalOrder

[herohde] [BEAM-3893] Add fallback to unauthenticated access for GCS IO

[robertwb] [BEAM-3865] Fix watermark hold handling bug.

[robertwb] [BEAM-2927] Python support for dataflow portable side inputs over Fn API

[Pablo] Fixing check for sideinput_io_metrics experiment flag.

[herohde] CR: fix typo

[herohde] [BEAM-3897] Add Go wordcount example with multi output DoFns

[ccy] Replace side inputs when applying PTransformOverrides

[tgroh] Add QueryablePipeline#forTransforms

[herohde] Allow manual override of Go worker binary

[aaltay] [BEAM-3861] Improve test infra in Python SDK for streaming end-to-end

[iemejia] Remove testing package-info from main package for GCP IO

[iemejia] Update maven failsafe/surefire plugin to version 2.21.0

[iemejia] [BEAM-3873] Update commons-compress to version 1.16.1 (fix

[iemejia] Remove maven warnings

[tgroh] Add Side Inputs to ExecutableStage

[herohde] [BEAM-3866] Remove windowed value requirement for External

[lcwik] [BEAM-3903] Build the Go SDK examples using gradle

[markliu] Clean up terminal state check in TestDataflowRunner

[robertwb] Add mapper microbenchmark.

[szewinho] Scope removed from root pom.xml and jaxb-api dependency set to

[Pablo] Updating dataflow API protocol buffers.

[tgroh] Make RemoteEnvironment public

[boyuanz] Add cython annotation to make DistributionAccumulator faster

[tgroh] Move getId to the top-level pipeline node

[tgroh] Add FusedPipeline#toPipeline

[robertwb] Remove from docs generation.

[wcn] BEAM-3894: Add Protocol Buffer support

[altay] Convert big query writes to beam.io.WriteToBigQuery

[aaltay] Improve test util for PubSub exist check (#4922)

[XuMingmin] [BEAM-3851] Option to preserve element timestamp while publishing to

[ankurgoenka] Send worker_id in all channels to runner Adding

[wcn] Updating generated files.

[ehudm] Reduce precommit test timeouts for Java and Go.

[altay] More graceful fallback when grpc is not present.

[markliu] [BEAM-3861] Complete streaming wordcount test in Python SDK

[aaltay] [BEAM-3738] Enable py3 lint and cleanup tox.ini. (#4877)

------------------------------------------
Started by timer
[EnvInject] - Loading node environment variables.
Building remotely on beam2 (beam) in workspace <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/>
 > git rev-parse --is-inside-work-tree # timeout=10
Fetching changes from the remote Git repository
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 8b72e5af513e5182c403f3b67ff6d16395c281db (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 8b72e5af513e5182c403f3b67ff6d16395c281db
Commit message: "Merge pull request #4940: Update generated files"
 > git rev-list --no-walk 5b6ca47fec0b5b720ad5afb9274dc0d418545b43 # timeout=10
Cleaning workspace
 > git rev-parse --verify HEAD # timeout=10
Resetting working tree
 > git reset --hard # timeout=10
 > git clean -fdx # timeout=10
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[beam_PerformanceTests_TFRecordIOIT_HDFS] $ /bin/bash -xe /tmp/jenkins8246568391384583897.sh
+ gcloud container clusters get-credentials io-datastores --zone=us-central1-a --verbosity=debug
DEBUG: Running gcloud.container.clusters.get-credentials with Namespace(__calliope_internal_deepest_parser=ArgumentParser(prog='gcloud.container.clusters.get-credentials', usage=None, description='See https://cloud.google.com/container-engine/docs/kubectl for\nkubectl documentation.', version=None, formatter_class=<class 'argparse.HelpFormatter'>, conflict_handler='error', add_help=False), account=None, api_version=None, authority_selector=None, authorization_token_file=None, cmd_func=<bound method Command.Run of <googlecloudsdk.calliope.backend.Command object at 0x7f6fef241510>>, command_path=['gcloud', 'container', 'clusters', 'get-credentials'], configuration=None, credential_file_override=None, document=None, format=None, h=None, help=None, http_timeout=None, log_http=None, name='io-datastores', project=None, quiet=None, trace_email=None, trace_log=None, trace_token=None, user_output_enabled=None, verbosity='debug', version=None, zone='us-central1-a').
WARNING: Accessing a Container Engine cluster requires the kubernetes commandline
client [kubectl]. To install, run
  $ gcloud components install kubectl

Fetching cluster endpoint and auth data.
DEBUG: Saved kubeconfig to /home/jenkins/.kube/config
kubeconfig entry generated for io-datastores.
INFO: Display format "default".
[beam_PerformanceTests_TFRecordIOIT_HDFS] $ /bin/bash -xe /tmp/jenkins3974261244662825482.sh
+ cp /home/jenkins/.kube/config <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/config-filebasedioithdfs-1521890089911>
[beam_PerformanceTests_TFRecordIOIT_HDFS] $ /bin/bash -xe /tmp/jenkins5848965985188276448.sh
+ kubectl --kubeconfig=<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/config-filebasedioithdfs-1521890089911> create namespace filebasedioithdfs-1521890089911
Error from server (AlreadyExists): namespaces "filebasedioithdfs-1521890089911" already exists
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: beam_PerformanceTests_TFRecordIOIT_HDFS #3

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/3/display/redirect>

------------------------------------------
[...truncated 78.84 KB...]
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Not a valid TFRecord. Fewer than 12 bytes.
	at org.apache.beam.sdk.repackaged.com.google.common.base.Preconditions.checkState(Preconditions.java:444)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordCodec.read(TFRecordIO.java:631)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordSource$TFRecordReader.readNextRecord(TFRecordIO.java:526)
	at org.apache.beam.sdk.io.CompressedSource$CompressedReader.readNextRecord(CompressedSource.java:426)
	at org.apache.beam.sdk.io.FileBasedSource$FileBasedReader.advanceImpl(FileBasedSource.java:473)
	at org.apache.beam.sdk.io.OffsetBasedSource$OffsetBasedReader.advance(OffsetBasedSource.java:267)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:602)
	... 14 more
java.io.IOException: Failed to advance reader of source: hdfs://146.148.62.9:9000/TEXTIO_IT__1521892046445-00001-of-00003.tfrecord range [0, 19333314)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:605)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation$SynchronizedReaderIterator.advance(ReadOperation.java:379)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.runReadLoop(ReadOperation.java:185)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.start(ReadOperation.java:150)
	at com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:74)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:381)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:353)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:284)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Invalid data
	at org.apache.beam.sdk.repackaged.com.google.common.base.Preconditions.checkState(Preconditions.java:444)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordCodec.read(TFRecordIO.java:642)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordSource$TFRecordReader.readNextRecord(TFRecordIO.java:526)
	at org.apache.beam.sdk.io.CompressedSource$CompressedReader.readNextRecord(CompressedSource.java:426)
	at org.apache.beam.sdk.io.FileBasedSource$FileBasedReader.advanceImpl(FileBasedSource.java:473)
	at org.apache.beam.sdk.io.OffsetBasedSource$OffsetBasedReader.advance(OffsetBasedSource.java:267)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:602)
	... 14 more
java.io.IOException: Failed to advance reader of source: hdfs://146.148.62.9:9000/TEXTIO_IT__1521892046445-00001-of-00003.tfrecord range [0, 19333314)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:605)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation$SynchronizedReaderIterator.advance(ReadOperation.java:379)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.runReadLoop(ReadOperation.java:185)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.start(ReadOperation.java:150)
	at com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:74)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:381)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:353)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:284)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Invalid data
	at org.apache.beam.sdk.repackaged.com.google.common.base.Preconditions.checkState(Preconditions.java:444)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordCodec.read(TFRecordIO.java:642)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordSource$TFRecordReader.readNextRecord(TFRecordIO.java:526)
	at org.apache.beam.sdk.io.CompressedSource$CompressedReader.readNextRecord(CompressedSource.java:426)
	at org.apache.beam.sdk.io.FileBasedSource$FileBasedReader.advanceImpl(FileBasedSource.java:473)
	at org.apache.beam.sdk.io.OffsetBasedSource$OffsetBasedReader.advance(OffsetBasedSource.java:267)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:602)
	... 14 more
java.io.IOException: Failed to advance reader of source: hdfs://146.148.62.9:9000/TEXTIO_IT__1521892046445-00000-of-00003.tfrecord range [0, 19222204)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:605)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation$SynchronizedReaderIterator.advance(ReadOperation.java:379)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.runReadLoop(ReadOperation.java:185)
	at com.google.cloud.dataflow.worker.util.common.worker.ReadOperation.start(ReadOperation.java:150)
	at com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:74)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:381)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:353)
	at com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:284)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Not a valid TFRecord. Fewer than 12 bytes.
	at org.apache.beam.sdk.repackaged.com.google.common.base.Preconditions.checkState(Preconditions.java:444)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordCodec.read(TFRecordIO.java:631)
	at org.apache.beam.sdk.io.TFRecordIO$TFRecordSource$TFRecordReader.readNextRecord(TFRecordIO.java:526)
	at org.apache.beam.sdk.io.CompressedSource$CompressedReader.readNextRecord(CompressedSource.java:426)
	at org.apache.beam.sdk.io.FileBasedSource$FileBasedReader.advanceImpl(FileBasedSource.java:473)
	at org.apache.beam.sdk.io.OffsetBasedSource$OffsetBasedReader.advance(OffsetBasedSource.java:267)
	at com.google.cloud.dataflow.worker.WorkerCustomSources$BoundedReaderIterator.advance(WorkerCustomSources.java:602)
	... 14 more
Workflow failed. Causes: S03:TFRecordIO.Read/Read+Transform bytes to strings/Map+Calculate hashcode/WithKeys/AddKeys/Map+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write failed., A work item was attempted 4 times without success. Each time the worker eventually lost contact with the service. The work item was attempted on: 
  tfrecordioit0writethenrea-03240451-eb64-harness-3ctb,
  tfrecordioit0writethenrea-03240451-eb64-harness-h06g,
  tfrecordioit0writethenrea-03240451-eb64-harness-48v0,
  tfrecordioit0writethenrea-03240451-eb64-harness-3ctb
	at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:134)
	at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:90)
	at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:55)
	at org.apache.beam.sdk.Pipeline.run(Pipeline.java:311)
	at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:346)
	at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:328)
	at org.apache.beam.sdk.io.tfrecord.TFRecordIOIT.writeThenReadAll(TFRecordIOIT.java:127)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
	at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:317)
	at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:317)
	at org.junit.rules.RunRules.evaluate(RunRules.java:20)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
	at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)

[INFO] 
[INFO] Results:
[INFO] 
[ERROR] Errors: 
[ERROR]   TFRecordIOIT.writeThenReadAll:127  Runtime java.io.IOException: Failed to adv...
[INFO] 
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0
[INFO] 
[INFO] 
[INFO] --- maven-dependency-plugin:3.0.2:analyze-only (default) @ beam-sdks-java-io-file-based-io-tests ---
[INFO] No dependency problems found
[INFO] 
[INFO] --- maven-failsafe-plugin:2.21.0:verify (default) @ beam-sdks-java-io-file-based-io-tests ---
[INFO] Failsafe report directory: <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/runs/9fb5402c/beam/sdks/java/io/file-based-io-tests/target/failsafe-reports>
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 07:27 min
[INFO] Finished at: 2018-03-24T11:54:34Z
[INFO] Final Memory: 91M/882M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.21.0:verify (default) on project beam-sdks-java-io-file-based-io-tests: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/runs/9fb5402c/beam/sdks/java/io/file-based-io-tests/target/failsafe-reports> for the individual test results.
[ERROR] Please refer to dump files (if any exist) [date]-jvmRun[N].dump, [date].dumpstream and [date]-jvmRun[N].dumpstream.
[ERROR] -> [Help 1]
org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.21.0:verify (default) on project beam-sdks-java-io-file-based-io-tests: There are test failures.

Please refer to <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/runs/9fb5402c/beam/sdks/java/io/file-based-io-tests/target/failsafe-reports> for the individual test results.
Please refer to dump files (if any exist) [date]-jvmRun[N].dump, [date].dumpstream and [date]-jvmRun[N].dumpstream.
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:213)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146)
    at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)
    at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81)
    at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:51)
    at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128)
    at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:309)
    at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:194)
    at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:107)
    at org.apache.maven.cli.MavenCli.execute (MavenCli.java:955)
    at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:290)
    at org.apache.maven.cli.MavenCli.main (MavenCli.java:194)
    at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke (Method.java:498)
    at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289)
    at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229)
    at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415)
    at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356)
Caused by: org.apache.maven.plugin.MojoFailureException: There are test failures.

Please refer to <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/runs/9fb5402c/beam/sdks/java/io/file-based-io-tests/target/failsafe-reports> for the individual test results.
Please refer to dump files (if any exist) [date]-jvmRun[N].dump, [date].dumpstream and [date]-jvmRun[N].dumpstream.
    at org.apache.maven.plugin.surefire.SurefireHelper.throwException (SurefireHelper.java:240)
    at org.apache.maven.plugin.surefire.SurefireHelper.reportExecution (SurefireHelper.java:112)
    at org.apache.maven.plugin.failsafe.VerifyMojo.execute (VerifyMojo.java:192)
    at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:134)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:208)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154)
    at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146)
    at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)
    at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81)
    at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:51)
    at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128)
    at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:309)
    at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:194)
    at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:107)
    at org.apache.maven.cli.MavenCli.execute (MavenCli.java:955)
    at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:290)
    at org.apache.maven.cli.MavenCli.main (MavenCli.java:194)
    at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke (Method.java:498)
    at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289)
    at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229)
    at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415)
    at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356)
[ERROR] 
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException

STDERR: 
2018-03-24 11:54:34,402 9fb5402c MainThread beam_integration_benchmark(1/1) ERROR    Error during benchmark beam_integration_benchmark
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 624, in RunBenchmark
    DoRunPhase(spec, collector, detailed_timer)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 527, in DoRunPhase
    samples = spec.BenchmarkRun(spec)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",> line 159, in Run
    job_type=job_type)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",> line 90, in SubmitJob
    assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2018-03-24 11:54:34,403 9fb5402c MainThread beam_integration_benchmark(1/1) INFO     Cleaning up benchmark beam_integration_benchmark
2018-03-24 11:54:34,404 9fb5402c MainThread beam_integration_benchmark(1/1) INFO     Running: kubectl --kubeconfig=<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/config-filebasedioithdfs-1521890089911> delete -f <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/src/.test-infra/kubernetes/hadoop/SmallITCluster/hdfs-single-datanode-cluster.yml>
2018-03-24 11:54:34,858 9fb5402c MainThread beam_integration_benchmark(1/1) INFO     Running: kubectl --kubeconfig=<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/config-filebasedioithdfs-1521890089911> delete -f <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/src/.test-infra/kubernetes/hadoop/SmallITCluster/hdfs-single-datanode-cluster-for-local-dev.yml>
2018-03-24 11:54:35,013 9fb5402c MainThread beam_integration_benchmark(1/1) ERROR    Exception running benchmark
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 758, in RunBenchmarkTask
    RunBenchmark(spec, collector)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 624, in RunBenchmark
    DoRunPhase(spec, collector, detailed_timer)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 527, in DoRunPhase
    samples = spec.BenchmarkRun(spec)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",> line 159, in Run
    job_type=job_type)
  File "<https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",> line 90, in SubmitJob
    assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2018-03-24 11:54:35,014 9fb5402c MainThread beam_integration_benchmark(1/1) ERROR    Benchmark 1/1 beam_integration_benchmark (UID: beam_integration_benchmark0) failed. Execution will continue.
2018-03-24 11:54:35,014 9fb5402c MainThread beam_integration_benchmark(1/1) INFO     Benchmark run statuses:
---------------------------------------------------------------------------------
Name                        UID                          Status  Failed Substatus
---------------------------------------------------------------------------------
beam_integration_benchmark  beam_integration_benchmark0  FAILED                  
---------------------------------------------------------------------------------
Success rate: 0.00% (0/1)
2018-03-24 11:54:35,014 9fb5402c MainThread beam_integration_benchmark(1/1) INFO     Complete logs can be found at: <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/runs/9fb5402c/pkb.log>
2018-03-24 11:54:35,014 9fb5402c MainThread beam_integration_benchmark(1/1) INFO     Completion statuses can be found at: <https://builds.apache.org/job/beam_PerformanceTests_TFRecordIOIT_HDFS/ws/runs/9fb5402c/completion_statuses.json>
Build step 'Execute shell' marked build as failure