You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/06/27 18:02:01 UTC

Build failed in Jenkins: beam_PostCommit_Python3_Verify #1239

See <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/1239/display/redirect>

------------------------------------------
Started by timer
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-15 (beam) in workspace <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/>
No credentials specified
 > git rev-parse --is-inside-work-tree # timeout=10
Fetching changes from the remote Git repository
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
ERROR: Error fetching remote repo 'origin'
hudson.plugins.git.GitException: Failed to fetch from https://github.com/apache/beam.git
	at hudson.plugins.git.GitSCM.fetchFrom(GitSCM.java:894)
	at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1161)
	at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1192)
	at hudson.scm.SCM.checkout(SCM.java:504)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1208)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:574)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:499)
	at hudson.model.Run.execute(Run.java:1810)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
Caused by: hudson.plugins.git.GitException: Command "git config remote.origin.url https://github.com/apache/beam.git" returned status code 4:
stdout: 
stderr: error: failed to write new configuration file <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/.git/config.lock>

	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2042)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2010)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2006)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommand(CliGitAPIImpl.java:1638)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommand(CliGitAPIImpl.java:1650)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.setRemoteUrl(CliGitAPIImpl.java:1284)
	at hudson.plugins.git.GitAPI.setRemoteUrl(GitAPI.java:160)
	at sun.reflect.GeneratedMethodAccessor49.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.perform(RemoteInvocationHandler.java:929)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.call(RemoteInvocationHandler.java:903)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.call(RemoteInvocationHandler.java:855)
	at hudson.remoting.UserRequest.perform(UserRequest.java:212)
	at hudson.remoting.UserRequest.perform(UserRequest.java:54)
	at hudson.remoting.Request$2.run(Request.java:369)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at hudson.remoting.Engine$1.lambda$newThread$0(Engine.java:93)
	at java.lang.Thread.run(Thread.java:748)
	Suppressed: hudson.remoting.Channel$CallSiteStackTrace: Remote call to JNLP4-connect connection from 103.55.66.34.bc.googleusercontent.com/34.66.55.103:46040
		at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
		at hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
		at hudson.remoting.Channel.call(Channel.java:955)
		at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:283)
		at com.sun.proxy.$Proxy114.setRemoteUrl(Unknown Source)
		at org.jenkinsci.plugins.gitclient.RemoteGitImpl.setRemoteUrl(RemoteGitImpl.java:295)
		at hudson.plugins.git.GitSCM.fetchFrom(GitSCM.java:882)
		at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1161)
		at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1192)
		at hudson.scm.SCM.checkout(SCM.java:504)
		at hudson.model.AbstractProject.checkout(AbstractProject.java:1208)
		at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:574)
		at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86)
		at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:499)
		at hudson.model.Run.execute(Run.java:1810)
		at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
		at hudson.model.ResourceController.execute(ResourceController.java:97)
		at hudson.model.Executor.run(Executor.java:429)
ERROR: Error fetching remote repo 'origin'
Retrying after 10 seconds
No credentials specified
 > git rev-parse --is-inside-work-tree # timeout=10
Fetching changes from the remote Git repository
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
ERROR: Error fetching remote repo 'origin'
hudson.plugins.git.GitException: Failed to fetch from https://github.com/apache/beam.git
	at hudson.plugins.git.GitSCM.fetchFrom(GitSCM.java:894)
	at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1161)
	at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1192)
	at hudson.scm.SCM.checkout(SCM.java:504)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1208)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:574)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:499)
	at hudson.model.Run.execute(Run.java:1810)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
Caused by: hudson.plugins.git.GitException: Command "git config remote.origin.url https://github.com/apache/beam.git" returned status code 4:
stdout: 
stderr: error: failed to write new configuration file <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/.git/config.lock>

	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2042)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2010)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2006)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommand(CliGitAPIImpl.java:1638)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommand(CliGitAPIImpl.java:1650)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.setRemoteUrl(CliGitAPIImpl.java:1284)
	at hudson.plugins.git.GitAPI.setRemoteUrl(GitAPI.java:160)
	at sun.reflect.GeneratedMethodAccessor49.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.perform(RemoteInvocationHandler.java:929)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.call(RemoteInvocationHandler.java:903)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.call(RemoteInvocationHandler.java:855)
	at hudson.remoting.UserRequest.perform(UserRequest.java:212)
	at hudson.remoting.UserRequest.perform(UserRequest.java:54)
	at hudson.remoting.Request$2.run(Request.java:369)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at hudson.remoting.Engine$1.lambda$newThread$0(Engine.java:93)
	at java.lang.Thread.run(Thread.java:748)
	Suppressed: hudson.remoting.Channel$CallSiteStackTrace: Remote call to JNLP4-connect connection from 103.55.66.34.bc.googleusercontent.com/34.66.55.103:46040
		at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
		at hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
		at hudson.remoting.Channel.call(Channel.java:955)
		at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:283)
		at com.sun.proxy.$Proxy114.setRemoteUrl(Unknown Source)
		at org.jenkinsci.plugins.gitclient.RemoteGitImpl.setRemoteUrl(RemoteGitImpl.java:295)
		at hudson.plugins.git.GitSCM.fetchFrom(GitSCM.java:882)
		at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1161)
		at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1192)
		at hudson.scm.SCM.checkout(SCM.java:504)
		at hudson.model.AbstractProject.checkout(AbstractProject.java:1208)
		at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:574)
		at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86)
		at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:499)
		at hudson.model.Run.execute(Run.java:1810)
		at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
		at hudson.model.ResourceController.execute(ResourceController.java:97)
		at hudson.model.Executor.run(Executor.java:429)
ERROR: Error fetching remote repo 'origin'
Retrying after 10 seconds
No credentials specified
 > git rev-parse --is-inside-work-tree # timeout=10
Fetching changes from the remote Git repository
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
ERROR: Error fetching remote repo 'origin'
hudson.plugins.git.GitException: Failed to fetch from https://github.com/apache/beam.git
	at hudson.plugins.git.GitSCM.fetchFrom(GitSCM.java:894)
	at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1161)
	at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1192)
	at hudson.scm.SCM.checkout(SCM.java:504)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1208)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:574)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:499)
	at hudson.model.Run.execute(Run.java:1810)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
Caused by: hudson.plugins.git.GitException: Command "git config remote.origin.url https://github.com/apache/beam.git" returned status code 4:
stdout: 
stderr: error: failed to write new configuration file <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/.git/config.lock>

	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2042)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2010)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:2006)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommand(CliGitAPIImpl.java:1638)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommand(CliGitAPIImpl.java:1650)
	at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.setRemoteUrl(CliGitAPIImpl.java:1284)
	at hudson.plugins.git.GitAPI.setRemoteUrl(GitAPI.java:160)
	at sun.reflect.GeneratedMethodAccessor49.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.perform(RemoteInvocationHandler.java:929)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.call(RemoteInvocationHandler.java:903)
	at hudson.remoting.RemoteInvocationHandler$RPCRequest.call(RemoteInvocationHandler.java:855)
	at hudson.remoting.UserRequest.perform(UserRequest.java:212)
	at hudson.remoting.UserRequest.perform(UserRequest.java:54)
	at hudson.remoting.Request$2.run(Request.java:369)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at hudson.remoting.Engine$1.lambda$newThread$0(Engine.java:93)
	at java.lang.Thread.run(Thread.java:748)
	Suppressed: hudson.remoting.Channel$CallSiteStackTrace: Remote call to JNLP4-connect connection from 103.55.66.34.bc.googleusercontent.com/34.66.55.103:46040
		at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
		at hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
		at hudson.remoting.Channel.call(Channel.java:955)
		at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:283)
		at com.sun.proxy.$Proxy114.setRemoteUrl(Unknown Source)
		at org.jenkinsci.plugins.gitclient.RemoteGitImpl.setRemoteUrl(RemoteGitImpl.java:295)
		at hudson.plugins.git.GitSCM.fetchFrom(GitSCM.java:882)
		at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1161)
		at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1192)
		at hudson.scm.SCM.checkout(SCM.java:504)
		at hudson.model.AbstractProject.checkout(AbstractProject.java:1208)
		at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:574)
		at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86)
		at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:499)
		at hudson.model.Run.execute(Run.java:1810)
		at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
		at hudson.model.ResourceController.execute(ResourceController.java:97)
		at hudson.model.Executor.run(Executor.java:429)
ERROR: Error fetching remote repo 'origin'

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Python3_Verify #1242

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/1242/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python3_Verify #1241

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/1241/display/redirect?page=changes>

Changes:

[kedin] Spotless config update to include java files only under src directory

------------------------------------------
[...truncated 453.24 KB...]
            },
            "output_name": "out",
            "user_name": "WriteUserScoreSums/Write/WriteImpl/FinalizeWrite.out"
          }
        ],
        "parallel_input": {
          "@type": "OutputReference",
          "output_name": "out",
          "step_name": "s7"
        },
        "serialized_fn": "eNrNV2t73MQV1u7aSRCXhqQJSellm5JWpuzKSRxiu5QCmwTMksXVOonakqojaXZHtqTR0Yy8Mc1SSLpO6BXaQqH3Unq/3++XX9Df1DOzXoctNi1feHgeW9o5M3PmnPe855X0ZMUKSEYCRj2fkqQuc5KKDs8TUQ94Ts0GiWPix/R8TrKM5if56dQEY+opKPWhbLkVwzC8TgqVMScRxz+fCGp6nSglcfQ49Xp5JKkJE+5O3JLlPKBCwCSbcG9ULuRaRj0WpVLAjvF4cELb6yHFgIjkuTAXHl1C80PKbMJODGZXqw83WO5N6IoXMiukdijAbGn3UXrddGOrWIeb/MK/ADdb/x21pBdlxE1vCe/tKF0x4Rb0/pY+7Lbc3eioE8XUy4hkXpbTTnQRbh3zwDPcnYr6KokLXJfz1SikudmWREbBOWVcHNlgDzre24e3Wu4kOtY7YJ8L+Lsr5m1b0iSrCcyWdGkN61GjaViTfHijQgo7W6tFshbEvAjtYdK2P+t3ghMn5mp0do7UZmaD6Zo/fRR/zR2fmZ2h9NiMH9o5FUWM+wtB85pQRYb9rjmKQaMEtwVhFMd1T11NL+Yk1HYTDuiKC5nDwQG8zYHbC18DE/AEEREC0x96eLs7gWZSSA7v0N7TIvEEI3ko4J1Nw71VuVFDLyUJ9RTjiIR3uXegvXbY2pgqkqnp42GNd9B03YGyQVUTyWOUIJzw7pZ7Aw6TKNnI4JA+VNXTzmISpfCe6wXUJ4qiowp4B1OFONyH91psn4srDHgfu42pvCZ1Xsr7OE+0TWzczCXelnmUdhtcl3UKvd3Zh/f77v7xDLsx90dp3qUna3dumVjNPagSU+2WIuw5iWJ076W0h3ckcP1a4Y/zNpU0xy6rFzKKzfvzbpHQVC7GJKCMxzoqG6OaZup6pFnS96O0D8cuwIzl3q3iRE5O186rFj2LtGgrVrSLRNjaNLwuJFlsn95oZ22B42NxZJpApnc2zaJgJaZhG70uqNYz4e4+nLB00XtRGvKel2B+Ki3UjtntBEhFpVtXmBo+Enuv2m3CnGZaSCSB+a3i2YziJC4x4QOY/D19+KDl3qIIGigdUi2tMIR73dvRqHbPq8M9ffo8pporDZxfPQIfYrNsTlNvNaI9Ff59Y70S5JRI5HORBkoLTLjfYgfcXbheEUTJFjwwgIYDJ61mqWngf6W5t3HzNcO4ZBjXSsaVstGGU60BnJ7Su0Znw4MDeMhdRovNeELtZZquRKkY3WsiJqvU7vF8RSAE1FY5eItcyAZPkkh6i2uS8fSYd47mUWfNFnlgi3BFqYiy26/AzR7iVs/WYEGHcE9MEj8k98LDZx4oNQxouvtUJ+U88fIilarlNqN8RGuwhmZDDOHMOrSmJDzqwOIYUl0qPYQd2flhfYxfRLHEVMDR8OK0moX2Oiw5cHZsa5RkPJdewsMiRlE65+5RpXxV6eH8AFwHPqLde7g3kJ4HH12HjznwGHu4tVXhAooDuMCwbFihMlao0tzZbDQuS1STkrFcVnUKdZ2ulIw+DsuGOGJcwqmKEZYNOWFcMXQV5aSaw0VhxVgpG3lTDcMJ42Bb7jDCSWNzXu4czVwfVPSgUzH24+0y+jCQFh+3Ws2yBiekHYIiDp7ikPsEWhYbPI6p5lyVd6oCe7R6OKz2IsmqCXZzVTKCMymt0pgqfagOqU/DKhFVghvSbkwl7lbFq1dPR7mQVdnjo/WiStOAF0pr1B70eOiwOHSXvtbhExKIpRsxjoQEX6uxYoXkPBYQuDvUWMRRQCHUzxCsL1Ddg6eSTK5tdil09HRMU+hqDdfPzVN5znNg7KCEyC1r37CsoRjRbEWfrh4HEF+FhC1ojfW2pml65uVSY7dRPlDaUdpb2lPaVaqUKmXgU0jUzAFgj7G0he8UuQThgBxA4cAqu9CH3jYadpHNiQGsOfD4OnyyD5ck9B14ovCLq/CpTYU98voV9kmGyvmUpUSHzTOlXZf7cMVi/1uoPq333LdB5KHUlJrtxsQlRaUBKsz6FENRufpGi8q1cVF5+sy/S6ypgP+MA59F4J9WwH9Owucd+IIG8IubAB59/QA+owB89pUAfqkPX/5/APzKFgBuodXPIZLPayS/+kYj+cI4ki+iPLMme4Sh4n4N8fy6A99APF9ssTennH1TyRl700jYtyR822I+C1jIKOuwLtNy8x22zFYYSsp32QuvJSkvbS0p31PMftmB72MlXlLM/oGEHzrwowH82IGfKEn56TaS8jMtKT934Bfr8Ms+/ErCrx34je6I39LX/GY7r32hGxN+h4T/fR/+YOlHpIpYSJJkHr6v+/g2mcMfmyWt1fgO2+3SHM/903a+N5aYJ4cPn6WNIfwZz/iLpuIwCfTx1+18DFeYD+q3uWGc+An3N/Twdw0DVrRIipgoEqjnO4V/YID6EzES3uix9098BZbwr/p/AKmW3SQ=",
        "user_name": "WriteUserScoreSums/Write/WriteImpl/FinalizeWrite/FinalizeWrite"
      }
    }
  ],
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: '2019-06-27T20:13:57.208654Z'
 currentStateTime: '1970-01-01T00:00:00Z'
 id: '2019-06-27_13_13_56-7997272284617937870'
 location: 'us-central1'
 name: 'beamapp-jenkins-0627201316-325310'
 projectId: 'apache-beam-testing'
 stageStates: []
 startTime: '2019-06-27T20:13:57.208654Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2019-06-27_13_13_56-7997272284617937870]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_56-7997272284617937870?project=apache-beam-testing
root: INFO: Job 2019-06-27_13_13_56-7997272284617937870 is in state JOB_STATE_RUNNING
root: INFO: 2019-06-27T20:13:56.347Z: JOB_MESSAGE_DETAILED: Autoscaling is enabled for job 2019-06-27_13_13_56-7997272284617937870. The number of workers will be between 1 and 1000.
root: INFO: 2019-06-27T20:13:56.401Z: JOB_MESSAGE_DETAILED: Autoscaling was automatically enabled for job 2019-06-27_13_13_56-7997272284617937870.
root: INFO: 2019-06-27T20:13:59.494Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-06-27T20:14:00.274Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-a.
root: INFO: 2019-06-27T20:14:00.919Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2019-06-27T20:14:00.987Z: JOB_MESSAGE_DEBUG: Combiner lifting skipped for step WriteUserScoreSums/Write/WriteImpl/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2019-06-27T20:14:01.063Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
root: INFO: 2019-06-27T20:14:01.127Z: JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-06-27T20:14:01.292Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
root: INFO: 2019-06-27T20:14:01.404Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2019-06-27T20:14:01.474Z: JOB_MESSAGE_DETAILED: Fusing consumer UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Reify into UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey+UserScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Partial
root: INFO: 2019-06-27T20:14:01.532Z: JOB_MESSAGE_DETAILED: Fusing consumer UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Write into UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Reify
root: INFO: 2019-06-27T20:14:01.584Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/Write/WriteImpl/GroupByKey/Write into WriteUserScoreSums/Write/WriteImpl/GroupByKey/Reify
root: INFO: 2019-06-27T20:14:01.649Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/Write/WriteImpl/GroupByKey/Reify into WriteUserScoreSums/Write/WriteImpl/WindowInto(WindowIntoFn)
root: INFO: 2019-06-27T20:14:01.703Z: JOB_MESSAGE_DETAILED: Fusing consumer UserScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Extract into UserScore/ExtractAndSumScore/CombinePerKey(sum)/Combine
root: INFO: 2019-06-27T20:14:01.752Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/Write/WriteImpl/Extract into WriteUserScoreSums/Write/WriteImpl/GroupByKey/GroupByWindow
root: INFO: 2019-06-27T20:14:01.793Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/Write/WriteImpl/Pair into WriteUserScoreSums/Write/WriteImpl/WriteBundles/WriteBundles
root: INFO: 2019-06-27T20:14:01.822Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/Write/WriteImpl/WindowInto(WindowIntoFn) into WriteUserScoreSums/Write/WriteImpl/Pair
root: INFO: 2019-06-27T20:14:01.881Z: JOB_MESSAGE_DETAILED: Fusing consumer UserScore/ParseGameEventFn into ReadInputText/Read
root: INFO: 2019-06-27T20:14:01.929Z: JOB_MESSAGE_DETAILED: Fusing consumer UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey+UserScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Partial into UserScore/ExtractAndSumScore/Map(<lambda at user_score.py:111>)
root: INFO: 2019-06-27T20:14:01.990Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/Write/WriteImpl/WriteBundles/WriteBundles into FormatUserScoreSums
root: INFO: 2019-06-27T20:14:02.048Z: JOB_MESSAGE_DETAILED: Fusing consumer FormatUserScoreSums into UserScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Extract
root: INFO: 2019-06-27T20:14:02.113Z: JOB_MESSAGE_DETAILED: Fusing consumer UserScore/ExtractAndSumScore/Map(<lambda at user_score.py:111>) into UserScore/ParseGameEventFn
root: INFO: 2019-06-27T20:14:02.159Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/Write/WriteImpl/GroupByKey/GroupByWindow into WriteUserScoreSums/Write/WriteImpl/GroupByKey/Read
root: INFO: 2019-06-27T20:14:02.227Z: JOB_MESSAGE_DETAILED: Fusing consumer UserScore/ExtractAndSumScore/CombinePerKey(sum)/Combine into UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Read
root: INFO: 2019-06-27T20:14:02.300Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/Write/WriteImpl/InitializeWrite into WriteUserScoreSums/Write/WriteImpl/DoOnce/Read
root: INFO: 2019-06-27T20:14:02.363Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
root: INFO: 2019-06-27T20:14:02.423Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
root: INFO: 2019-06-27T20:14:02.492Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
root: INFO: 2019-06-27T20:14:02.560Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-06-27T20:14:02.882Z: JOB_MESSAGE_DEBUG: Executing wait step start35
root: INFO: 2019-06-27T20:14:03.042Z: JOB_MESSAGE_BASIC: Executing operation WriteUserScoreSums/Write/WriteImpl/DoOnce/Read+WriteUserScoreSums/Write/WriteImpl/InitializeWrite
root: INFO: 2019-06-27T20:14:03.100Z: JOB_MESSAGE_BASIC: Executing operation WriteUserScoreSums/Write/WriteImpl/GroupByKey/Create
root: INFO: 2019-06-27T20:14:03.112Z: JOB_MESSAGE_DEBUG: Starting worker pool setup.
root: INFO: 2019-06-27T20:14:03.145Z: JOB_MESSAGE_BASIC: Executing operation UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Create
root: INFO: 2019-06-27T20:14:03.173Z: JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-a...
root: INFO: 2019-06-27T20:14:03.334Z: JOB_MESSAGE_BASIC: Finished operation WriteUserScoreSums/Write/WriteImpl/GroupByKey/Create
root: INFO: 2019-06-27T20:14:03.334Z: JOB_MESSAGE_BASIC: Finished operation UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Create
root: INFO: 2019-06-27T20:14:03.513Z: JOB_MESSAGE_DEBUG: Value "WriteUserScoreSums/Write/WriteImpl/GroupByKey/Session" materialized.
root: INFO: 2019-06-27T20:14:03.570Z: JOB_MESSAGE_DEBUG: Value "UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Session" materialized.
root: INFO: 2019-06-27T20:14:03.718Z: JOB_MESSAGE_BASIC: Executing operation ReadInputText/Read+UserScore/ParseGameEventFn+UserScore/ExtractAndSumScore/Map(<lambda at user_score.py:111>)+UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey+UserScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Partial+UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Reify+UserScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Write
root: INFO: 2019-06-27T20:14:39.593Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 1 to 5.
root: INFO: 2019-06-27T20:15:47.394Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 1 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:15:47.450Z: JOB_MESSAGE_DETAILED: Resized worker pool to 1, though goal was 5.  This could be a quota issue.
root: INFO: 2019-06-27T20:16:19.746Z: JOB_MESSAGE_DETAILED: Workers have started successfully.
root: INFO: 2019-06-27T20:16:19.797Z: JOB_MESSAGE_DETAILED: Workers have started successfully.
root: INFO: 2019-06-27T20:16:48.188Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 3 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:16:48.252Z: JOB_MESSAGE_DETAILED: Resized worker pool to 3, though goal was 5.  This could be a quota issue.
root: INFO: 2019-06-27T20:18:13.293Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 4 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:18:13.348Z: JOB_MESSAGE_DETAILED: Resized worker pool to 4, though goal was 5.  This could be a quota issue.
root: INFO: 2019-06-27T20:19:12.271Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 5 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:19:42.421Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 5 to 9.
root: INFO: 2019-06-27T20:20:12.746Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 9 to 13.
root: INFO: 2019-06-27T20:20:28.988Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 6 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:20:29.046Z: JOB_MESSAGE_DETAILED: Resized worker pool to 6, though goal was 13.  This could be a quota issue.
root: INFO: 2019-06-27T20:20:35.638Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 8 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:20:35.692Z: JOB_MESSAGE_DETAILED: Resized worker pool to 8, though goal was 17.  This could be a quota issue.
root: INFO: 2019-06-27T20:20:40.759Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 13 to 17.
root: INFO: 2019-06-27T20:20:41.436Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 9 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:20:41.493Z: JOB_MESSAGE_DETAILED: Resized worker pool to 9, though goal was 17.  This could be a quota issue.
root: INFO: 2019-06-27T20:21:03.801Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 10 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:21:03.871Z: JOB_MESSAGE_DETAILED: Resized worker pool to 10, though goal was 17.  This could be a quota issue.
root: INFO: 2019-06-27T20:21:09.168Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 12 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:21:09.219Z: JOB_MESSAGE_DETAILED: Resized worker pool to 12, though goal was 17.  This could be a quota issue.
root: INFO: 2019-06-27T20:21:31.200Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 15 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:21:31.250Z: JOB_MESSAGE_DETAILED: Resized worker pool to 15, though goal was 19.  This could be a quota issue.
root: INFO: 2019-06-27T20:21:35.535Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 17 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-27T20:21:35.589Z: JOB_MESSAGE_DETAILED: Resized worker pool to 17, though goal was 19.  This could be a quota issue.
root: INFO: 2019-06-27T20:21:40.655Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 17 to 19.
apache_beam.io.filesystem: DEBUG: Listing files in 'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/b8bfc779-e89a-48c0-b02a-495484ee34bd/results'
apache_beam.io.filesystem: DEBUG: translate_pattern: 'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/b8bfc779-e89a-48c0-b02a-495484ee34bd/results*' -> 'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/b8bfc779\\-e89a\\-48c0\\-b02a\\-495484ee34bd\\/results[^/\\\\]*'
root: DEBUG: Connecting using Google Application Default Credentials.
root: INFO: Starting the size estimation of the input
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: INFO: Finished listing 0 files in 0.06506705284118652 seconds.
root: DEBUG: Connecting using Google Application Default Credentials.
--------------------- >> end captured logging << ---------------------
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_57-2150893249261969016?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1137: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  method_to_use = self._compute_method(p, p.options)
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery_file_loads.py>:557: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  temp_location = p.options.view_as(GoogleCloudOptions).temp_location
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_28_58-16706981435663458945?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_37_27-4041501762654797870?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:687: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_54-5734116694373971580?project=apache-beam-testing.
  kms_key=transform.kms_key))
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/fileio_test.py>:232: FutureWarning: MatchAll is experimental.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_36_35-11128685942147359719?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_45_00-13852432633697403541?project=apache-beam-testing.
  | 'GetPath' >> beam.Map(lambda metadata: metadata.path))
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/fileio_test.py>:243: FutureWarning: MatchAll is experimental.
  | 'Checksums' >> beam.Map(compute_hash))
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/fileio_test.py>:243: FutureWarning: ReadMatches is experimental.
  | 'Checksums' >> beam.Map(compute_hash))
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_55-4173845759093182923?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1137: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_27_06-9343907808090350603?project=apache-beam-testing.
  method_to_use = self._compute_method(p, p.options)
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_36_19-14501074158301855725?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_44_42-16419364952283482405?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_56-7997272284617937870?project=apache-beam-testing.
Exception in thread Thread-1:
Traceback (most recent call last):
  File "/usr/lib/python3.5/threading.py", line 914, in _bootstrap_inner
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_27_20-18152486730939230067?project=apache-beam-testing.
    self.run()
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_34_41-5303780965774690905?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_42_22-1829962271628920304?project=apache-beam-testing.
  File "/usr/lib/python3.5/threading.py", line 862, in run
    self._target(*self._args, **self._kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 190, in poll_for_job_completion
    job_id, page_token=page_token, start_time=last_message_time)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/utils/retry.py",> line 197, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 748, in list_messages
    response = self._client.projects_locations_jobs_messages.List(request)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py",> line 553, in List
    config, request, global_params=global_params)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967054/lib/python3.5/site-packages/apitools/base/py/base_api.py",> line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967054/lib/python3.5/site-packages/apitools/base/py/base_api.py",> line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967054/lib/python3.5/site-packages/apitools/base/py/base_api.py",> line 604, in __ProcessHttpResponse
    http_response, method_config=method_config, request=request)
apitools.base.py.exceptions.HttpNotFoundError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs/2019-06-27_13_13_56-7997272284617937870/messages?alt=json&startTime=2019-06-27T20%3A21%3A40.655Z>: response: <{'transfer-encoding': 'chunked', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', '-content-encoding': 'gzip', 'date': 'Thu, 27 Jun 2019 20:22:52 GMT', 'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'server': 'ESF', 'cache-control': 'private', 'x-content-type-options': 'nosniff', 'content-length': '279', 'status': '404'}>, content <{
  "error": {
    "code": 404,
    "message": "(f6edd7203539588b): Information about job 2019-06-27_13_13_56-7997272284617937870 could not be found in our system. Please double check the id is correct. If it is please contact customer support.",
    "status": "NOT_FOUND"
  }
}
>

Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_54-12714228266890436570?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:687: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
  kms_key=transform.kms_key))
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_23_46-12951024352623691913?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_32_28-13685993652260595888?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_41_01-221194846515054151?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_48_53-2955327925777682029?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_55-10652175835653268118?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1137: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  method_to_use = self._compute_method(p, p.options)
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery_file_loads.py>:557: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_26_56-14723539225137016945?project=apache-beam-testing.
  temp_location = p.options.view_as(GoogleCloudOptions).temp_location
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_35_17-9432516871372980835?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_43_55-5062290127227603128?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_51_55-17777614111712220917?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_51-5360721188581380524?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_24_59-4028830795442355958?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_33_56-250049340778017327?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_42_08-17079616723281310432?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_50_39-17084894420546299236?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:687: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
  kms_key=transform.kms_key))
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/big_query_query_to_table_pipeline.py>:73: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
  kms_key=kms_key))
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_13_53-2190904417282297212?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_24_04-9274260789737294663?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_32_00-8917021748254858038?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_40_32-6804696549179376206?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_49_14-2080721347009464466?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_13_57_54-6998512472638205853?project=apache-beam-testing.

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 42 tests in 3194.435s

FAILED (SKIP=5, errors=2, failures=1)

> Task :sdks:python:test-suites:dataflow:py35:postCommitIT FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/test-suites/dataflow/py36/build.gradle'> line: 48

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py36:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/test-suites/dataflow/py35/build.gradle'> line: 48

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py35:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 54m 16s
77 actionable tasks: 60 executed, 17 from cache

Publishing build scan...
https://gradle.com/s/prmljbbbzwtxu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python3_Verify #1240

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/1240/display/redirect?page=changes>

Changes:

[valentyn] Match Python 3 warning message in __init__.py with the one in setup.py.

------------------------------------------
[...truncated 293.63 KB...]
 stageStates: []
 startTime: '2019-06-27T18:21:15.942838Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_STREAMING, 2)>
root: INFO: Created job with id: [2019-06-27_11_21_14-15039767703063158577]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_14-15039767703063158577?project=apache-beam-testing
root: INFO: Job 2019-06-27_11_21_14-15039767703063158577 is in state JOB_STATE_RUNNING
root: INFO: 2019-06-27T18:21:18.944Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-06-27T18:21:19.689Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-4 in us-central1-b.
root: INFO: 2019-06-27T18:21:20.224Z: JOB_MESSAGE_DETAILED: Expanding SplittableParDo operations into optimizable parts.
root: INFO: 2019-06-27T18:21:20.228Z: JOB_MESSAGE_DETAILED: Expanding CollectionToSingleton operations into optimizable parts.
root: INFO: 2019-06-27T18:21:20.245Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2019-06-27T18:21:20.262Z: JOB_MESSAGE_DETAILED: Expanding SplittableProcessKeyed operations into optimizable parts.
root: INFO: 2019-06-27T18:21:20.268Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into streaming Read/Write steps
root: INFO: 2019-06-27T18:21:20.284Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
root: INFO: 2019-06-27T18:21:20.315Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2019-06-27T18:21:20.324Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateUserScores/LeaderboardUserGlobalWindows into AddEventTimestamps
root: INFO: 2019-06-27T18:21:20.330Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateTeamScores/LeaderboardTeamFixedWindows into AddEventTimestamps
root: INFO: 2019-06-27T18:21:20.340Z: JOB_MESSAGE_DETAILED: Fusing consumer AddEventTimestamps into ParseGameEventFn
root: INFO: 2019-06-27T18:21:20.344Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/MergeBuckets into CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/ReadStream
root: INFO: 2019-06-27T18:21:20.355Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateUserScores/ExtractAndSumScore/Map(<lambda at leader_board.py:147>) into CalculateUserScores/LeaderboardUserGlobalWindows
root: INFO: 2019-06-27T18:21:20.359Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/WriteStream into CalculateTeamScores/ExtractAndSumScore/Map(<lambda at leader_board.py:147>)
root: INFO: 2019-06-27T18:21:20.368Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteTeamScoreSums/WriteToBigQuery/StreamInsertRows/ParDo(BigQueryWriteFn) into WriteTeamScoreSums/WriteToBigQuery/AppendDestination
root: INFO: 2019-06-27T18:21:20.373Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/WriteToBigQuery/StreamInsertRows/ParDo(BigQueryWriteFn) into WriteUserScoreSums/WriteToBigQuery/AppendDestination
root: INFO: 2019-06-27T18:21:20.379Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/WriteToBigQuery/AppendDestination into WriteUserScoreSums/ConvertToRow
root: INFO: 2019-06-27T18:21:20.384Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteUserScoreSums/ConvertToRow into FormatUserScoreSums
root: INFO: 2019-06-27T18:21:20.388Z: JOB_MESSAGE_DETAILED: Fusing consumer FormatUserScoreSums into CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/Combine
root: INFO: 2019-06-27T18:21:20.391Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/Combine into CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/MergeBuckets
root: INFO: 2019-06-27T18:21:20.395Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/MergeBuckets into CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/ReadStream
root: INFO: 2019-06-27T18:21:20.398Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/WriteStream into CalculateUserScores/ExtractAndSumScore/Map(<lambda at leader_board.py:147>)
root: INFO: 2019-06-27T18:21:20.403Z: JOB_MESSAGE_DETAILED: Fusing consumer DecodeString into ReadPubSub/Read
root: INFO: 2019-06-27T18:21:20.406Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateTeamScores/ExtractAndSumScore/Map(<lambda at leader_board.py:147>) into CalculateTeamScores/LeaderboardTeamFixedWindows
root: INFO: 2019-06-27T18:21:20.409Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteTeamScoreSums/WriteToBigQuery/AppendDestination into WriteTeamScoreSums/ConvertToRow
root: INFO: 2019-06-27T18:21:20.414Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteTeamScoreSums/ConvertToRow into TeamScoresDict
root: INFO: 2019-06-27T18:21:20.419Z: JOB_MESSAGE_DETAILED: Fusing consumer ParseGameEventFn into DecodeString
root: INFO: 2019-06-27T18:21:20.423Z: JOB_MESSAGE_DETAILED: Fusing consumer CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/Combine into CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/MergeBuckets
root: INFO: 2019-06-27T18:21:20.427Z: JOB_MESSAGE_DETAILED: Fusing consumer TeamScoresDict into CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/Combine
root: INFO: 2019-06-27T18:21:20.443Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
root: INFO: 2019-06-27T18:21:20.498Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
root: INFO: 2019-06-27T18:21:20.549Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-06-27T18:21:20.750Z: JOB_MESSAGE_DEBUG: Executing wait step start2
root: INFO: 2019-06-27T18:21:20.770Z: JOB_MESSAGE_DEBUG: Starting worker pool setup.
root: INFO: 2019-06-27T18:21:20.777Z: JOB_MESSAGE_BASIC: Starting 1 workers...
root: INFO: 2019-06-27T18:21:24.836Z: JOB_MESSAGE_BASIC: Executing operation ReadPubSub/Read+DecodeString+ParseGameEventFn+AddEventTimestamps+CalculateUserScores/LeaderboardUserGlobalWindows+CalculateTeamScores/LeaderboardTeamFixedWindows+CalculateUserScores/ExtractAndSumScore/Map(<lambda at leader_board.py:147>)+CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/WriteStream+CalculateTeamScores/ExtractAndSumScore/Map(<lambda at leader_board.py:147>)+CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/WriteStream
root: INFO: 2019-06-27T18:21:24.836Z: JOB_MESSAGE_BASIC: Executing operation CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/ReadStream+CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/MergeBuckets+CalculateTeamScores/ExtractAndSumScore/CombinePerKey(sum)/Combine+TeamScoresDict+WriteTeamScoreSums/ConvertToRow+WriteTeamScoreSums/WriteToBigQuery/AppendDestination+WriteTeamScoreSums/WriteToBigQuery/StreamInsertRows/ParDo(BigQueryWriteFn)
root: INFO: 2019-06-27T18:21:24.836Z: JOB_MESSAGE_BASIC: Executing operation CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/ReadStream+CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/MergeBuckets+CalculateUserScores/ExtractAndSumScore/CombinePerKey(sum)/Combine+FormatUserScoreSums+WriteUserScoreSums/ConvertToRow+WriteUserScoreSums/WriteToBigQuery/AppendDestination+WriteUserScoreSums/WriteToBigQuery/StreamInsertRows/ParDo(BigQueryWriteFn)
root: WARNING: Timing out on waiting for job 2019-06-27_11_21_14-15039767703063158577 after 606 seconds
root: INFO: Start verify Bigquery data.
google.auth.transport._http_client: DEBUG: Making request: GET http://169.254.169.254
google.auth.transport._http_client: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/project/project-id
urllib3.util.retry: DEBUG: Converted retries value: 3 -> Retry(total=3, connect=None, read=None, redirect=None, status=None)
google.auth.transport.requests: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true
urllib3.connectionpool: DEBUG: Starting new HTTP connection (1): metadata.google.internal:80
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET /computeMetadata/v1/instance/service-accounts/default/?recursive=true HTTP/1.1" 200 144
google.auth.transport.requests: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/844138762903-compute@developer.gserviceaccount.com/token
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET /computeMetadata/v1/instance/service-accounts/844138762903-compute@developer.gserviceaccount.com/token HTTP/1.1" 200 176
urllib3.connectionpool: DEBUG: Starting new HTTPS connection (1): www.googleapis.com:443
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST /bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET /bigquery/v2/projects/apache-beam-testing/queries/4cecabf0-3437-43e7-9822-6cb0d7075e01?maxResults=0&location=US HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET /bigquery/v2/projects/apache-beam-testing/jobs/4cecabf0-3437-43e7-9822-6cb0d7075e01?location=US HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET /bigquery/v2/projects/apache-beam-testing/datasets/_7357fab0f784d2a7327ddbe81cdd1f4ca7e429cd/tables/anon568e62d7_3be7_4c4c_b300_b8cd45cd9f42/data HTTP/1.1" 200 None
root: INFO: Read from given query (SELECT total_score FROM `apache-beam-testing.leader_board_it_dataset1561659659.leader_board_users` WHERE total_score=5000 LIMIT 1), total rows 1
root: INFO: Generate checksum: de00231fe6730b972c0ff60a99988438911cda53
root: INFO: Start verify Bigquery data.
google.auth.transport._http_client: DEBUG: Making request: GET http://169.254.169.254
google.auth.transport._http_client: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/project/project-id
urllib3.util.retry: DEBUG: Converted retries value: 3 -> Retry(total=3, connect=None, read=None, redirect=None, status=None)
google.auth.transport.requests: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true
urllib3.connectionpool: DEBUG: Starting new HTTP connection (1): metadata.google.internal:80
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET /computeMetadata/v1/instance/service-accounts/default/?recursive=true HTTP/1.1" 200 144
google.auth.transport.requests: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/844138762903-compute@developer.gserviceaccount.com/token
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET /computeMetadata/v1/instance/service-accounts/844138762903-compute@developer.gserviceaccount.com/token HTTP/1.1" 200 176
urllib3.connectionpool: DEBUG: Starting new HTTPS connection (1): www.googleapis.com:443
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST /bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
root: WARNING: Retry with exponential backoff: waiting for 2.5667962602553205 seconds before retrying _query_with_retry because we caught exception: google.api_core.exceptions.NotFound: 404 Not found: Table apache-beam-testing:leader_board_it_dataset1561659659.leader_board_teams was not found in location US
 Traceback for above exception (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/utils/retry.py",> line 197, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/tests/bigquery_matcher.py",> line 99, in _query_with_retry
    return [row.values() for row in query_job]
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 2718, in __iter__
    return iter(self.result())
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 2685, in result
    super(QueryJob, self).result(timeout=timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 697, in result
    return super(_AsyncJob, self).result(timeout=timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/api_core/future/polling.py",> line 127, in result
    raise self._exception

urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST /bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
root: WARNING: Retry with exponential backoff: waiting for 5.417510467871727 seconds before retrying _query_with_retry because we caught exception: google.api_core.exceptions.NotFound: 404 Not found: Table apache-beam-testing:leader_board_it_dataset1561659659.leader_board_teams was not found in location US
 Traceback for above exception (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/utils/retry.py",> line 197, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/tests/bigquery_matcher.py",> line 99, in _query_with_retry
    return [row.values() for row in query_job]
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 2718, in __iter__
    return iter(self.result())
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 2685, in result
    super(QueryJob, self).result(timeout=timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 697, in result
    return super(_AsyncJob, self).result(timeout=timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/api_core/future/polling.py",> line 127, in result
    raise self._exception

urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST /bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
root: WARNING: Retry with exponential backoff: waiting for 10.917736116501898 seconds before retrying _query_with_retry because we caught exception: google.api_core.exceptions.NotFound: 404 Not found: Table apache-beam-testing:leader_board_it_dataset1561659659.leader_board_teams was not found in location US
 Traceback for above exception (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/utils/retry.py",> line 197, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/tests/bigquery_matcher.py",> line 99, in _query_with_retry
    return [row.values() for row in query_job]
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 2718, in __iter__
    return iter(self.result())
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 2685, in result
    super(QueryJob, self).result(timeout=timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 697, in result
    return super(_AsyncJob, self).result(timeout=timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/api_core/future/polling.py",> line 127, in result
    raise self._exception

urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST /bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
root: WARNING: Retry with exponential backoff: waiting for 28.487514440342785 seconds before retrying _query_with_retry because we caught exception: google.api_core.exceptions.NotFound: 404 Not found: Table apache-beam-testing:leader_board_it_dataset1561659659.leader_board_teams was not found in location US
 Traceback for above exception (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/utils/retry.py",> line 197, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/tests/bigquery_matcher.py",> line 99, in _query_with_retry
    return [row.values() for row in query_job]
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 2718, in __iter__
    return iter(self.result())
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 2685, in result
    super(QueryJob, self).result(timeout=timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/cloud/bigquery/job.py",> line 697, in result
    return super(_AsyncJob, self).result(timeout=timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/google/api_core/future/polling.py",> line 127, in result
    raise self._exception

urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST /bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
google.auth.transport._http_client: DEBUG: Making request: GET http://169.254.169.254
google.auth.transport._http_client: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/project/project-id
urllib3.util.retry: DEBUG: Converted retries value: 3 -> Retry(total=3, connect=None, read=None, redirect=None, status=None)
google.auth.transport.requests: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true
urllib3.connectionpool: DEBUG: Starting new HTTP connection (1): metadata.google.internal:80
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET /computeMetadata/v1/instance/service-accounts/default/?recursive=true HTTP/1.1" 200 144
google.auth.transport.requests: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/844138762903-compute@developer.gserviceaccount.com/token
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET /computeMetadata/v1/instance/service-accounts/844138762903-compute@developer.gserviceaccount.com/token HTTP/1.1" 200 176
urllib3.connectionpool: DEBUG: Starting new HTTPS connection (1): www.googleapis.com:443
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "DELETE /bigquery/v2/projects/apache-beam-testing/datasets/leader_board_it_dataset1561659659?deleteContents=true HTTP/1.1" 204 0
--------------------- >> end captured logging << ---------------------
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1137: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_15-12389678040481198720?project=apache-beam-testing.
  method_to_use = self._compute_method(p, p.options)
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_37_12-960883275587796161?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_46_58-3814286599063432897?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_54_33-13346598284442458478?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:687: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_11-615940475463559441?project=apache-beam-testing.
  kms_key=transform.kms_key))
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_45_13-17985987452624809073?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_53_48-5510950477228618353?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1137: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  method_to_use = self._compute_method(p, p.options)
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/fileio_test.py>:232: FutureWarning: MatchAll is experimental.
  | 'GetPath' >> beam.Map(lambda metadata: metadata.path))
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/fileio_test.py>:243: FutureWarning: MatchAll is experimental.
  | 'Checksums' >> beam.Map(compute_hash))
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/fileio_test.py>:243: FutureWarning: ReadMatches is experimental.
  | 'Checksums' >> beam.Map(compute_hash))
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_14-15039767703063158577?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_35_38-18044910366310283199?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_45_21-1108022985918176837?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_53_51-543035880815273578?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_12_01_51-3269068709917556615?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:687: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_12-9660353995653841975?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_43_48-11196328050531719094?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_52_40-11442956218216567202?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_12_01_09-17727655758564341191?project=apache-beam-testing.
  kms_key=transform.kms_key))
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_13-13251294892337466203?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_33_53-14783270863838739624?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_43_01-13334170186257560357?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_50_56-8842417339630429752?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_57_34-2908525405295890699?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:687: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
  kms_key=transform.kms_key))
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/big_query_query_to_table_pipeline.py>:73: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
  kms_key=kms_key))
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_16-12879557786636645854?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1137: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  method_to_use = self._compute_method(p, p.options)
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery_file_loads.py>:557: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  temp_location = p.options.view_as(GoogleCloudOptions).temp_location
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1137: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  method_to_use = self._compute_method(p, p.options)
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/io/gcp/bigquery_file_loads.py>:557: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  temp_location = p.options.view_as(GoogleCloudOptions).temp_location
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_30_03-8221616964317028141?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_38_14-5976029510644698695?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_46_48-10813681871349375293?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_13-8339368527426944894?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_34_11-18199207548725928748?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_44_07-18278998614158871141?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_52_50-11773408247184126469?project=apache-beam-testing.
<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:687: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_21_10-6232782436336644687?project=apache-beam-testing.
  kms_key=transform.kms_key))
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_32_47-11479446652272137610?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_40_57-8035064405539412294?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_11_52_22-6776960090671605256?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_12_01_10-15282477350379204684?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-27_12_09_41-5124406333126387929?project=apache-beam-testing.

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 42 tests in 3427.945s

FAILED (SKIP=5, errors=1)

> Task :sdks:python:test-suites:dataflow:py37:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python3_Verify/ws/src/sdks/python/test-suites/dataflow/py37/build.gradle'> line: 78

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py37:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 58m 7s
77 actionable tasks: 60 executed, 17 from cache

Publishing build scan...
https://gradle.com/s/3byoxrlghsmsm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org