You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by "ASF GitHub Bot (JIRA)" <ji...@apache.org> on 2018/10/10 13:15:00 UTC

[jira] [Work logged] (BEAM-5299) Define max global window as a shared value in protos like URN enums.

     [ https://issues.apache.org/jira/browse/BEAM-5299?focusedWorklogId=153126&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-153126 ]

ASF GitHub Bot logged work on BEAM-5299:
----------------------------------------

                Author: ASF GitHub Bot
            Created on: 10/Oct/18 13:14
            Start Date: 10/Oct/18 13:14
    Worklog Time Spent: 10m 
      Work Description: mxm commented on issue #6381: [BEAM-5299] Define max timestamp for global window in proto
URL: https://github.com/apache/beam/pull/6381#issuecomment-428566326
 
 
   It appears this won't work with Dataflow. Any ideas how to fix the failures below?
   
   ```
   com.google.cloud.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException: java.lang.IllegalArgumentException: unable to deserialize Serialized DoFnInfo
       	at com.google.cloud.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:192)
       	at com.google.cloud.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:163)
       	at com.google.cloud.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:63)
       	at com.google.cloud.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:50)
       	at com.google.cloud.dataflow.worker.graph.Networks.replaceDirectedNetworkNodes(Networks.java:87)
       	at com.google.cloud.dataflow.worker.IntrinsicMapTaskExecutorFactory.create(IntrinsicMapTaskExecutorFactory.java:123)
       	at com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:336)
       	at com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:290)
       	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
       	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
       	at com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
       	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
       	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
       	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
       	at java.lang.Thread.run(Thread.java:745)
       Caused by: com.google.cloud.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException: java.lang.IllegalArgumentException: unable to deserialize Serialized DoFnInfo
       	at com.google.cloud.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2214)
       	at com.google.cloud.dataflow.worker.repackaged.com.google.common.cache.LocalCache.get(LocalCache.java:4053)
       	at com.google.cloud.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4899)
       	at com.google.cloud.dataflow.worker.UserParDoFnFactory.create(UserParDoFnFactory.java:90)
       	at com.google.cloud.dataflow.worker.DefaultParDoFnFactory.create(DefaultParDoFnFactory.java:74)
       	at com.google.cloud.dataflow.worker.IntrinsicMapTaskExecutorFactory.createParDoOperation(IntrinsicMapTaskExecutorFactory.java:262)
       	at com.google.cloud.dataflow.worker.IntrinsicMapTaskExecutorFactory.access$000(IntrinsicMapTaskExecutorFactory.java:84)
       	at com.google.cloud.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:181)
       	... 14 more
       Caused by: java.lang.IllegalArgumentException: unable to deserialize Serialized DoFnInfo
       	at org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:74)
       	at com.google.cloud.dataflow.worker.UserParDoFnFactory$UserDoFnExtractor.getDoFnInfo(UserParDoFnFactory.java:61)
       	at com.google.cloud.dataflow.worker.UserParDoFnFactory.lambda$create$0(UserParDoFnFactory.java:92)
       	at com.google.cloud.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4904)
       	at com.google.cloud.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3628)
       	at com.google.cloud.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2336)
       	at com.google.cloud.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2295)
       	at com.google.cloud.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2208)
       	... 21 more
       Caused by: java.io.InvalidClassException: org.apache.beam.sdk.values.WindowingStrategy; local class incompatible: stream classdesc serialVersionUID = -6607512772692666907, local class serialVersionUID = -3616600070988263902
       	at java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:616)
       	at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1630)
       	at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1521)
       	at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1781)
       	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353)
       	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2018)
       	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1942)
       	at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1808)
       	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353)
       	at java.io.ObjectInputStream.readObject(ObjectInputStream.java:373)
       	at org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:71)
       	... 28 more
   ```

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
users@infra.apache.org


Issue Time Tracking
-------------------

    Worklog Id:     (was: 153126)
    Time Spent: 3.5h  (was: 3h 20m)

> Define max global window as a shared value in protos like URN enums.
> --------------------------------------------------------------------
>
>                 Key: BEAM-5299
>                 URL: https://issues.apache.org/jira/browse/BEAM-5299
>             Project: Beam
>          Issue Type: Improvement
>          Components: beam-model, sdk-go, sdk-java-core, sdk-py-core
>            Reporter: Luke Cwik
>            Assignee: Maximilian Michels
>            Priority: Minor
>              Labels: portability
>          Time Spent: 3.5h
>  Remaining Estimate: 0h
>
> Instead of having each language define a max timestamp themselves, define the max timestamps within proto to be shared across different languages.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)