You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@beam.apache.org by be...@gmail.com on 2022/06/21 10:57:26 UTC

P1 issues report (71)

This is your daily summary of Beam's current P1 issues, not including flaky tests.

    See https://beam.apache.org/contribute/issue-priorities/#p1-critical for the meaning and expectations around P1 issues.



https://api.github.com/repos/apache/beam/issues/21946: [Bug]: No way to read or write to file when running Beam in Flink
https://api.github.com/repos/apache/beam/issues/21941: [Bug]: No output timestamp incorrectly handled in Dataflow runner
https://api.github.com/repos/apache/beam/issues/21935: [Bug]: Reject illformed GBK Coders
https://api.github.com/repos/apache/beam/issues/21897: [Feature Request]: Flink runner savepoint backward compatibility 
https://api.github.com/repos/apache/beam/issues/21893: [Bug]: BigQuery Storage Write API implementation does not support table partitioning
https://api.github.com/repos/apache/beam/issues/21794: Dataflow runner creates a new timer whenever the output timestamp is change
https://api.github.com/repos/apache/beam/issues/21763: [Playground Task]: Migrate from Google Analytics to Matomo Cloud
https://api.github.com/repos/apache/beam/issues/21715: Data missing when using CassandraIO.Read
https://api.github.com/repos/apache/beam/issues/21713: 404s in BigQueryIO don't get output to Failed Inserts PCollection
https://api.github.com/repos/apache/beam/issues/21711: Python Streaming job failing to drain with BigQueryIO write errors
https://api.github.com/repos/apache/beam/issues/21703: pubsublite.ReadWriteIT failing in beam_PostCommit_Java_DataflowV1 and V2
https://api.github.com/repos/apache/beam/issues/21702: SpannerWriteIT failing in beam PostCommit Java V1
https://api.github.com/repos/apache/beam/issues/21700: --dataflowServiceOptions=use_runner_v2 is broken
https://api.github.com/repos/apache/beam/issues/21699: Changing the output timestamp of a timer does not clear the previously set timer
https://api.github.com/repos/apache/beam/issues/21695: DataflowPipelineResult does not raise exception for unsuccessful states.
https://api.github.com/repos/apache/beam/issues/21694: BigQuery Storage API insert with writeResult retry and write to error table
https://api.github.com/repos/apache/beam/issues/21479: Install Python wheel and dependencies to local venv in SDK harness
https://api.github.com/repos/apache/beam/issues/21478: KafkaIO.read.withDynamicRead() doesn't pick up new TopicPartitions
https://api.github.com/repos/apache/beam/issues/21477: Add integration testing for BQ Storage API  write modes
https://api.github.com/repos/apache/beam/issues/21476: WriteToBigQuery Dynamic table destinations returns wrong tableId
https://api.github.com/repos/apache/beam/issues/21475: Beam x-lang Dataflow tests failing due to _InactiveRpcError
https://api.github.com/repos/apache/beam/issues/21473: PVR_Spark2_Streaming perma-red
https://api.github.com/repos/apache/beam/issues/21466: Simplify version override for Dev versions of the Go SDK.
https://api.github.com/repos/apache/beam/issues/21465: Kafka commit offset drop data on failure for runners that have non-checkpointing shuffle
https://api.github.com/repos/apache/beam/issues/21269: Delete orphaned files
https://api.github.com/repos/apache/beam/issues/21268: Race between member variable being accessed due to leaking uninitialized state via OutboundObserverFactory
https://api.github.com/repos/apache/beam/issues/21267: WriteToBigQuery submits a duplicate BQ load job if a 503 error code is returned from googleapi
https://api.github.com/repos/apache/beam/issues/21265: apache_beam.runners.portability.fn_api_runner.translations_test.TranslationsTest.test_run_packable_combine_globally 'apache_beam.coders.coder_impl._AbstractIterable' object is not reversible
https://api.github.com/repos/apache/beam/issues/21263: (Broken Pipe induced) Bricked Dataflow Pipeline 
https://api.github.com/repos/apache/beam/issues/21262: Python AfterAny, AfterAll do not follow spec
https://api.github.com/repos/apache/beam/issues/21260: Python DirectRunner does not emit data at GC time
https://api.github.com/repos/apache/beam/issues/21259: Consumer group with random prefix
https://api.github.com/repos/apache/beam/issues/21258: Dataflow error in CombinePerKey operation
https://api.github.com/repos/apache/beam/issues/21257: Either Create or DirectRunner fails to produce all elements to the following transform
https://api.github.com/repos/apache/beam/issues/21123: Multiple jobs running on Flink session cluster reuse the persistent Python environment.
https://api.github.com/repos/apache/beam/issues/21119: Migrate to the next version of Python `requests` when released
https://api.github.com/repos/apache/beam/issues/21117: "Java IO IT Tests" - missing data in grafana
https://api.github.com/repos/apache/beam/issues/21115: JdbcIO date conversion is sensitive to OS
https://api.github.com/repos/apache/beam/issues/21112: Dataflow SocketException (SSLException) error while trying to send message from Cloud Pub/Sub to BigQuery
https://api.github.com/repos/apache/beam/issues/21111: Java creates an incorrect pipeline proto when core-construction-java jar is not in the CLASSPATH
https://api.github.com/repos/apache/beam/issues/21110: codecov/patch has poor behavior
https://api.github.com/repos/apache/beam/issues/21109: SDF BoundedSource seems to execute significantly slower than 'normal' BoundedSource
https://api.github.com/repos/apache/beam/issues/21108: java.io.InvalidClassException With Flink Kafka
https://api.github.com/repos/apache/beam/issues/20979: Portable runners should be able to issue checkpoints to Splittable DoFn
https://api.github.com/repos/apache/beam/issues/20978: PubsubIO.readAvroGenericRecord creates SchemaCoder that fails to decode some Avro logical types
https://api.github.com/repos/apache/beam/issues/20973: Python Beam SDK Harness hangs when installing pip packages
https://api.github.com/repos/apache/beam/issues/20818: XmlIO.Read does not handle XML encoding per spec
https://api.github.com/repos/apache/beam/issues/20814: JmsIO is not acknowledging messages correctly
https://api.github.com/repos/apache/beam/issues/20813: No trigger early repeatedly for session windows
https://api.github.com/repos/apache/beam/issues/20812: Cross-language consistency (RequiresStableInputs) is quietly broken (at least on portable flink runner)
https://api.github.com/repos/apache/beam/issues/20692: Timer with dataflow runner can be set multiple times (dataflow runner)
https://api.github.com/repos/apache/beam/issues/20691: Beam metrics should be displayed in Flink UI "Metrics" tab
https://api.github.com/repos/apache/beam/issues/20689: Kafka commitOffsetsInFinalize OOM on Flink
https://api.github.com/repos/apache/beam/issues/20532: Support for coder argument in WriteToBigQuery
https://api.github.com/repos/apache/beam/issues/20531: FileBasedSink: allow setting temp directory provider per dynamic destination
https://api.github.com/repos/apache/beam/issues/20530: Make non-portable Splittable DoFn the only option when executing Java "Read" transforms
https://api.github.com/repos/apache/beam/issues/20529: SpannerIO tests don't actually assert anything.
https://api.github.com/repos/apache/beam/issues/20528: python CombineGlobally().with_fanout() cause duplicate combine results for sliding windows
https://api.github.com/repos/apache/beam/issues/20333: beam_PerformanceTests_Kafka_IO failing due to " provided port is already allocated"
https://api.github.com/repos/apache/beam/issues/20332: FileIO writeDynamic with AvroIO.sink not writing all data
https://api.github.com/repos/apache/beam/issues/20330: Remove insecure ssl options from MongoDBIO
https://api.github.com/repos/apache/beam/issues/20109: SortValues should fail if SecondaryKey coder is not deterministic
https://api.github.com/repos/apache/beam/issues/20108: Python direct runner doesn't emit empty pane when it should
https://api.github.com/repos/apache/beam/issues/20009: Environment-sensitive provisioning for Dataflow
https://api.github.com/repos/apache/beam/issues/19971: [SQL] Some Hive tests throw NullPointerException, but get marked as passing (Direct Runner)
https://api.github.com/repos/apache/beam/issues/19817: datetime and decimal should be logical types
https://api.github.com/repos/apache/beam/issues/19815: Add support for remaining data types in python RowCoder 
https://api.github.com/repos/apache/beam/issues/19813: PubsubIO returns empty message bodies for all messages read
https://api.github.com/repos/apache/beam/issues/19556: User reports protobuf ClassChangeError running against 2.6.0 or above
https://api.github.com/repos/apache/beam/issues/19369: KafkaIO doesn't commit offsets while being used as bounded source
https://api.github.com/repos/apache/beam/issues/17950: [Bug]: Java Precommit permared