You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Dongjoon Hyun (Jira)" <ji...@apache.org> on 2020/03/16 22:52:08 UTC

[jira] [Updated] (SPARK-30244) Emit pre/post events for "Partition" methods in ExternalCatalogWithListener

     [ https://issues.apache.org/jira/browse/SPARK-30244?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Dongjoon Hyun updated SPARK-30244:
----------------------------------
    Affects Version/s:     (was: 3.0.0)
                       3.1.0

> Emit pre/post events for "Partition" methods in ExternalCatalogWithListener
> ---------------------------------------------------------------------------
>
>                 Key: SPARK-30244
>                 URL: https://issues.apache.org/jira/browse/SPARK-30244
>             Project: Spark
>          Issue Type: Improvement
>          Components: SQL
>    Affects Versions: 3.1.0
>            Reporter: Jackson Westeen
>            Priority: Major
>
> Many of the methods in `ExternalCatalogWithListener` support emitting pre/post events to the ListenerBus when delegating calls to the underlying ExternalCatalog implementation. 
> Is there a reason the "Partition" series of methods don't support this feature?
> [https://github.com/apache/spark/blob/v2.4.4/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalogWithListener.scala#L193]
> We're looking to hook into post "ALTER TABLE ADD PARTITION" calls, so we can publish a message to Kafka for example, whenever new partitions are added to the metastore.
> It appears there's already event hooks for createDatabase / createTable.. but not createPartition. Would you consider adding hooks here? Or is there a better way to hook into to successful "ALTER TABLE ADD PARTITION" calls and subscribe to changes in partition status?
> Thank you!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org