You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Jarus Local <ja...@gmail.com> on 2023/03/13 22:19:04 UTC
Extending Spark Connector versus Providing utility library
Hi team,
Had a design question around wether it’s a good idea to write wrappers over
all existing spark connectors for adding some functionality/improving
usability in terms of options passed to the connector. In contrast to
providing utility libraries that takes parameters and calls the underlying
connectors as it is.
Example:
Option 1:
Extends snowflake-connector to add some options and use it as:
>> spark.read.format(“my-snowflake”).option(“myoption”, “”)
Option 2:
Write a utility
>> util.readWithMyOption()
Also what kind of logic can be put inside the DataSourceV2 connector? Is it
a good idea to manipulate/change the incoming dataset inside the connector
Thanks.