You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "ASF GitHub Bot (JIRA)" <ji...@apache.org> on 2018/01/05 17:08:00 UTC

[jira] [Commented] (FLINK-8356) JDBCAppendTableSink does not work for Hbase Phoenix Driver

    [ https://issues.apache.org/jira/browse/FLINK-8356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16313457#comment-16313457 ] 

ASF GitHub Bot commented on FLINK-8356:
---------------------------------------

GitHub user paulzwu opened a pull request:

    https://github.com/apache/flink/pull/5247

    [FLINK-8356] Need to add the commit in flush() in JDBCOutputFormat

    *Thank you very much for contributing to Apache Flink - we are happy that you want to help us improve Flink. To help the community review your contribution in the best possible way, please go through the checklist below, which will get the contribution into a shape in which it can be best reviewed.*
    
    *Please understand that we do not do this to make contributions to Flink a hassle. In order to uphold a high standard of quality for code contributions, while at the same time managing a large number of contributions, we need contributors to prepare the contributions well, and give reviewers enough contextual information for the review. Please also understand that contributions that do not follow this guide will take longer to review and thus typically be picked up with lower priority by the community.*
    
    ## Contribution Checklist
    
      - Make sure that the pull request corresponds to a [JIRA issue](https://issues.apache.org/jira/projects/FLINK/issues). Exceptions are made for typos in JavaDoc or documentation files, which need no JIRA issue.
      
      - Name the pull request in the form "[FLINK-XXXX] [component] Title of the pull request", where *FLINK-XXXX* should be replaced by the actual issue number. Skip *component* if you are unsure about which is the best component.
      Typo fixes that have no associated JIRA issue should be named following this pattern: `[hotfix] [docs] Fix typo in event time introduction` or `[hotfix] [javadocs] Expand JavaDoc for PuncuatedWatermarkGenerator`.
    
      - Fill out the template below to describe the changes contributed by the pull request. That will give reviewers the context they need to do the review.
      
      - Make sure that the change passes the automated tests, i.e., `mvn clean verify` passes. You can set up Travis CI to do that following [this guide](http://flink.apache.org/contribute-code.html#best-practices).
    
      - Each pull request should address only one issue, not mix up code from multiple issues.
      
      - Each commit in the pull request has a meaningful commit message (including the JIRA id)
    
      - Once all items of the checklist are addressed, remove the above text and this checklist, leaving only the filled out template below.
    
    
    **(The sections below can be removed for hotfixes of typos)**
    
    ## What is the purpose of the change
    
    *(For example: This pull request makes task deployment go through the blob server, rather than through RPC. That way we avoid re-transferring them on each deployment (during recovery).)*
    
    
    ## Brief change log
    
    *(for example:)*
      - *The TaskInfo is stored in the blob store on job creation time as a persistent artifact*
      - *Deployments RPC transmits only the blob storage reference*
      - *TaskManagers retrieve the TaskInfo from the blob cache*
    
    
    ## Verifying this change
    
    *(Please pick either of the following options)*
    
    This change is a trivial rework / code cleanup without any test coverage.
    
    *(or)*
    
    This change is already covered by existing tests, such as *(please describe tests)*.
    
    *(or)*
    
    This change added tests and can be verified as follows:
    
    *(example:)*
      - *Added integration tests for end-to-end deployment with large payloads (100MB)*
      - *Extended integration test for recovery after master (JobManager) failure*
      - *Added test that validates that TaskInfo is transferred only once across recoveries*
      - *Manually verified the change by running a 4 node cluser with 2 JobManagers and 4 TaskManagers, a stateful streaming program, and killing one JobManager and two TaskManagers during the execution, verifying that recovery happens correctly.*
    
    ## Does this pull request potentially affect one of the following parts:
    
      - Dependencies (does it add or upgrade a dependency): (yes / no)
      - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (yes / no)
      - The serializers: (yes / no / don't know)
      - The runtime per-record code paths (performance sensitive): (yes / no / don't know)
      - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes / no / don't know)
      - The S3 file system connector: (yes / no / don't know)
    
    ## Documentation
    
      - Does this pull request introduce a new feature? (yes / no)
      - If yes, how is the feature documented? (not applicable / docs / JavaDocs / not documented)


You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/apache/flink master

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/flink/pull/5247.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #5247
    
----
commit d98ba08a7a73dc93c839c77d11245aba70869ab5
Author: Aljoscha Krettek <al...@...>
Date:   2017-11-10T09:54:16Z

    [FLINK-6163] Document per-window state in ProcessWindowFunction

commit 1cd3ba3f2af454bc33f2c880163c01dddd4d1738
Author: Piotr Nowojski <pi...@...>
Date:   2017-11-10T14:15:11Z

    [hotfix][docs] Change mailing list link in quickstart to flink-user
    
    Previously it was pointing to flink-dev

commit 07af8c125bba88912dd39664532b1ec8ee035439
Author: Aljoscha Krettek <al...@...>
Date:   2017-11-10T15:28:46Z

    [FLINK-7702] Remove Javadoc aggregation for Scala code
    
    genjavadoc generated some Java code that was making Javadoc fail.

commit 2929eda7e244c1cd7250645702a808af967fae84
Author: Aljoscha Krettek <al...@...>
Date:   2017-11-10T17:13:26Z

    [FLINK-7702] Add maven-bundle-plugin to root pom
    
    Before, we had it in places that require it. This doesn't work when
    running mvn javadoc:aggregate because this will only run for the root
    pom and can then not find the "bundle" dependencies.

commit ad8ef6d01b23c74fcfb1f39a151dd00bfbd78ca3
Author: Stefan Richter <s....@...>
Date:   2017-11-13T10:50:07Z

    [FLINK-8040] [tests] Fix test instability in ResourceGuardTest
    
    This closes #5004.

commit 3b333b289cbbe43f722edef2c36c370ff4550128
Author: Kent Murra <ke...@...>
Date:   2017-09-27T20:48:55Z

    [FLINK-7657] [table] Add time types FilterableTableSource push down
    
    This closes #4746.

commit 32e5194d95742dc0115a2b4d8c34e3f025feebd0
Author: twalthr <tw...@...>
Date:   2017-11-13T13:09:45Z

    [FLINK-7657] [table] Add all basic types to RexProgramExtractor

commit 0e5fb0b78cd0a3ccb144071a47579eb6c3d0570a
Author: Nico Kruber <ni...@...>
Date:   2017-11-06T18:53:37Z

    [FLINK-7973] Fix shading and relocating Hadoop for the S3 filesystems
    
    - do not shade everything, especially not JDK classes!
    -> instead define include patterns explicitly
    - do not shade core Flink classes (only those imported from flink-hadoop-fs)
    - hack around Hadoop loading (unshaded/non-relocated) classes based on names in
      the core-default.xml by overwriting the Configuration class (we may need to
      extend this for the mapred-default.xml and hdfs-defaults.xml):
    -> provide a core-default-shaded.xml file with shaded class names and copy and
      adapt the Configuration class of the respective Hadoop version to load this
      file instead of core-default.xml.
    
    Add checkstyle suppression pattern for the Hadoop Configuration classes
    
    Also fix the (integration) tests not working because they tried to load the
    relocated classes which are apparently not available there
    
    Remove minimizeJar from shading of flink-s3-fs-presto because this was
    causing "java.lang.ClassNotFoundException:
    org.apache.flink.fs.s3presto.shaded.org.apache.commons.logging.impl.LogFactoryImpl"
    since these classes are not statically imported and thus removed when
    minimizing.
    
    Fix s3-fs-presto not shading org.HdrHistogram
    
    Fix log4j being relocated in the S3 fs implementations
    
    Add shading checks to travis

commit e9e7c3372189db7e933ff59114b9ec6245838eda
Author: Aljoscha Krettek <al...@...>
Date:   2017-11-08T16:49:17Z

    [FLINK-7973] Add shaded S3 FileSystem end-to-end tests

commit 8fa6cf947d0b00095a0e35b3d4545239e340103e
Author: Aljoscha Krettek <al...@...>
Date:   2017-11-08T17:57:22Z

    [hotfix] Make end-to-end test scripts more robust
    
    This uses traps to ensure that we properly do cleanups, remove config
    values and shutdown things.

commit 613fc445f0a2669672c90f0718c5fcfe6822b15c
Author: Nico Kruber <ni...@...>
Date:   2017-11-10T12:16:40Z

    [hotfix] fix presto end-to-end test not cleaning up

commit ab2b9f8f9629cf33135b923fd81ee28a4ea8b8bd
Author: Nico Kruber <ni...@...>
Date:   2017-11-10T13:04:29Z

    [hotfix] ignore a warning from the error check of the S3 e2e tests

commit d139e634087aaba7db43429d803d12203b7bc18a
Author: Nico Kruber <ni...@...>
Date:   2017-11-10T15:42:10Z

    [hotfix] let end-to-end tests check for empty .out files again

commit 2906698b4a87f21c6fd099cf8a028f68fc311b1f
Author: Stefan Richter <s....@...>
Date:   2017-11-13T13:31:45Z

    [FLINK-8053] [checkpoints] Default to asynchronous snapshots for FsStateBackend and MemoryStateBackend.
    
    This closes #5005.

commit b98a4aa8b8ba1881f938125d0765620a4289a3c7
Author: Alejandro Alcalde <al...@...>
Date:   2017-11-07T12:09:46Z

    [FLINK-8006] [Startup Shell Scripts] - Fixing $pid
    
    This closes #4968.

commit 119006752e190334fbf4f90fa53f6dfa9374e81b
Author: zentol <ch...@...>
Date:   2017-11-07T16:13:57Z

    [FLINK-8011][dist] Set flink-python to provided
    
    This closes #4973.

commit 3f80f8ef334959392ac4a22b064fb6c8b5378462
Author: gyao <ga...@...>
Date:   2017-11-10T13:54:32Z

    [hotfix][docs] Fix broken link to FLINK-7811
    
    This closes #4995.

commit 49c8686b81eb7cb13912b4fbb6f468840f1cc7ad
Author: yew1eb <ye...@...>
Date:   2017-11-11T12:45:42Z

    [hotfix][docs][javadocs] Remove double "of"
    
    This closes #4999.

commit c5371ee6430a36f55001048248b1b2adc901a1d1
Author: yew1eb <ye...@...>
Date:   2017-11-11T13:08:21Z

    [hotfix][docs] Fix typos in deployment AWS documentation
    
    This closes #5000.

commit fcbfbb9126d458feb17527b73be1eb38a2e640f7
Author: Michael Fong <mc...@...>
Date:   2017-11-13T02:17:02Z

    [FLINK-4500][docs] Update cassandra documentation regarding data loss
    
    As of FLINK-4500 the Cassandra connector will wait for pending updates to finish upon checkpoint.
    
    This closes #5002.

commit 16107cf67416391ccf1493bda747a108bbbb3a20
Author: Piotr Nowojski <pi...@...>
Date:   2017-11-13T15:51:18Z

    [FLINK-7845][runtime] Make NettyMessage public
    
    This a walkaround strange javaassist bug. The issue should go away
    once we upgrade netty dependency.
    
    Please check the ticket for more information.
    
    This closes #5007.

commit ed90379e3b96b03e24414169f92e6cb8371a6250
Author: zentol <ch...@...>
Date:   2017-11-14T09:58:43Z

    [FLINK-8056][dist] Use 'web.port' instead of 'jobmanager.web.port'
    
    This closes #5010.

commit 408e186995f4e6bacc0904b8dd6b6f9b7e28e60e
Author: Vetriselvan1187 <ve...@...>
Date:   2017-11-06T17:11:41Z

    [FLINK-7998][examples] Fix TPCHQuery3 examples
    
    This closes #4959.

commit e16472953cef75743791c91f17e8114f2f045054
Author: zentol <ch...@...>
Date:   2017-11-14T13:47:00Z

    [FLINK-8071][build] Bump shade-plugin asm version to 5.1
    
    This closes #5014.

commit fe98cb77c4fcf2ff0e2840c2254fb8b517274917
Author: zentol <ch...@...>
Date:   2017-11-08T10:58:38Z

    [FLINK-7419][build][avro] Shade jackson dependency in flink-dist
    
    This closes #4981.

commit 5f00294af4b5527d4e320c2adc66fe9c6c2fd5b6
Author: Nico Kruber <ni...@...>
Date:   2017-11-14T13:36:22Z

    [FLINK-7973] disable JNI bridge for relocated hadoop classes in s3-fs-*

commit c32fcbc2673233bb1c16a8b96fb2b7e88a55fdca
Author: Jark Wu <ja...@...>
Date:   2017-08-15T11:24:57Z

    [FLINK-7451] [table] Support non-ascii character literals in Table API and SQL
    
    This closes #4544.

commit 9d34715747040352f6c0ca13c052763cd8049dad
Author: twalthr <tw...@...>
Date:   2017-11-15T13:39:21Z

    [FLINK-7451] [table] Disable testing of the charset in TableTestBase

commit 11218a35dc0fdd7439142a313e6628c51cffe689
Author: twalthr <tw...@...>
Date:   2017-11-14T10:06:54Z

    [FLINK-8013] [table] Support aggregate functions with generic arrays
    
    This closes #5011.

commit 54eeccfe18ace636b77ff0226ce1f91d35e4d49d
Author: twalthr <tw...@...>
Date:   2017-09-26T10:10:33Z

    [FLINK-7678] [table] Support composite inputs for user-defined functions
    
    This closes #4726.

----


> JDBCAppendTableSink does not work for Hbase Phoenix Driver 
> -----------------------------------------------------------
>
>                 Key: FLINK-8356
>                 URL: https://issues.apache.org/jira/browse/FLINK-8356
>             Project: Flink
>          Issue Type: Bug
>          Components: Table API & SQL
>    Affects Versions: 1.4.0
>            Reporter: Paul Wu
>
> The following code runs without errors, but the data is not inserted into the HBase table. However, it does work for MySQL (see the commented out code). The Phoenix driver is from https://mvnrepository.com/artifact/org.apache.phoenix/phoenix/4.7.0-HBase-1.1
> String query = "select CURRENT_DATE SEGMENTSTARTTIME, CURRENT_DATE SEGMENTENDTIME, cast (imsi as varchar) imsi, cast(imei as varchar) imei from ts ";
>         
>         Table table = ste.sqlQuery(query);
>         JDBCAppendTableSinkBuilder jdbc = JDBCAppendTableSink.builder();
>         jdbc.setDrivername("org.apache.phoenix.jdbc.PhoenixDriver");
>         jdbc.setDBUrl("jdbc:phoenix:hosts:2181:/hbase-unsecure");
>         jdbc.setQuery("upsert INTO GEO_ANALYTICS_STREAMING_DATA (SEGMENTSTARTTIME,SEGMENTENDTIME, imsi, imei) values (?,?,?, ?)");
> //     JDBCAppendTableSinkBuilder jdbc = JDBCAppendTableSink.builder();
> //        jdbc.setDrivername("com.mysql.jdbc.Driver");
> //        jdbc.setDBUrl("jdbc:mysql://localhost/test");
> //        jdbc.setUsername("root").setPassword("");
> //        jdbc.setQuery("insert INTO GEO_ANALYTICS_STREAMING_DATA (SEGMENTSTARTTIME,SEGMENTENDTIME, imsi, imei) values (?,?,?, ?)");
> //        jdbc.setBatchSize(1);
>         jdbc.setParameterTypes(Types.SQL_DATE, Types.SQL_DATE, Types.STRING, Types.STRING);
>         JDBCAppendTableSink sink = jdbc.build();
>         table.writeToSink(sink);



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)