You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Chesnay Schepler (JIRA)" <ji...@apache.org> on 2016/03/16 17:04:33 UTC

[jira] [Updated] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package

     [ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Chesnay Schepler updated FLINK-3524:
------------------------------------
    Summary: Provide a JSONDeserialisationSchema in the kafka connector package  (was: Provide a JSONSerialisationSchema in the kafka connector package)

> Provide a JSONDeserialisationSchema in the kafka connector package
> ------------------------------------------------------------------
>
>                 Key: FLINK-3524
>                 URL: https://issues.apache.org/jira/browse/FLINK-3524
>             Project: Flink
>          Issue Type: Improvement
>          Components: Kafka Connector
>            Reporter: Robert Metzger
>            Assignee: Chesnay Schepler
>              Labels: starter
>
> (I don't want to include this into 1.0.0)
> Currently, there is no standardized way of parsing JSON data from a Kafka stream. I see a lot of users using JSON in their topics. It would make things easier for our users to provide a serializer for them.
> I suggest to use the jackson library because we have that aready as a dependency in Flink and it allows to parse from a byte[].
> I would suggest to provide the following classes:
>  - JSONDeserializationSchema()
>  - JSONDeKeyValueSerializationSchema(bool includeMetadata)
> The second variant should produce a record like this:
> {code}
> {"key": "keydata",
> "value": "valuedata",
> "metadata": {"offset": 123, "topic": "<topic>", "partition": 2 }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)