You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@flink.apache.org by Aljoscha Krettek <al...@apache.org> on 2017/12/04 09:57:19 UTC

[CANCEL] [VOTE] Release 1.4.0, release candidate #2

Canceling this vote in favour of a new RC that will come once the discovered blocker issues are resolved.

> On 4. Dec 2017, at 10:17, Aljoscha Krettek <al...@apache.org> wrote:
> 
> Neat! I wil try and make sure that we get the fix in ASAP.
> 
>> On 3. Dec 2017, at 12:34, Eron Wright <er...@gmail.com> wrote:
>> 
>> Update: PR for "FLINK-8174 - Mesos RM unable to accept offers for
>> unreserved resources" is ready for review.  We were able to solve (3),
>> supporting any combination of reserved and unreserved resources.
>> 
>> On Fri, Dec 1, 2017 at 2:24 PM, Eron Wright <er...@gmail.com> wrote:
>> 
>>> There are three levels of support we could land on.
>>> 1. Flink works with unreserved resources (revert FLINK-7294).
>>> 2. Flink works with unreserved resources, and correctly ignores reserved
>>> resources (revert FLINK-7294 and mitigate Fenzo bug).
>>> 3. Flink works with unreserved resources and reserved resources.
>>> 
>>> 3 is a moon shot.  Striving for 2.  Fallback on 1.
>>> 
>>> 
>>> 
>>> On Fri, Dec 1, 2017 at 2:10 PM, Aljoscha Krettek <al...@apache.org>
>>> wrote:
>>> 
>>>> Thanks for the update!
>>>> 
>>>> Just to be clear, you're proposing going forward with the "simple fix" of
>>>> reverting FLINK-7294?
>>>> 
>>>>> On 1. Dec 2017, at 18:39, Eron Wright <er...@gmail.com> wrote:
>>>>> 
>>>>> Update on reported Mesos issue (FLINK-8174):
>>>>> 
>>>>> TLDR; a PR will be ready within 24 hours that will undo reservation
>>>> support.
>>>>> 
>>>>> A couple of months ago, a fix (FLINK-7294) was merged related to how
>>>> Flink
>>>>> accepts Mesos resource offers.  The intention was to allow Flink to make
>>>>> use of so-called +reserved+ resources, a Mesos feature which makes it
>>>>> possible to reserve hosts for use by a specific framework/role.  The fix
>>>>> inadvertently regressed the ability to use +unreserved+ resources.
>>>> This is
>>>>> a serious regression because unreserved resources are the common case.
>>>>> 
>>>>> The simple solution is to revert the earlier fix, deferring support for
>>>>> reservations to another release.   We are spending some time to find a
>>>> fix
>>>>> that works for all scenarios, but seems unlikely at this time.   I am
>>>>> reaching out to the original contributor to get their feedback.
>>>>> 
>>>>> In the course of the investigation, a related flaw was discovered in
>>>> Fenzo
>>>>> that causes Flink to misinterpret offers that contain a mix of reserved
>>>> and
>>>>> unreserved resources.   I believe that a small fix is possible purely
>>>>> within Flink; an update to Fenzo does not appear necessary.
>>>>> 
>>>>> Going forward, we will contribute an improved integration test suite
>>>> with
>>>>> which to test Flink under diverse Mesos conditions (e.g. reservations).
>>>>> 
>>>>> Thanks,
>>>>> Eron
>>>>> 
>>>>> On Thu, Nov 30, 2017 at 9:47 PM, Tzu-Li (Gordon) Tai <
>>>> tzulitai@apache.org>
>>>>> wrote:
>>>>> 
>>>>>> Hi,
>>>>>> 
>>>>>> I’ve noticed a behavioral regression in the Kafka producer, that should
>>>>>> also be considered a blocker: https://issues.apache.org/
>>>>>> jira/browse/FLINK-8181
>>>>>> There’s already a PR for the issue here: https://github.com/
>>>>>> apache/flink/pull/5108
>>>>>> 
>>>>>> Best,
>>>>>> Gordon
>>>>>> 
>>>>>> On 30 November 2017 at 5:27:22 PM, Fabian Hueske (fhueske@gmail.com)
>>>>>> wrote:
>>>>>> 
>>>>>> I've created a JIRA issue for the the Hadoop 2.9.0 build problem [1].
>>>>>> 
>>>>>> Best, Fabian
>>>>>> 
>>>>>> [1] https://issues.apache.org/jira/browse/FLINK-8177
>>>>>> 
>>>>>> 2017-11-30 4:35 GMT+01:00 Eron Wright <er...@gmail.com>:
>>>>>> 
>>>>>>> Unfortunately we've identified a blocker bug for Flink on Mesos -
>>>>>>> FLINK-8174. We'll have a patch ready on Thursday.
>>>>>>> 
>>>>>>> Thanks,
>>>>>>> Eron
>>>>>>> 
>>>>>>> On Wed, Nov 29, 2017 at 3:40 PM, Eron Wright <er...@gmail.com>
>>>>>> wrote:
>>>>>>> 
>>>>>>>> On Dell EMC side, we're testing the RC2 on DCOS 1.10.0. Seeing a
>>>>>>>> potential issue with offer acceptance and we'll update the thread
>>>> with
>>>>>> a
>>>>>>> +1
>>>>>>>> or with a more concrete issue within 24 hours.
>>>>>>>> 
>>>>>>>> Thanks,
>>>>>>>> Eron
>>>>>>>> 
>>>>>>>> On Wed, Nov 29, 2017 at 6:54 AM, Chesnay Schepler <
>>>> chesnay@apache.org>
>>>>>>>> wrote:
>>>>>>>> 
>>>>>>>>> I don't think anyone has taken a look yet, nor was there a
>>>> discussion
>>>>>> as
>>>>>>>>> to postponing it.
>>>>>>>>> 
>>>>>>>>> It just slipped through the cracks i guess...
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> On 29.11.2017 15:47, Gyula Fóra wrote:
>>>>>>>>> 
>>>>>>>>>> Hi guys,
>>>>>>>>>> I ran into this again while playing with savepoint/restore
>>>>>> parallelism:
>>>>>>>>>> 
>>>>>>>>>> https://issues.apache.org/jira/browse/FLINK-7595
>>>>>>>>>> https://github.com/apache/flink/pull/4651
>>>>>>>>>> 
>>>>>>>>>> Anyone has some idea about the status of this PR or were we
>>>> planning
>>>>>> to
>>>>>>>>>> postpone this to 1.5?
>>>>>>>>>> 
>>>>>>>>>> Thanks,
>>>>>>>>>> Gyula
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> Fabian Hueske <fh...@gmail.com> ezt írta (időpont: 2017. nov.
>>>> 29.,
>>>>>>>>>> Sze,
>>>>>>>>>> 13:10):
>>>>>>>>>> 
>>>>>>>>>> OK, the situation is the following:
>>>>>>>>>>> 
>>>>>>>>>>> The test class (org.apache.flink.yarn.UtilsTest) implements a
>>>>>> Hadoop
>>>>>>>>>>> interface (Container) that was extended in Hadoop 2.9.0 by a
>>>> getter
>>>>>>> and
>>>>>>>>>>> setter.
>>>>>>>>>>> By adding the methods, we can compile Flink for Hadoop 2.9.0.
>>>>>> However,
>>>>>>>>>>> the
>>>>>>>>>>> getter/setter add a dependency on a class that was also added in
>>>>>>> Hadoop
>>>>>>>>>>> 2.9.0.
>>>>>>>>>>> Therefore, the implementation is not backwards compatible with
>>>>>> Hadoop
>>>>>>>>>>> versions < 2.9.0.
>>>>>>>>>>> 
>>>>>>>>>>> Not sure how we can fix the problem. We would need two version of
>>>>>> the
>>>>>>>>>>> class
>>>>>>>>>>> that are chosen based on the Hadoop version. Do we have something
>>>>>> like
>>>>>>>>>>> that
>>>>>>>>>>> somewhere else?
>>>>>>>>>>> 
>>>>>>>>>>> Since this is only a problem in a test class, Flink 1.4.0 might
>>>>>> still
>>>>>>>>>>> work
>>>>>>>>>>> very well with Hadoop 2.9.0.
>>>>>>>>>>> However, this has not been tested AFAIK.
>>>>>>>>>>> 
>>>>>>>>>>> Cheers, Fabian
>>>>>>>>>>> 
>>>>>>>>>>> 2017-11-29 12:47 GMT+01:00 Fabian Hueske <fh...@gmail.com>:
>>>>>>>>>>> 
>>>>>>>>>>> I just tried to build the release-1.4 branch for Hadoop 2.9.0
>>>>>>> (released
>>>>>>>>>>>> a
>>>>>>>>>>>> few days ago) and got a compilation failure in a test class.
>>>>>>>>>>>> 
>>>>>>>>>>>> Right now, I'm assessing how much we need to fix to support
>>>> Hadoop
>>>>>>>>>>>> 2.9.0.
>>>>>>>>>>>> I'll report later.
>>>>>>>>>>>> 
>>>>>>>>>>>> Best, Fabian
>>>>>>>>>>>> 
>>>>>>>>>>>> 2017-11-29 11:16 GMT+01:00 Aljoscha Krettek <aljoscha@apache.org
>>>>> :
>>>>>>>>>>>> 
>>>>>>>>>>>> Agreed, this is a regression compared to the previous
>>>>>> functionality.
>>>>>>> I
>>>>>>>>>>>>> updated the issue to "Blocker".
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On 29. Nov 2017, at 10:01, Gyula Fóra <gy...@gmail.com>
>>>>>> wrote:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Hi all,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I have found the following issue:
>>>>>>>>>>>>>> https://issues.apache.org/jira/browse/FLINK-8165
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I would say this is a blocker (I personally pass the
>>>>>> ParameterTool
>>>>>>>>>>>>>> all
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> over
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> the place in my production apps), but a pretty trivial issue to
>>>>>>> fix,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> we
>>>>>>>>>>> 
>>>>>>>>>>>> can
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> wait a little to find other potential problems.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I can submit a fix in a little bit.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>>> Gyula
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Tzu-Li (Gordon) Tai <tz...@apache.org> ezt írta (időpont:
>>>>>> 2017.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> nov.
>>>>>>>>>>> 
>>>>>>>>>>>> 29., Sze, 9:23):
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> +1
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Verified:
>>>>>>>>>>>>>>> - No missing release Maven artifacts
>>>>>>>>>>>>>>> - Staged Apache source & binary convenience releases looks
>>>> good
>>>>>>>>>>>>>>> - NOTICE / LICENSE is correct, README is sane
>>>>>>>>>>>>>>> - Built from source (macOS, Scala 2.11, Hadoop-free & Hadoop
>>>>>> 2.8)
>>>>>>>>>>>>>>> - Cluster testing on AWS EMR (see release-testing-doc for
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> configuration
>>>>>>>>>>> 
>>>>>>>>>>>> details)
>>>>>>>>>>>>>>> - Tested Kinesis / Elasticsearch connector (no dependency
>>>>>> clashes
>>>>>>> on
>>>>>>>>>>>>>>> cluster execution, works locally in IDE)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Thanks a lot for managing the release Aljoscha!
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>>>> Gordon
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On 28 November 2017 at 8:32:42 PM, Stefan Richter (
>>>>>>>>>>>>>>> s.richter@data-artisans.com) wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> +1 (non-binding)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I tested Flink in a cluster setup on Google Cloud,
>>>> YARN-per-job,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> checked
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> that for all backends that HA, recovery, at-least-once,
>>>>>> end-to-end
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> exactly
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> once (with Kafka11 Producer), savepoints, externalized
>>>>>> checkpoints,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> and
>>>>>>>>>>> 
>>>>>>>>>>>> rescaling work correctly.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Am 28.11.2017 um 11:47 schrieb Aljoscha Krettek <
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> aljoscha@apache.org
>>>>>>>>>>> 
>>>>>>>>>>>> :
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> +1
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Verified:
>>>>>>>>>>>>>>>> - NOTICE and LICENSE are correct
>>>>>>>>>>>>>>>> - source doesn't contain binaries
>>>>>>>>>>>>>>>> - verified signatures
>>>>>>>>>>>>>>>> - verified hashes
>>>>>>>>>>>>>>>> - cluster testing on AWS and Cloudera VM (with Kerberos) (see
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> release-testing doc)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> On 28. Nov 2017, at 11:20, Aljoscha Krettek <
>>>>>> aljoscha@apache.org
>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Phew, thanks for the update!
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On 28. Nov 2017, at 11:19, Gyula Fóra <gy...@apache.org>
>>>>>>> wrote:
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Ok seems like I had to remove the snappy jar as it was
>>>>>>> corrupted
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> (makes
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> total sense) :P
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Gyula Fóra <gy...@apache.org> ezt írta (időpont: 2017.
>>>> nov.
>>>>>>>>>>>>>>>>>> 28.,
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> K,
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 11:13):
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Hi Aljoscha,
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Thanks for the release candidate. I am having a hard time
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> building
>>>>>>>>>>> 
>>>>>>>>>>>> the rc,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> I seem to get this error no matter what I do:
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> [ERROR] Failed to execute goal
>>>>>>>>>>>>>>>>>>> org.apache.maven.plugins:maven-shade-plugin:2.4.1:shade
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> (shade-hadoop) on
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> project flink-shaded-hadoop2-uber: Error creating shaded jar:
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> invalid
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> LOC
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> header (bad signature) -> [Help 1]
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> (Apache Maven 3.3.9)
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Any idea what I am missing?
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>> Gyula
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Aljoscha Krettek <al...@apache.org> ezt írta (időpont:
>>>>>>> 2017.
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> nov.
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 27.,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> H, 19:35):
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Hi everyone,
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Please review and vote on release candidate #2 for the
>>>>>>> version
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> 1.4.0, as
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> follows:
>>>>>>>>>>>>>>>>>>>> [ ] +1, Approve the release
>>>>>>>>>>>>>>>>>>>> [ ] -1, Do not approve the release (please provide
>>>> specific
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> comments)
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> The complete staging area is available for your review,
>>>>>> which
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> includes:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> * JIRA release notes [1],
>>>>>>>>>>>>>>>>>>>> * the official Apache source release and binary
>>>> convenience
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> releases
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> to
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> be deployed to dist.apache.org[2], which are signed with the
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> key
>>>>>>>>>>> 
>>>>>>>>>>>> with
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> fingerprint F2A67A8047499BBB3908D17AA8F4FD97121D7293 [3],
>>>>>>>>>>>>>>>>>>>> * all artifacts to be deployed to the Maven Central
>>>>>>> Repository
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> [4],
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> * source code tag "release-1.4.0-rc1" [5],
>>>>>>>>>>>>>>>>>>>> * website pull request listing the new release [6].
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Please have a careful look at the website PR because I
>>>>>>> changed
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> some
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> wording and we're now also releasing a binary without Hadoop
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> dependencies.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Please use this document for coordinating testing efforts:
>>>> [7]
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> The only change between RC1 and this RC2 is that the
>>>> source
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> release
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> package does not include the erroneously included binary Ruby
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> dependencies
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> of the documentation anymore. Because of this I would like to
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> propose a
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> shorter voting time and close the vote around the time that
>>>> RC1
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> would have
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> closed. This would mean closing by end of Wednesday. Please
>>>> let
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> me
>>>>>>>>>>> 
>>>>>>>>>>>> know if
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> you disagree with this. The vote is adopted by majority
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> approval,
>>>>>>>>>>> 
>>>>>>>>>>>> with at
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> least 3 PMC affirmative votes.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>> Your friendly Release Manager
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> [1]
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?
>>>>>> proje
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> ctId=12315522&version=12340533
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> [2] http://people.apache.org/~aljoscha/flink-1.4.0-rc2/
>>>>>>>>>>>>>>>>>>>> [3] https://dist.apache.org/repos/
>>>> dist/release/flink/KEYS
>>>>>>>>>>>>>>>>>>>> [4]
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>>> content/repositories/orgapache
>>>>>>>>>>> flink-1140
>>>>>>>>>>> 
>>>>>>>>>>>> [5]
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> https://git-wip-us.apache.org/
>>>>>> repos/asf?p=flink.git;a=tag;h=
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> ea751b7b23b23446ed3fcdeed564bbe8bf4adf9c
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> [6] https://github.com/apache/flink-web/pull/95
>>>>>>>>>>>>>>>>>>>> [7]
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> https://docs.google.com/document/d/
>>>>>> 1HqYyrNoMSXwo8zBpZj7s39Uz
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> UdlFcFO8TRpHNZ_cl44/edit?usp=sharing
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Pro-tip: you can create a settings.xml file with these
>>>> contents:
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> <settings>
>>>>>>>>>>>>>>>>>>>> <activeProfiles>
>>>>>>>>>>>>>>>>>>>> <activeProfile>flink-1.4.0</activeProfile>
>>>>>>>>>>>>>>>>>>>> </activeProfiles>
>>>>>>>>>>>>>>>>>>>> <profiles>
>>>>>>>>>>>>>>>>>>>> <profile>
>>>>>>>>>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>>>>>>>>>> <repositories>
>>>>>>>>>>>>>>>>>>>> <repository>
>>>>>>>>>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>>>>>>>>>> <url>
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>>> content/repositories/orgapache
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> flink-1140/
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> </url>
>>>>>>>>>>>>>>>>>>>> </repository>
>>>>>>>>>>>>>>>>>>>> <repository>
>>>>>>>>>>>>>>>>>>>> <id>archetype</id>
>>>>>>>>>>>>>>>>>>>> <url>
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>>> content/repositories/orgapache
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> flink-1140/
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> </url>
>>>>>>>>>>>>>>>>>>>> </repository>
>>>>>>>>>>>>>>>>>>>> </repositories>
>>>>>>>>>>>>>>>>>>>> </profile>
>>>>>>>>>>>>>>>>>>>> </profiles>
>>>>>>>>>>>>>>>>>>>> </settings>
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> And reference that in you maven commands via --settings
>>>>>>>>>>>>>>>>>>>> path/to/settings.xml. This is useful for creating a
>>>>>>> quickstart
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> based
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> on the
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> staged release and for building against the staged jars.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>>>> 
>>> 
>