You are viewing a plain text version of this content. The canonical link for it is here.
Posted to github@beam.apache.org by GitBox <gi...@apache.org> on 2020/07/29 07:24:22 UTC

[GitHub] [beam] ibzib edited a comment on pull request #12397: [BEAM-7632] Add Python quickstart instructions for Flink and Spark.

ibzib edited a comment on pull request #12397:
URL: https://github.com/apache/beam/pull/12397#issuecomment-665351269


   > For an existing cluster, isn't it as simple as passing the flink master URL? If so, I think that's worth stating.
   
   For a local cluster, yes (we can just use loopback). Or at least I thought so. I tried this on my local Flink 1.10 cluster:
   
   ```sh
   python -m apache_beam.examples.wordcount --input /path/to/inputfile \
                                            --output /path/to/write/counts \
                                            --runner FlinkRunner \
                                            --environment_type LOOPBACK \
                                            --flink_master localhost:8081 \
                                            --flink_version 1.10
   ```
   
   But I got an error:
   
   `ERROR:root:org.apache.flink.runtime.rest.util.RestClientException: [Failed to deserialize JobGraph.]`
   
   This [Stack Overflow post](https://stackoverflow.com/questions/58358308/flink-submitting-job-failed-to-deserialize-jobgraph) suggests a Flink version incompatibility might be the cause, but both the job server jar and the Flink cluster are Flink 1.10, so I'm not sure what's going on.
   


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org