Now it fails with this error:

WARNING:root:Make sure that locally built Python SDK docker image has Python 
3.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.7_sdk:2.21.0. If 
the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:====================
 <function lift_combiners at 0x0000027A25480CA8> ====================
Traceback (most recent call last):
  File 
"C:\Users\ashish.raghav\AppData\Local\Programs\Python\Python37\lib\runpy.py", 
line 193, in _run_module_as_main
    "__main__", mod_spec)
  File 
"C:\Users\ashish.raghav\AppData\Local\Programs\Python\Python37\lib\runpy.py", 
line 85, in _run_code
    exec(code, run_globals)
  File "C:\Users\ashish.raghav\Desktop\projects\test-flink\wc_minimal.py", line 
145, in <module>
    run()
  File "C:\Users\ashish.raghav\Desktop\projects\test-flink\wc_minimal.py", line 
140, in run
    output | WriteToText(known_args.output)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\pipeline.py",
 line 524, in __exit__
    self.run().wait_until_finish()
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\pipeline.py",
 line 497, in run
    self._options).run(False)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\pipeline.py",
 line 510, in run
    return self.runner.run_pipeline(self, self._options)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\portable_runner.py",
 line 406, in run_pipeline
    job_service_handle.submit(proto_pipeline)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\portable_runner.py",
 line 107, in submit
    prepare_response.staging_session_token)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\portable_runner.py",
 line 204, in stage
    stager.stage_job_resources(resources, staging_location='')
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\stager.py",
 line 305, in stage_job_resources
    file_path, FileSystems.join(staging_location, staged_path))
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\portable_stager.py",
 line 98, in stage_artifact
    self._artifact_staging_stub.PutArtifact(artifact_request_generator())
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\grpc\_channel.py",
 line 1011, in __call__
    return _end_unary_response_blocking(state, call, False, None)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\grpc\_channel.py",
 line 729, in _end_unary_response_blocking
    raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
        status = StatusCode.UNAVAILABLE
        details = "failed to connect to all addresses"
        debug_error_string = 
"{"created":"@1590667551.301000000","description":"Failed to pick 
subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3962,"referenced_errors":[{"created":"@1590667551.301000000","description":"failed
 to connect to all 
addresses","file":"src/core/ext/filters/client_channel/lb_policy/pick_first/pick_first.cc","file_line":394,"grpc_status":14}]}"
>                                                              python -m 
> wc_minimal --input 
> C:\Users\ashish.raghav\Desktop\projects\test-flink\input.txt --output 
> C:\Users\ashish.raghav\Desktop\projects\test-flink\output>
WARNING:root:Make sure that locally built Python SDK docker image has Python 
3.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.7_sdk:2.21.0. If 
the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:====================
 <function lift_combiners at 0x00000243E7ADECA8> ====================
Traceback (most recent call last):
  File 
"C:\Users\ashish.raghav\AppData\Local\Programs\Python\Python37\lib\runpy.py", 
line 193, in _run_module_as_main
    "__main__", mod_spec)
  File 
"C:\Users\ashish.raghav\AppData\Local\Programs\Python\Python37\lib\runpy.py", 
line 85, in _run_code
    exec(code, run_globals)
  File "C:\Users\ashish.raghav\Desktop\projects\test-flink\wc_minimal.py", line 
147, in <module>
    run()
  File "C:\Users\ashish.raghav\Desktop\projects\test-flink\wc_minimal.py", line 
142, in run
    output | WriteToText(known_args.output)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\pipeline.py",
 line 524, in __exit__
    self.run().wait_until_finish()
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\pipeline.py",
 line 497, in run
    self._options).run(False)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\pipeline.py",
 line 510, in run
    return self.runner.run_pipeline(self, self._options)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\portable_runner.py",
 line 406, in run_pipeline
    job_service_handle.submit(proto_pipeline)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\portable_runner.py",
 line 107, in submit
    prepare_response.staging_session_token)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\portable_runner.py",
 line 204, in stage
    stager.stage_job_resources(resources, staging_location='')
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\stager.py",
 line 305, in stage_job_resources
    file_path, FileSystems.join(staging_location, staged_path))
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\apache_beam\runners\portability\portable_stager.py",
 line 98, in stage_artifact
    self._artifact_staging_stub.PutArtifact(artifact_request_generator())
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\grpc\_channel.py",
 line 1011, in __call__
    return _end_unary_response_blocking(state, call, False, None)
  File 
"C:\Users\ashish.raghav\Desktop\projects\test-flink\.venv\lib\site-packages\grpc\_channel.py",
 line 729, in _end_unary_response_blocking
    raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
        status = StatusCode.UNAVAILABLE
        details = "failed to connect to all addresses"
        debug_error_string = 
"{"created":"@1590667588.146000000","description":"Failed to pick 
subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3962,"referenced_errors":[{"created":"@1590667588.146000000","description":"failed
 to connect to all 
addresses","file":"src/core/ext/filters/client_channel/lb_policy/pick_first/pick_first.cc","file_line":394,"grpc_status":14}]}"



From: Ashish Raghav <ashish.rag...@corecompete.com>
Sent: 28 May 2020 17:05
To: user@beam.apache.org
Subject: RE: Issue while submitting python beam pipeline on flink - local

EXTERNAL EMAIL
Do not click links or open attachments unless you recognise the sender and know 
the content is safe. Report suspicious email to 
info...@corecompete.com<mailto:info...@corecompete.com>.

Ok. Will test with latest version.

From: Kyle Weaver <kcwea...@google.com<mailto:kcwea...@google.com>>
Sent: 28 May 2020 17:03
To: user@beam.apache.org<mailto:user@beam.apache.org>
Subject: Re: Issue while submitting python beam pipeline on flink - local

EXTERNAL EMAIL
Do not click links or open attachments unless you recognise the sender and know 
the content is safe. Report suspicious email to 
info...@corecompete.com<mailto:info...@corecompete.com>.

Oh, I see the issue. The documentation you were looking at might contain an 
outdated reference to the apachebeam repo on Docker hub. We have migrated all 
Docker images to the apache top-level repository. So instead of 
apachebeam/flink1.9_job_server, you should use apache/beam_flink1.9_job_server.

On Thu, May 28, 2020 at 7:31 AM Kyle Weaver 
<kcwea...@google.com<mailto:kcwea...@google.com>> wrote:
2.21.0 should be available now: 
https://hub.docker.com/layers/apache/beam_flink1.9_job_server/2.21.0/images/sha256-eeac6dd4571794a8f985e9967fa0c1522aa56a28b5b0a0a34490a600065f096d?context=explore

On Thu, May 28, 2020 at 7:27 AM Ashish Raghav 
<ashish.rag...@corecompete.com<mailto:ashish.rag...@corecompete.com>> wrote:
Hi Kyle,

The Latest Version available on docker hub is 2.20.0


From: Kyle Weaver <kcwea...@google.com<mailto:kcwea...@google.com>>
Sent: 28 May 2020 16:51
To: user@beam.apache.org<mailto:user@beam.apache.org>
Subject: Re: Issue while submitting python beam pipeline on flink - local

EXTERNAL EMAIL
Do not click links or open attachments unless you recognise the sender and know 
the content is safe. Report suspicious email to 
info...@corecompete.com<mailto:info...@corecompete.com>.

Hi Ashish, can you check to make sure apachebeam/flink1.9_job_server is also on 
version 2.21.0?

On Thu, May 28, 2020 at 7:13 AM Ashish Raghav 
<ashish.rag...@corecompete.com<mailto:ashish.rag...@corecompete.com>> wrote:
Hello Guys ,

I am trying to run a python beam pipeline on flink. I am trying to run 
apache_beam.examples.wordcount_minimal but with Pipelineoptions as
"--runner=PortableRunner",
"--job_endpoint=192.168.99.100:8099<http://192.168.99.100:8099>",
"--environment_type=LOOPBACK",

I have a apachebeam/flink1.9_job_server running on local container.
Whenever I submit the job , I get errors ( attached the log).When I run it with 
DirectRunner, the code runs fine though.

versions:
Apache Beam SDK=2.21.0
python=3.7.6

I have followed all instruction from the apache beam UI but I seem to have hit 
a roadblock here. Please suggest.

Please let me know if any other information is required.

Ashish Raghav | DE
Core Compete<https://corecompete.com/> | 
ashish.rag...@corecompete.com<mailto:adwait.th...@corecompete.com>
Accelerating Cloud Analytics
[cid:image006.jpg@01D63516.96A35ED0][cid:image005.png@01D63516.96A35ED0]

Reply via email to