Configurar o IAM e a rede para jobs do Dataflow avaliações
12730 avaliações
A few of the commands are outdated.
Mal W. · Revisado há almost 4 years
File "/usr/lib/python3.9/runpy.py", line 197, in _run_module_as_main return _run_code(code, main_globals, None, File "/usr/lib/python3.9/runpy.py", line 87, in _run_code exec(code, run_globals) File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/examples/wordcount.py", line 103, in <module> run() File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/examples/wordcount.py", line 98, in run output | 'Write' >> WriteToText(known_args.output) File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/pipeline.py", line 596, in __exit__ self.result = self.run() File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/pipeline.py", line 546, in run return Pipeline.from_runner_api( File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/pipeline.py", line 573, in run return self.runner.run_pipeline(self, self._options) File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/runners/dataflow/dataflow_runner.py", line 582, in run_pipeline self.dataflow_client.create_job(self.job), self) File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/utils/retry.py", line 253, in wrapper return fun(*args, **kwargs) File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/runners/dataflow/internal/apiclient.py", line 724, in create_job self.create_job_description(job) File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/runners/dataflow/internal/apiclient.py", line 822, in create_job_description job.proto.environment = Environment( File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/runners/dataflow/internal/apiclient.py", line 184, in __init__ _verify_interpreter_version_is_supported(options) File "/home/student_04_ceb375744f38/df-env/lib/python3.9/site-packages/apache_beam/runners/dataflow/internal/apiclient.py", line 1272, in _verify_interpreter_version_is_supported raise Exception(
Tristan T. · Revisado há almost 4 years
ROHIT P. · Revisado há almost 4 years
Ashish S. · Revisado há almost 4 years
Had to go off trails with --experiment use_unsupported_python_version Can't complete this lab with provided code. No time left to diagnose or correct packages. "Error syncing pod d3f2b51794d3fd272770f868250b1791 ("df-beamapp-student005a377216-02201449-5d9m-harness-5d9g_default(d3f2b51794d3fd272770f868250b1791)"), skipping: failed to "StartContainer" for "python" with CrashLoopBackOff: "back-off 10s restarting failed container=python pod=df-beamapp-student005a377216-02201449-5d9m-harness-5d9g_default(d3f2b51794d3fd272770f868250b1791)""
Jordon M. · Revisado há almost 4 years
Catarina D. · Revisado há almost 4 years
Balaram D. · Revisado há almost 4 years
Sourav S. · Revisado há almost 4 years
The lab instructions do not play well with Python 3.9, which is the default for the Cloud Shell at the moment. The Dataflow job failed twice and the lab time ran out on my third attempt.
Mohamed A. · Revisado há almost 4 years
Please note the Labs seems to be intended for Python version 3.7, but most of the labs that are currently running are using Python 3.9 Version which are not compatible and fail, So we need to give this --experiment use_unsupported_python_version to avoid failures. It would be good to have this updated in the Labs... but a good learning experience that I wanted to share.
Krishna V. · Revisado há almost 4 years
Not working the code provided
Souvik B. · Revisado há almost 4 years
This lab use python3.9 but dataflow can run only 3.6 - 3.8 . Then I can't run this lab.
Chittanutchote W. · Revisado há almost 4 years
H F. · Revisado há almost 4 years
Python version in lab does not correspond to the one needed for Dataflow SDK. Tried to launch the job using command line parameter : --experiment use_unsupported_python_version But the job later crashes repeatedly on kubelet error : "Error syncing pod d3f2b51794d3fd272770f868250b1791 ("df-beamapp-student0059c260e1-02180722-gtoz-harness-hqlc_default(d3f2b51794d3fd272770f868250b1791)"), skipping: failed to "StartContainer" for "python" with CrashLoopBackOff: "back-off 10s restarting failed container=python pod=df-beamapp-student0059c260e1-02180722-gtoz-harness-hqlc_default(d3f2b51794d3fd272770f868250b1791)""
Christophe S. · Revisado há almost 4 years
Javier M. · Revisado há almost 4 years
Javier M. · Revisado há almost 4 years
Error Default python version is 3.9.x but apache-beam[gcp] only support 3.8.x
Robin J. · Revisado há almost 4 years
Sachin K. · Revisado há almost 4 years
Md S. · Revisado há almost 4 years
The version of Python is 3.9 and I got errors that says I need 3.8 and I could not install it, so I was not able to complete the lab. Please fix this.
Jean-Sébastien P. · Revisado há almost 4 years
Alceu D. · Revisado há almost 4 years
Jagdeep S. · Revisado há almost 4 years
Pyhton 3.9 not supported. I am not able to finish the course!!!!!!
Hugo M. · Revisado há almost 4 years
Bruno P. · Revisado há almost 4 years
The wordcount job wouldn't run to completion. First issue was the python version mismatch, then it didn't run. I wasn't able to complete. I'll try again.
tim t. · Revisado há almost 4 years
Não garantimos que as avaliações publicadas sejam de consumidores que compraram ou usaram os produtos. As avaliações não são verificadas pelo Google.