Dataflow を使用したサーバーレスのデータ処理 - Apache Beam と Dataflow を使用して ETL パイプラインを作成する(Python) のレビュー
11486 件のレビュー
Unsure how to resolve error when running pipeline WARNING:google.auth.compute engine.metadata:Compute Engine Metadata server unavailable on attempt 1 of 3. Reason: http.client transport only supports the http scheme, https is specified
Liao Ivan · 5日前にレビュー済み
the lab cfails to provission workers to the dataflow jobs, no matter if the machine types are stated explicitely.
Ulloa Bermeo Iván Marcelo · 5日前にレビュー済み
Honeycutt Yan · 5日前にレビュー済み
Sundhar Svetak · 6日前にレビュー済み
Moya Rabadán Francisco · 6日前にレビュー済み
Moreira Leonardo · 6日前にレビュー済み
Issue with the project id. There is no parent organization to select.
Honnappa Vanitha · 7日前にレビュー済み
médiocre
Wessim Kasraoui · 7日前にレビュー済み
de Oliveira Patrocínio Igor · 7日前にレビュー済み
Narita Ana · 8日前にレビュー済み
having error on Run your pipeline task, even using the solution code provided in the lab. Traceback (most recent call last): File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/my_pipeline.py", line 108, in <module> run() File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/my_pipeline.py", line 92, in run | 'ReadFromGCS' >> beam.io.ReadFromText(input) File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/df-env/lib/python3.10/site-packages/apache_beam/io/textio.py", line 808, in __init__ self._source = self._source_class( File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/df-env/lib/python3.10/site-packages/apache_beam/io/textio.py", line 144, in __init__ super().__init__( File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/df-env/lib/python3.10/site-packages/apache_beam/io/filebasedsource.py", line 127, in __init__ self._validate() File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/df-env/lib/python3.10/site-packages/apache_beam/options/value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/df-env/lib/python3.10/site-packages/apache_beam/io/filebasedsource.py", line 190, in _validate match_result = FileSystems.match([pattern], limits=[1])[0] File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/df-env/lib/python3.10/site-packages/apache_beam/io/filesystems.py", line 240, in match return filesystem.match(patterns, limits) File "/home/jupyter/training-data-analyst/quests/dataflow_python/1_Basic_ETL/lab/df-env/lib/python3.10/site-packages/apache_beam/io/filesystem.py", line 779, in match raise BeamIOError("Match operation failed", exceptions) apache_beam.io.filesystem.BeamIOError: Match operation failed with exceptions {'gs://qwiklabs-gcp-02-c763640af21b/events.json': RefreshError(TransportError("Failed to retrieve https://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true from the Google Compute Engine metadata service. Compute Engine Metadata server unavailable. Last exception: HTTPSConnectionPool(host='metadata.google.internal', port=443): Max retries exceeded with url: /computeMetadata/v1/instance/service-accounts/default/?recursive=true (Caused by SSLError(SSLCertVerificationError(1, '[SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1017)')))"))}
ye yumeng · 8日前にレビュー済み
proper folders were not cloned for all the tasks
Shafi Shaik · 8日前にレビュー済み
Shaw Sriyansh · 8日前にレビュー済み
Shaw Sriyansh · 8日前にレビュー済み
Cezar Junior Luis Antonio · 8日前にレビュー済み
Sharma Jyoti · 8日前にレビュー済み
Cezar Junior Luis Antonio · 8日前にレビュー済み
Cezar Junior Luis Antonio · 8日前にレビュー済み
Could not get the pipeline to run due to multiple errors. Even the solution would not run. Very disappointing to be kicked out of the lab and lose progress before being able to resolve / troubleshoot. Errors were around: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1017)'))': /computeMetadata/v1/instance/service-accounts/default/?recursive=true and Compute Engine Medadata server unavailable.
De Meillon Mariette · 9日前にレビュー済み
too complex, needs more direction on what to do
Orccotoma Oscar · 10日前にレビュー済み
Longas Daniela · 10日前にレビュー済み
problema con la infrectuctura (falta de espacio)
Chiguay Gabriela · 10日前にレビュー済み
Cezar Junior Luis Antonio · 11日前にレビュー済み
Could not complete Part1 Task 6 run the pipeline (using the provided Solution code) due to the following error: raise BeamIOError("Match operation failed", exceptions) apache_beam.io.filesystem.BeamIOError: Match operation failed with exceptions {'gs://qwiklabs-gcp-00-f5855126f119/events.json': RefreshError(TransportError("Failed to retrieve https://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true from the Google Compute Engine metadata service. Compute Engine Metadata server unavailable. Last exception: HTTPSConnectionPool(host='metadata.google.internal', port=443): Max retries exceeded with url: /computeMetadata/v1/instance/service-accounts/default/?recursive=true (Caused by SSLError(SSLCertVerificationError(1, '[SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1017)')))"))}
Meng Lingmin · 11日前にレビュー済み
Qwiklabs Dataflow Lab Fix Summary Problems SSL/Metadata auth failure — google-auth 2.44.0+ broke the default HTTP transport, causing CERTIFICATE_VERIFY_FAILED errors when the notebook tried to reach GCP's metadata server Zone resource exhaustion — n1-standard-1 (Dataflow's default) was completely unavailable across all us-central1 zones for Qwiklabs accounts Pipeline exits early — p.run() doesn't wait for completion, causing silent failures argparse rejects extra flags — parse_args() blocks passing extra Beam arguments like --worker_machine_type Fixes 1. Fix SSL auth error bashexport GCE_METADATA_MTLS_MODE=none 2. Use e2-standard-2 machine type Add to your run command: bash--worker_machine_type=e2-standard-2 3. Fix pipeline code In my_pipeline.py: python# Change this: opts = parser.parse_args() options = PipelineOptions() p.run() # To this: opts, pipeline_args = parser.parse_known_args() options = PipelineOptions(pipeline_args) p.run().wait_until_finish() 4. Full working command bashcd $BASE_DIR export PROJECT_ID=$(gcloud config get-value project) export GCE_METADATA_MTLS_MODE=none python3 my_pipeline.py \ --project=${PROJECT_ID} \ --region=us-central1 \ --stagingLocation=gs://$PROJECT_ID/staging/ \ --tempLocation=gs://$PROJECT_ID/temp/ \ --runner=DataflowRunner \ --worker_machine_type=e2-standard-2 5. For the Dataflow Template UI Under Optional Parameters → uncheck "Use default machine type" → Series: E2 → Machine type: e2-standard-2
Olson David · 11日前にレビュー済み
公開されたレビューが、製品を購入または使用した人によるものであることは保証されません。Google はこれらのレビューの検証を行っていません。