Topic 1 Question 177
You want to rebuild your batch pipeline for structured data on Google Cloud. You are using PySpark to conduct data transformations at scale, but your pipelines are taking over twelve hours to run. To expedite development and pipeline run time, you want to use a serverless tool and SOL syntax. You have already moved your raw data into Cloud Storage. How should you build the pipeline on Google Cloud while meeting speed and processing requirements?
Convert your PySpark commands into SparkSQL queries to transform the data, and then run your pipeline on Dataproc to write the data into BigQuery.
Ingest your data into Cloud SQL, convert your PySpark commands into SparkSQL queries to transform the data, and then use federated quenes from BigQuery for machine learning.
Ingest your data into BigQuery from Cloud Storage, convert your PySpark commands into BigQuery SQL queries to transform the data, and then write the transformations to a new table.
Use Apache Beam Python SDK to build the transformation pipelines, and write the data into BigQuery.
ユーザの投票
コメント(11)
- 正解だと思う選択肢: C
The question is C but not because the SQL Syntax, as you can perfectly use SparkSQL on Dataproc reading files from GCS. It's because the "serverless" requirement.
👍 8devaid2022/10/05 - 正解だと思う選択肢: C
This same question is there on Google's Professional Machine Learning Engineer, Question 4 Answer is C.
👍 3TNT872022/09/22 Answer C
👍 2TNT872022/09/07
シャッフルモード