Topic 1 Question 104
You used Dataprep to create a recipe on a sample of data in a BigQuery table. You want to reuse this recipe on a daily upload of data with the same schema, after the load job with variable execution time completes. What should you do?
Create a cron schedule in Dataprep.
Create an App Engine cron job to schedule the execution of the Dataprep job.
Export the recipe as a Dataprep template, and create a job in Cloud Scheduler.
Export the Dataprep job as a Dataflow template, and incorporate it into a Composer job.
ユーザの投票
コメント(13)
I'd pick D because it's the only option which allows variable execution (since we need to execute the dataprep job only after the prior load job). Although D suggests the export of Dataflow templates, this discussion suggests that the export option is no longer available (https://stackoverflow.com/questions/72544839/how-to-get-the-dataflow-template-of-a-dataprep-job), there are already Airflow Operators for Dataprep which we should be using instead - https://airflow.apache.org/docs/apache-airflow-providers-google/stable/operators/cloud/dataprep.html
👍 5jkhong2022/12/18- 正解だと思う選択肢: D
Since the load job execution time is unexpected, schedule the Dataprep based on a fixed time window may not work. When the Dataprep job run the first time, we can find the Dataflow job for that in the console. We can use that to create the Template --> With the help of the Composer to determine if the load job is completed, we can then trigger the Dataflow job
👍 5midgoo2023/03/09 - 正解だと思う選択肢: A
It's A. You can set it directly in Dataprep a job and it will use Dataflow under the hood.
👍 4anicloudgirl2022/12/06
シャッフルモード