Topic 1 Question 176
You work for a food product company. Your company’s historical sales data is stored in BigQuery.You need to use Vertex AI’s custom training service to train multiple TensorFlow models that read the data from BigQuery and predict future sales. You plan to implement a data preprocessing algorithm that performs mm-max scaling and bucketing on a large number of features before you start experimenting with the models. You want to minimize preprocessing time, cost, and development effort. How should you configure this workflow?
Write the transformations into Spark that uses the spark-bigquery-connector, and use Dataproc to preprocess the data.
Write SQL queries to transform the data in-place in BigQuery.
Add the transformations as a preprocessing layer in the TensorFlow models.
Create a Dataflow pipeline that uses the BigQuerylO connector to ingest the data, process it, and write it back to BigQuery.
ユーザの投票
コメント(3)
- 正解だと思う選択肢: B
BigQuery (SQL) is the easiest, cheapest approach
👍 1b1a8fae2024/01/10 - 正解だと思う選択肢: B
BigQuery can do both transformations https://cloud.google.com/bigquery/docs/manual-preprocessing#numerical_functions
👍 1BlehMaks2024/01/12 C This option allows you to leverage the power and simplicity of TensorFlow to preprocess and transform the data with simple Python code
👍 136bdc1e2024/01/13
シャッフルモード