Topic 1 Question 292
You are developing a TensorFlow Extended (TFX) pipeline with standard TFX components. The pipeline includes data preprocessing steps. After the pipeline is deployed to production, it will process up to 100 TB of data stored in BigQuery. You need the data preprocessing steps to scale efficiently, publish metrics and parameters to Vertex AI Experiments, and track artifacts by using Vertex ML Metadata. How should you configure the pipeline run?
Run the TFX pipeline in Vertex AI Pipelines. Configure the pipeline to use Vertex AI Training jobs with distributed processing.
Run the TFX pipeline in Vertex AI Pipelines. Set the appropriate Apache Beam parameters in the pipeline to run the data preprocessing steps in Dataflow.
Run the TFX pipeline in Dataproc by using the Apache Beam TFX orchestrator. Set the appropriate Vertex AI permissions in the job to publish metadata in Vertex AI.
Run the TFX pipeline in Dataflow by using the Apache Beam TFX orchestrator. Set the appropriate Vertex AI permissions in the job to publish metadata in Vertex AI.
ユーザの投票
コメント(1)
- 正解だと思う選択肢: B
A (Vertex AI Training jobs): While Vertex AI Training jobs are useful for model training, they are not the primary way to scale data preprocessing within a TFX pipeline. C and D (Dataproc and Dataflow with Apache Beam TFX orchestrator): While you can run TFX pipelines on Dataproc or Dataflow directly, using Vertex AI Pipelines as the orchestrator provides better integration with Vertex AI services and simplifies metadata tracking and experiment management.
👍 3AB_C2024/11/27
シャッフルモード