Topic 1 Question 45
You are training a TensorFlow model on a structured dataset with 100 billion records stored in several CSV files. You need to improve the input/output execution performance. What should you do?
Load the data into BigQuery, and read the data from BigQuery.
Load the data into Cloud Bigtable, and read the data from Bigtable.
Convert the CSV files into shards of TFRecords, and store the data in Cloud Storage.
Convert the CSV files into shards of TFRecords, and store the data in the Hadoop Distributed File System (HDFS).
ユーザの投票
コメント(17)
C - not enough info in the question, but C is the "most correct" one
👍 22ralf_cc2021/07/09Answer is C. TFRecords in cloud storage for big data is the recommended practice by Google for training TF models.
👍 4David_ml2022/05/09- 正解だと思う選択肢: A
https://cloud.google.com/architecture/ml-on-gcp-best-practices#store-tabular-data-in-bigquery BigQuery for structured data, cloud storage for unstructed data
👍 3behzadsw2023/01/06
シャッフルモード