You have thousands of Apache Spark jobs running in your on-premises Apache Hadoop cluster. You want to migrate the jobs to Google Cloud. You want to use managed services to run your jobs instead of maintaining a long-lived Hadoop cluster yourself. You have a tight timeline and want to keep code changes to a minimum. What should you do?

  • A. Move your data to BigQuery. Convert your Spark scripts to a SQL-based processing approach.
  • B. Rewrite your jobs in Apache Beam. Run your jobs in Dataflow.
  • C. Copy your data to Compute Engine disks. Manage and run your jobs directly on those instances.
  • D. Move your data to Cloud Storage. Run your jobs on Dataproc.
Show Suggested Answer Hide Answer
Suggested Answer: D 🗳️

Exam Professional Data Engineer topic 1 question 286 discussion - ExamTopics


Click on the Run Some AI Magic button and choose an AI action to run on this article