
Answer-first summary for fast verification
Answer: Leverage AWS Database Migration Service (DMS) to replicate the dataset and its schema to a target database in AWS, then use AWS Glue to catalog the data.
The most efficient way to transfer a large dataset while maintaining its schema and relationships is to use AWS Database Migration Service (DMS) to replicate the dataset to a target database in AWS. Once the dataset is replicated, you can use AWS Glue to catalog the data, ensuring that the schema and relationships are preserved. Manually exporting and importing the dataset or creating an AWS Glue job may be time-consuming and error-prone. AWS Data Pipeline can be used for orchestration, but it does not directly support the replication of large datasets and their schemas.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
You need to migrate a large dataset from an on-premises data warehouse to AWS Glue Data Catalog. The dataset contains millions of records and spans multiple tables. What is the most efficient way to transfer the dataset while maintaining its schema and relationships?
A
Manually export the dataset from the on-premises data warehouse and import it into the AWS Glue Data Catalog.
B
Use AWS Data Pipeline to orchestrate the data transfer and schema migration from the on-premises data warehouse to the AWS Glue Data Catalog.
C
Leverage AWS Database Migration Service (DMS) to replicate the dataset and its schema to a target database in AWS, then use AWS Glue to catalog the data.
D
Create an AWS Glue job that extracts the dataset from the on-premises data warehouse and writes it to the AWS Glue Data Catalog.
No comments yet.