
Answer-first summary for fast verification
Answer: Use Azure Databricks to read the data from the sources, perform the upsert operation using the Delta Lake API, and write the results back to the Delta Lake.
Option B is the correct approach as it leverages Azure Databricks and the Delta Lake API to perform the upsert operation. Azure Databricks can read data from multiple sources and interact with the Delta Lake using its APIs. The Delta Lake API provides the functionality to perform upserts efficiently. Options A, C, and D do not provide the required functionality for upserts in the context of a Delta Lake.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
You are working on a batch processing solution that reads data from multiple sources and writes it to a Delta Lake. The data needs to be upserts, meaning that if a record with the same primary key exists, it should be updated, and if not, it should be inserted. How would you implement this functionality?
A
Use Azure Data Factory to orchestrate the data flow and use the Copy Data activity to upsert the data.
B
Use Azure Databricks to read the data from the sources, perform the upsert operation using the Delta Lake API, and write the results back to the Delta Lake.
C
Use Azure SQL Data Warehouse to store the data and perform the upsert operation using T-SQL.
D
Use Azure Cosmos DB to store the data and perform the upsert operation using its built-in support for upserts.
No comments yet.