
Ultimate access to all questions.
You have an Azure Databricks workspace and an Azure Data Lake Storage Gen2 account named storage1. New files are uploaded daily to storage1.
You need to recommend a solution to configure storage1 as a structured streaming source that meets the following requirements:
Incrementally process new files as they arrive.
Minimize implementation and maintenance effort.
Minimize the cost of processing millions of files.
Support schema inference and schema evolution.
What should you include in the recommendation?
A
COPY INTO
B
Azure Data Factory
C
Auto Loader
D
Apache Spark FileStreamSource