
Answer-first summary for fast verification
Answer: ```spark.readStream.format("cloudFiles").option("cloudFiles.format","csv").option("cloudFiles.schemaLocation", checkpoint_directory).load("landing").writeStream.option("checkpointLocation", checkpoint_directory).table(raw)```
The correct command uses `readStream` with `format("cloudFiles")` to leverage Auto Loader for streaming data from the landing area, specifies the CSV format, and includes the schema location and checkpoint directory for metadata and progress tracking. This ensures the data is correctly streamed into the Bronze layer. Incorrect options either miss `readStream` or use incorrect formats or methods, which are crucial for the Auto Loader's operation and the medallion architecture's requirements.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
Which Auto Loader structured streaming command correctly transfers data from the landing area to the Bronze layer in Databricks?
A
spark.readStream.format("csv").option("cloudFiles.schemaLocation", checkpoint_directory).load("landing").writeStream.option("checkpointLocation", checkpoint_directory).table(raw)
B
spark.read.format("cloudFiles").option("cloudFiles.format","csv").option("cloudFiles.schemaLocation", checkpoint_directory).load("landing").writeStream.option("checkpointLocation", checkpoint_directory).table(raw)
C
spark.readStream.load(rawSalesLocation).writeStream.option("checkpointLocation", checkpointPath).outputMode("append").table("uncleanedSales")
D
spark.read.load(rawSalesLocation).writeStream.option("checkpointLocation", checkpointPath).outputMode("append").table("uncleanedSales")
E
spark.readStream.format("cloudFiles").option("cloudFiles.format","csv").option("cloudFiles.schemaLocation", checkpoint_directory).load("landing").writeStream.option("checkpointLocation", checkpoint_directory).table(raw)
No comments yet.