
Ultimate access to all questions.
A data engineer observes that a critical field present in the Kafka source was inadvertently omitted when writing to Delta Lake, causing it to also be missing in downstream long-term storage. The Kafka retention period is seven days, and the pipeline has been running in production for three months.
How can Delta Lake help prevent this type of data loss in the future?
A
The Delta log and Structured Streaming checkpoints record the full history of the Kafka producer.
B
Delta Lake schema evolution can retroactively calculate the correct value for newly added fields, as long as the data was in the original source.
C
Delta Lake automatically checks that all fields present in the source data are included in the ingestion layer.
D
Data can never be permanently dropped or deleted from Delta Lake, so data loss is not possible under any circumstance.
E
Ingesting all raw data and metadata from Kafka to a bronze Delta table creates a permanent, replayable history of the data state.