
Ultimate access to all questions.
In the context of Azure Databricks, Structured Streaming and Delta Lake enable several design patterns for efficient data processing. Consider a scenario where a financial services company needs to monitor transactions in real-time to detect fraudulent activities immediately. The solution must process data continuously, support event-time processing to handle late-arriving data, and maintain state to track transactions over time. Given these requirements, which design pattern would be the BEST choice to implement? Choose the correct option from the following:
A
Micro-batch processing pattern, as it processes data in small batches, providing a balance between latency and throughput.
B
Continuous processing pattern, as it allows for real-time data processing with minimal latency, suitable for immediate fraud detection.
C
Event-time processing pattern, as it ensures accurate processing of data based on when events actually occurred, crucial for handling late-arriving transaction data.
D
Stateful processing pattern, as it maintains the state of transactions over time, essential for tracking and analyzing fraudulent activities across multiple transactions.