Ultimate access to all questions.
You are tasked with designing a real-time data processing pipeline for a financial services company that requires processing high volumes of transaction data with minimal latency. The solution must ensure data integrity, support scalability to handle peak loads, and comply with financial regulatory standards. Considering these requirements, which of the following approaches would BEST meet the company's needs? Choose one option.
Explanation:
Option B is the best choice because it addresses all the company's requirements. Apache Kafka ensures high-throughput, low-latency data ingestion. Databricks Structured Streaming provides scalable, fault-tolerant real-time processing. Delta Lake offers ACID transactions, which are crucial for data integrity and regulatory compliance. This combination effectively meets the need for scalability, low latency, and compliance with financial standards.