Ultimate access to all questions.
Upgrade Now 🚀
Sign in to unlock AI tutor
You are tasked with designing a data pipeline that needs to handle a large volume of data in real-time. Which distributed computing framework would be most suitable for this task, and why?
A
Apache Hadoop
B
Apache Spark
C
AWS Lambda
D
SQL Server Integration Services