
Answer-first summary for fast verification
Answer: No
## Detailed Analysis ### Understanding the Requirements **Workload Requirements:** - **Data Engineers**: Use Python and SQL, must share a single cluster - **Jobs**: Run notebooks with Python, Scala, and SQL, managed via request process - **Data Scientists**: Perform ad hoc analysis in Scala and R, each must have their own cluster that terminates after 120 minutes of inactivity ### Cluster Type Analysis **Standard Cluster Characteristics:** - Supports all languages: Python, SQL, Scala, and R - Can be configured with auto-termination (default is 120 minutes) - Suitable for single-user scenarios - Can be shared among multiple users **High Concurrency Cluster Characteristics:** - Optimized for multiple concurrent users - Provides better isolation and security through separate processes - **Critical Limitation**: Does not support Scala workloads - Does not auto-terminate by default ### Solution Evaluation **✅ What Works Correctly:** - **Data Scientists**: Standard clusters with auto-termination are appropriate since they support Scala and R, and can be configured to terminate after 120 minutes of inactivity. - **Data Engineers**: A shared Standard cluster works well for Python and SQL workloads. **❌ What Fails:** - **Jobs Cluster**: The solution proposes a High Concurrency cluster for jobs, but the jobs require Scala support. High Concurrency clusters do not support Scala workloads, making this configuration invalid for the specified job requirements. ### Optimal Configuration The correct cluster assignment should be: - **Data Scientists**: Standard clusters (individual, auto-terminating) - **Data Engineers**: Standard cluster (shared) - **Jobs**: Standard cluster (supports all required languages including Scala) ### Why the Solution Fails The fundamental issue is the use of a High Concurrency cluster for jobs that require Scala execution. Since High Concurrency clusters cannot run Scala code, the job workloads would fail or be incomplete. This violates the requirement that jobs must support Python, Scala, and SQL notebooks. **Conclusion**: The proposed solution does not meet the goal because the High Concurrency cluster cannot support the Scala workloads required by the job cluster.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
You plan to create an Azure Databricks workspace with three workloads: one for data engineers using Python and SQL, one for jobs running notebooks with Python, Scala, and SQL, and one for data scientists performing ad hoc analysis in Scala and R. The following company standards must be met:
Proposed Solution: You create a Standard cluster for each data scientist, a Standard cluster for the data engineers, and a High Concurrency cluster for the jobs.
Does this solution meet the goal?
A
Yes
B
No
No comments yet.