
Ultimate access to all questions.
A junior data engineer is migrating a workload from a relational database system to the Databricks Lakehouse. The source system uses a star schema, leveraging foreign key constraints and multi-table inserts to validate records on write. Which consideration will impact the decisions made by the engineer while migrating this workload?
A
Databricks only allows foreign key constraints on hashed identifiers, which avoid collisions in highly-parallel writes.
B
Databricks supports Spark SQL and JDBC; all logic can be directly migrated from the source system without refactoring.
C
Committing to multiple tables simultaneously requires taking out multiple table locks and can lead to a state of deadlock.
D
All Delta Lake transactions are ACID compliant against a single table, and Databricks does not enforce foreign key constraints.
E
Foreign keys must reference a primary key field; multi-table inserts must leverage Delta Lake's upsert functionality.
Explanation:
Correct Answer: D
This is the correct answer because it accurately describes key architectural differences between traditional relational databases and Databricks Lakehouse:
Delta Lake ACID compliance: Delta Lake provides ACID (Atomicity, Consistency, Isolation, Durability) transactions, but these are guaranteed within a single table, not across multiple tables in a single transaction.
No foreign key constraint enforcement: Databricks does not enforce foreign key constraints at the database level. While you can define foreign key relationships in your data model for documentation purposes, Databricks will not validate or enforce these constraints during write operations.
Why other options are incorrect:
Migration Implications: When migrating from a traditional RDBMS with foreign key constraints to Databricks Lakehouse, the data engineer must: