
Ultimate access to all questions.
You are working with a Spark SQL table df_employees that contains the following schema:
employee_id ARRAY<ARRAY<INT>>,
name STRING,
department STRING,
salary DOUBLE
employee_id ARRAY<ARRAY<INT>>,
name STRING,
department STRING,
salary DOUBLE
Each row in employee_id contains multiple arrays of employee IDs (for example: [[101,102],[103,104]]).
Your task is to convert the employee_id column into a single array of integers per row, without increasing the number of rows, using Spark SQL.
Which of the following queries correctly accomplishes this task?
A
SELECT flatten(employee_id) as employee_ids FROM df_employees
SELECT flatten(employee_id) as employee_ids FROM df_employees
B
SELECT explode(employee_id) AS employee_id
FROM df_employees
SELECT explode(employee_id) AS employee_id
FROM df_employees
C
SELECT explode(flatten(employee_id)) AS employee_id
FROM df_employees
SELECT explode(flatten(employee_id)) AS employee_id
FROM df_employees
D
SELECT employee_id
FROM df_employees
LATERAL VIEW explode(employee_id) e AS employee_id
SELECT employee_id
FROM df_employees
LATERAL VIEW explode(employee_id) e AS employee_id