
Ultimate access to all questions.
A data scientist is transitioning their pandas DataFrame code to utilize the pandas API on Spark. They are working with the following incomplete code snippet:
________BLANK_________
df = ps.read_parquet(path)
df["category"].value_counts()
________BLANK_________
df = ps.read_parquet(path)
df["category"].value_counts()
Which line of code should they use to successfully complete the refactoring with the pandas API on Spark?
A
import pandas as ps
B
import databricks.pandas as ps
C
import pyspark.pandas as ps
D
import pandas.spark as ps
E
import databricks.pyspark as ps