
Ultimate access to all questions.
Your company consistently deals with both batch- and stream-based event data for various applications. To handle and process this data effectively, you intend to use Google Cloud Dataflow within a predictable time frame. Nevertheless, you have identified that, at times, the data might arrive later than expected or in a non-sequential manner. In light of these potential delays and disorder, how should you structure your Cloud Dataflow pipeline to appropriately manage and process data that arrives late or out of order?
A
Set a single global window to capture all the data.
B
Set sliding windows to capture all the lagged data.
C
Use watermarks and timestamps to capture the lagged data.
D
Ensure every datasource type (stream or batch) has a timestamp, and use the timestamps to define the logic for lagged data.