
Ultimate access to all questions.
Deep dive into the quiz with AI chat providers.
We prepare a focused prompt with your quiz and certificate details so each AI can offer a more tailored, in-depth explanation.
Why do Transformers use positional encodings?
A
To increase randomness in generation
B
To provide sequence order information to the model
C
To reduce model size
D
To avoid gradient vanishing
Explanation:
Transformers use positional encodings to provide sequence order information to the model. Unlike RNNs and LSTMs which process sequences sequentially and inherently capture order, Transformers process all tokens in parallel through self-attention mechanisms. This parallel processing means they don't have built-in awareness of token positions. Positional encodings add information about the position of each token in the sequence, allowing the model to understand the order of words or tokens, which is crucial for understanding language structure and meaning.