
Answer-first summary for fast verification
Answer: Number of tokens consumed
## Explanation For Amazon Bedrock, inference costs are primarily driven by the **number of tokens consumed**. Here's why: ### How Amazon Bedrock Pricing Works: 1. **Token-based pricing**: Amazon Bedrock charges based on the number of input and output tokens processed during inference. 2. **Input tokens**: The text you provide to the model (prompt) 3. **Output tokens**: The text generated by the model (response) ### Why Other Options Are Incorrect: - **B. Temperature value**: Temperature is a parameter that controls the randomness/creativity of the model's output (0-1 scale), but it doesn't directly affect pricing. - **C. Amount of data used to train the LLM**: This is a fixed cost incurred by the model provider during training, not a variable cost for inference. - **D. Total training time**: Similar to option C, this is a one-time training cost, not an ongoing inference cost. ### Key Points: - Different foundation models on Amazon Bedrock have different per-token pricing - Pricing may vary between input and output tokens - Some models may have minimum charges or different pricing tiers - The number of tokens directly correlates with the computational resources required to generate the inference This token-based pricing model is common across many LLM services as it accurately reflects the actual computational work performed by the model.
Author: Ritesh Yadav
Ultimate access to all questions.
A company wants to assess the costs that are associated with using a large language model (LLM) to generate inferences. The company wants to use Amazon Bedrock to build generative AI applications. Which factor will drive the inference costs?
A
Number of tokens consumed
B
Temperature value
C
Amount of data used to train the LLM
D
Total training time
No comments yet.