
Answer-first summary for fast verification
Answer: Average response time
The correct answer is **C: Average response time**. **Why C is correct:** Average response time directly measures how quickly an AI model processes input and generates output during inference or operational runtime. This metric is crucial for assessing runtime efficiency because it quantifies the latency between receiving a request and delivering a response. In production environments, lower average response times indicate higher efficiency, which is essential for real-time applications like chatbots, recommendation systems, fraud detection, and interactive services where user experience depends on minimal delay. **Why other options are incorrect:** - **A: Customer satisfaction score (CSAT):** This measures user satisfaction with a service or product, not the technical performance or efficiency of an AI model. It is a business metric rather than a runtime efficiency metric. - **B: Training time for each epoch:** This evaluates efficiency during the model training phase, not during operational runtime. Training metrics are relevant for development and optimization but do not reflect how efficiently the model runs in production. - **D: Number of training instances:** This refers to the infrastructure or resources used during training, such as the count of virtual machines or containers. It is a configuration or scaling metric, not a direct measure of runtime efficiency during model operation. In summary, average response time is the most appropriate metric for evaluating the runtime efficiency of operating AI models, as it focuses on performance during inference, aligning with best practices for monitoring deployed AI systems in AWS and other cloud environments.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
No comments yet.