
Ultimate access to all questions.
Deep dive into the quiz with AI chat providers.
We prepare a focused prompt with your quiz and certificate details so each AI can offer a more tailored, in-depth explanation.
Question: 34
You are developing an AI-powered knowledge base application for a global research organization. The application will generate detailed technical reports based on user queries. Evaluation metrics include perplexity (response quality), throughput (tokens generated per second), and memory usage. The LLM must deliver highly accurate, contextually relevant information, while minimizing resource consumption. Which of the following LLM configurations would best meet the application's requirements for high accuracy, moderate throughput, and efficient memory usage?
A
A 6-billion parameter model with moderate perplexity, low memory usage, and high throughput.
B
A 1-billion parameter model with high perplexity, low memory usage, and very high throughput.
C
A 13-billion parameter model with low perplexity, moderate memory usage, and moderate throughput.
D
A 30-billion parameter model with very low perplexity but high memory usage and low throughput.
Explanation:
A 13-billion parameter model with low perplexity, moderate memory usage, and moderate throughput is the best choice for this use case because it provides a balance of high accuracy (low perplexity), sufficient performance (moderate throughput), and manageable resource consumption (moderate memory usage). This configuration ensures that the LLM can generate contextually relevant and accurate technical reports while operating efficiently in terms of memory and response speed.
C strikes the right balance between accuracy, performance, and resource efficiency, meeting the requirements of the application effectively.