
Ultimate access to all questions.
Question: 34
You are developing an AI-powered knowledge base application for a global research organization. The application will generate detailed technical reports based on user queries. Evaluation metrics include perplexity (response quality), throughput (tokens generated per second), and memory usage. The LLM must deliver highly accurate, contextually relevant information, while minimizing resource consumption. Which of the following LLM configurations would best meet the application's requirements for high accuracy, moderate throughput, and efficient memory usage?
Explanation:
A 13-billion parameter model with low perplexity, moderate memory usage, and moderate throughput is the best choice for this use case because it provides a balance of high accuracy (low perplexity), sufficient performance (moderate throughput), and manageable resource consumption (moderate memory usage). This configuration ensures that the LLM can generate contextually relevant and accurate technical reports while operating efficiently in terms of memory and response speed.
C strikes the right balance between accuracy, performance, and resource efficiency, meeting the requirements of the application effectively.