
Ultimate access to all questions.
A company needs to deploy language models for inference on edge devices with minimal latency. Which solution best meets these requirements?
A
Deploy optimized small language models (SLMs) on edge devices.
B
Deploy optimized large language models (LLMs) on edge devices.
C
Incorporate a centralized small language model (SLM) API for asynchronous communication with edge devices.
D
Incorporate a centralized large language model (LLM) API for asynchronous communication with edge devices.