
Answer-first summary for fast verification
Answer: Graphic Processing Unit (GPU)
CUDA (Compute Unified Device Architecture) is NVIDIA's parallel computing platform that enables general-purpose computing on GPUs. To support CUDA computations, a DLVM must have a compatible GPU, as CUDA specifically leverages GPU hardware for parallel processing acceleration, which is essential for deep learning model training. Option C (GPU) is correct because CUDA requires GPU hardware. The community discussion strongly supports this, with 88% selecting C and detailed explanations noting that CUDA is designed for GPU computation. Other options are less suitable: A (SSD) improves storage performance but doesn't enable CUDA; B (CPU overclocking) may boost CPU speed but CUDA relies on GPU, not CPU enhancements; D (High RAM) aids memory-intensive tasks but isn't specific to CUDA; E (Intel SGX) is a security feature unrelated to CUDA. While some comments mention the question may be out of scope, the technical consensus aligns with GPU implementation for CUDA support.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
You need to configure a Deep Learning Virtual Machine (DLVM) to support Compute Unified Device Architecture (CUDA) computations for training deep learning models. What should you implement?
A
Solid State Drives (SSD)
B
Computer Processing Unit (CPU) speed increase by using overclocking
C
Graphic Processing Unit (GPU)
D
High Random Access Memory (RAM) configuration
E
Intel Software Guard Extensions (Intel SGX) technology
No comments yet.