Ultimate access to all questions.
You have deployed a machine learning model on Google's Vertex AI platform for real-time inference. During an online prediction request, you encounter an 'Out of Memory' error, indicating that the prediction node lacks sufficient memory to process the request. What should you do to resolve this issue?