New Year Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: ac4s65

An administrator observes that inference requests to Nutanix Enterprise AI are experiencing increased latency.

An administrator observes that inference requests to Nutanix Enterprise AI are experiencing increased latency. After accessing the Infrastructure Usage and Health page, the administrator finds that the Service Health status is marked as Critical, and the GPU utilization on one node is consistently at 100%, while other nodes show moderate usage.

What is the most appropriate next step to resolve the performance issue?

A.

Rebuild the endpoint to include multiple instances.3

B.

Restart the Kubernetes cluster to reset all node metrics and clear the Critical status.

C.

Disable inference endpoints on the overloaded node.

D.

Increase the memory allocation for the node with high GPU usage.

NCP-AI PDF/Engine
  • Printable Format
  • Value of Money
  • 100% Pass Assurance
  • Verified Answers
  • Researched by Industry Experts
  • Based on Real Exams Scenarios
  • 100% Real Questions
buy now NCP-AI pdf
Get 65% Discount on All Products, Use Coupon: "ac4s65"