The foundation of the NVIDIA software stack is the DGX OS. Which of the following Linux distributions is DGX OS built upon?
Answer : A
DGX OS, the operating system powering NVIDIA DGX systems, is built on Ubuntu Linux, specifically the Long-Term Support (LTS) version. It integrates Ubuntu's robust base with NVIDIA-specific enhancements, including GPU drivers, tools, and optimizations tailored for AI and high-performance computing workloads. Neither Red Hat nor CentOS serves as the foundation for DGX OS, making Ubuntu the correct choice.
(Reference: NVIDIA DGX OS Documentation, System Requirements Section)
Which solution should be recommended to support real-time collaboration and rendering among a team?
Answer : C
An NVIDIA Certified Server with RTX GPUs is optimized for real-time collaboration and rendering, supporting NVIDIA Virtual Workstation (vWS) software. This setup enables low-latency, multi-user graphics workloads, ideal for team-based design or visualization. T4 GPUs focus on inference efficiency, and DGX SuperPOD targets large-scale AI training, not collaborative rendering.
(Reference: NVIDIA AI Infrastructure and Operations Study Guide, Section on GPU Selection for Collaboration)
Which of the following aspects have led to an increase in the adoption of AI? (Choose two.)
Answer : C, D
The surge in AI adoption is driven by two key enablers: high-powered GPUs and large amounts of data. High-powered GPUs provide the massive parallel compute capabilities necessary to train complex AI models, particularly deep neural networks, by processing numerous operations simultaneously, significantly reducing training times. Simultaneously, the availability of large datasets---spanning text, images, and other modalities---provides the raw material that modern AI algorithms, especially data-hungry deep learning models, require to learn patterns and make accurate predictions. While Moore's Law (the doubling of transistor counts) has historically aided computing, its impact has slowed, and rule-based machine learning has largely been supplanted by data-driven approaches.
(Reference: NVIDIA AI Infrastructure and Operations Study Guide, Section on AI Adoption Drivers)
Which phase of deep learning benefits the greatest from a multi-node architecture?
Answer : B
Training is the deep learning phase that benefits most from a multi-node architecture. It involves compute-intensive operations---forward and backward passes, gradient computation, and synchronization---across large datasets and complex models. Distributing these tasks across multiple nodes with GPUs accelerates processing, reduces time to convergence, and enables handling models too large for a single node. While data augmentation and inference can leverage multiple nodes, their gains are less pronounced, as they typically involve lighter or more localized computation.
(Reference: NVIDIA AI Infrastructure and Operations Study Guide, Section on Multi-Node Training)
What is a key value of using NVIDIA NIMs?
Answer : A
NVIDIA NIMs (NVIDIA Inference Microservices) are pre-built, GPU-accelerated microservices with standardized APIs, designed to simplify and accelerate AI model deployment across diverse environments---clouds, data centers, and edge devices. Their key value lies in enabling fast, turnkey inference without requiring custom deployment pipelines, reducing setup time and complexity. While community support and SDK deployment may be tangential benefits, they are not the primary focus of NIMs.
(Reference: NVIDIA NIMs Documentation, Overview Section)
Which two components are included in GPU Operator? (Choose two.)
Answer : A, C
The NVIDIA GPU Operator is a tool for automating GPU resource management in Kubernetes environments. It includes two key components: GPU drivers, which provide the necessary software to interface with NVIDIA GPUs, and the NVIDIA Data Center GPU Manager (DCGM), which offers health monitoring, telemetry, and diagnostics for GPU clusters. Frameworks like PyTorch and TensorFlow are separate AI development tools, not part of the GPU Operator, which focuses on infrastructure rather than application layers.
(Reference: NVIDIA GPU Operator Documentation, Components Section)
When should RoCE be considered to enhance network performance in a multi-node AI computing environment?
Answer : C
RoCE (RDMA over Converged Ethernet) enhances network performance by offloading data transport to the NIC via RDMA, bypassing CPU involvement. It's particularly valuable when high CPU utilization limits bandwidth usage, as it reduces overhead and unlocks full link capacity. While RoCE can handle storage traffic, it's less effective with high packet loss (requiring reliable networks), making CPU-bound scenarios its prime use case.
(Reference: NVIDIA Networking Documentation, Section on RoCE Benefits)