The capacity to divvy up a single physical graphics processing unit (GPU) among multiple virtual machines (VMs) has solidified as a functional technology, enabling a more granular approach to resource allocation. This 'GPU partitioning' (GPU-P) strategy, primarily driven by NVIDIA's vGPU software, allows a single GPU to be shared, offering a fraction of its power to each VM rather than dedicating the entire unit to one.
This capability is not a hypothetical future but a deployed reality. For users of Microsoft Azure Stack HCI and Windows Server 2025, the architecture for GPU partitioning is laid out in operational guides. These systems require specific host and guest drivers, with NVIDIA's Virtual GPU Manager and corresponding graphics drivers being central to the process. Dell Technologies, a prominent player in hardware solutions, specifically points to its servers' need for separate GPU-P drivers and advocates for Windows Admin Center (WAC) as a streamlined tool for managing these configurations. NVIDIA's licensing portal is identified as the source for these necessary drivers.
Read More: NVIDIA releases SANA-WM video model on May 17 2026 for faster creation
The implementation of GPU partitioning extends across various platforms. On Hyper-V, the virtualization technology underpinning Windows Server, partitioning can be managed through Windows Admin Center or PowerShell. Projects like 'Easy-GPU-PV' on GitHub are actively seeking to simplify this process on Windows Hyper-V environments, automating VM creation and driver installations alongside GPU partitioning. The ability to assign only a single GPU partition to any given VM is a consistent limitation across these systems.
Divergent Paths: GPU-P versus Discrete Device Assignment
While GPU partitioning enables one-to-many GPU sharing, Discrete Device Assignment (DDA) represents a contrasting approach, assigning a whole GPU to a single VM. Both technologies are relevant within Azure Stack HCI environments, but they cater to different needs and have distinct operational requirements.
Read More: System Freezes Cause "Not Responding" Errors on PCs
| Feature | GPU Partitioning (GPU-P) | Discrete Device Assignment (DDA) |
|---|---|---|
| GPU Resource Model | High (one GPU to many VMs) | Low (one GPU to one VM) |
| Driver Requirement | GPU vendor driver (e.g., NVIDIA) in guest | GPU driver in guest |
| Compatibility | Supported by NVIDIA vGPU software | Broad hardware support |
Crucially, a physical GPU cannot simultaneously function as both a DDA-assigned device and a partitionable GPU. This segregation ensures that the intended operational mode of the GPU is maintained. For users engaging with GPU partitioning, ensuring compatible CPUs, operating systems, and GPUs is paramount, particularly if live migration of VMs is a desired feature. The Azure Local catalog lists supported solutions and GPU models for those adopting these advanced configurations.
Licensing and Driver Nuances
The underlying technology for GPU partitioning, particularly NVIDIA's vGPU, is tied to a specific licensing model. Separate licensing from NVIDIA is a mandatory requirement for utilizing GPU-P features. The installation process for these drivers involves distinct steps for both the host machines and the individual virtual machines. Within Azure Stack HCI, the integration of NVIDIA's vGPU manager and graphics drivers is a critical step in enabling these capabilities. When managing these partitions, information such as the partition ID, VM ID, VRAM allocation, and encode/decode capabilities becomes visible.
Read More: YouTube AI Likeness Detection Tool Now Open to All Creators 18 Plus