US 12,124,867 B2
Network function placement in vGPU-enabled environments
Uday Pundalik Kurkure, Palo Alto, CA (US); Sairam Veeraswamy, Bangalore (IN); Hari Sivaraman, Palo Alto, CA (US); Lan Vu, Palo Alto, CA (US); and Avinash Kumar Chaurasia, Bangalore (IN)
Assigned to VMware LLC, Palo Alto, CA (US)
Filed by VMware LLC, Palo Alto, CA (US)
Filed on Apr. 7, 2021, as Appl. No. 17/224,293.
Claims priority of application No. 202141003210 (IN), filed on Jan. 22, 2021.
Prior Publication US 2022/0237014 A1, Jul. 28, 2022
Int. Cl. G06F 9/455 (2018.01); G06F 9/50 (2006.01); G06T 1/20 (2006.01)
CPC G06F 9/45558 (2013.01) [G06F 9/5077 (2013.01); G06T 1/20 (2013.01); G06F 2009/45587 (2013.01); G06F 2009/45591 (2013.01); G06F 2009/45595 (2013.01); G06F 9/5033 (2013.01); G06F 9/505 (2013.01); G06F 2209/503 (2013.01); G06F 2209/509 (2013.01)] 20 Claims
OG exemplary drawing
 
8. A method performed by at least one computing device executing machine-readable instructions, the method comprising:
monitoring a virtual graphics processing unit (vGPU)-enabled computing environment comprising a plurality of vGPU-enabled graphics processing units (GPUs);
identifying a network function request associated with a network function, the network function comprising Compute Unified Device Architecture (CUDA) compatible instructions;
determining to instantiate a vGPU-based implementation of the CUDA compatible instructions to handle the network function request;
selecting, in response to the determination to instantiate a vGPU-based implementation of the CUDA compatible instructions to handle the network function request, a vGPU-enabled GPU to handle the network function request, the vGPU-enabled GPU being selected based on at least one of: a network function memory requirement, and a network function IO requirement; and
processing the network function request using an instance of the network function within a virtual machine that is executed using the vGPU-enabled GPU.