US 12,443,450 B2
Supercluster network of graphical processing units (GPUS)
Jagwinder Singh Brar, Bellevue, WA (US); David Dale Becker, Seattle, WA (US); and Jacob Robert Uecker, Las Vegas, NV (US)
Assigned to Oracle International Corporation, Redwood Shores, CA (US)
Filed by Oracle International Corporation, Redwood Shores, CA (US)
Filed on Nov. 2, 2023, as Appl. No. 18/500,463.
Claims priority of provisional application 63/583,512, filed on Sep. 18, 2023.
Claims priority of provisional application 63/460,766, filed on Apr. 20, 2023.
Claims priority of provisional application 63/425,646, filed on Nov. 15, 2022.
Claims priority of provisional application 63/424,282, filed on Nov. 10, 2022.
Claims priority of provisional application 63/422,650, filed on Nov. 4, 2022.
Prior Publication US 2024/0152396 A1, May 9, 2024
Int. Cl. G06F 9/50 (2006.01); G06F 13/40 (2006.01); H04L 45/24 (2022.01); H04L 45/74 (2022.01)
CPC G06F 9/5027 (2013.01) [G06F 9/5083 (2013.01); G06F 13/4022 (2013.01); H04L 45/24 (2013.01); H04L 45/74 (2013.01); G06F 2209/505 (2013.01)] 20 Claims
OG exemplary drawing
 
1. A method comprising:
providing a network fabric including: (i) a plurality of graphical processing unit (GPU) clusters, the plurality of GPU clusters including at least a first GPU cluster operating at a first speed and a second GPU cluster operating at a second speed that is different than the first speed, and (ii) a plurality of blocks, wherein each block includes: (a) one or more racks that host a GPU cluster, and (b) a plurality of switches arranged in a hierarchical structure that communicatively couple the block to other blocks included in the network fabric; and
responsive to receiving a request to execute a workload, allocating one or more GPUs from the plurality of GPU clusters to execute the workload.