| CPC G06N 5/04 (2013.01) | 20 Claims |

|
1. A gateway computing device adapted for processing contextualized AI inference requests, the gateway computing device comprising:
communication circuitry to:
receive, from an edge device, an artificial intelligence (AI) inferencing operation request, wherein the AI inferencing operation request includes: a model identifier of an AI inferencing model, and contextual data; and
transmit, to a remote computing system, a request to execute a selected flavor of the AI inferencing model at the remote computing system; and
processing circuitry to:
select, in response to the AI inferencing operation request, a flavor of the AI inferencing model from among a plurality of flavors of the AI inferencing model based on the model identifier and the contextual data, wherein the flavors operate on respective hardware configurations at the remote computing system; and
transmit a communication to the remote computing system, the remote computing system configured to use the communication to initiate execution of the selected flavor of the AI inferencing model locally at the remote computing system, wherein inferencing results from the execution of the selected flavor of the AI inferencing model at the remote computing system are provided to the edge device via the gateway computing device;
wherein the gateway computing device is connected via a network to the edge device and the remote computing system.
|