US 12,293,260 B2
Generating and deploying packages for machine learning at edge devices
Calvin Yue-Ren Kuo, Mercer Island, WA (US); Jiazhen Chen, Bellevue, WA (US); Jingwei Sun, Issaquah, WA (US); and Haiyang Liu, Issaquah, WA (US)
Assigned to Amazon Technologies, Inc., Seattle, WA (US)
Filed by Amazon Technologies, Inc., Seattle, WA (US)
Filed on Jan. 30, 2018, as Appl. No. 15/884,279.
Claims priority of provisional application 62/589,512, filed on Nov. 21, 2017.
Prior Publication US 2019/0156246 A1, May 23, 2019
Int. Cl. G06N 20/00 (2019.01); G06F 8/60 (2018.01); G06F 18/214 (2023.01); G06N 5/04 (2023.01); H04W 4/38 (2018.01)
CPC G06N 20/00 (2019.01) [G06F 8/60 (2013.01); G06F 18/214 (2023.01); G06N 5/04 (2013.01); H04W 4/38 (2018.02)] 20 Claims
OG exemplary drawing
 
1. A system, comprising:
one or more computing devices of a provider network comprising respective processors and memory to implement a machine learning deployment service to:
receive, from a user of a client of the machine learning deployment service via a management interface of the machine learning deployment service a set of indications including:
a first indication, from the user, of an inference application, wherein the inference application comprises one or more functions configured to perform one or more actions based on inference data generated by a machine learning model;
a second indication, from the user, of a machine learning framework to be used by the inference application, wherein the second indication of the machine learning framework is based on a selection by the user of the machine learning framework from among a plurality of machine learning frameworks stored at the provider network that are available for selection, based on user input via the management interface at the provider network, to be used by the inference application when the inference application is executed on at least one connected device of a remote network of the client, wherein the machine learning framework is configured to run at least a portion of a machine learning model;
a third indication, from the user, of the machine learning model to be used by the inference application, wherein the third indication of the machine learning model is based on a selection by the user of the machine learning model from among a plurality of machine learning models stored at the provider network that are available for selection, based on the user input via the management interface at the provider network, to be used by the inference application when the inference application is executed on the at least one connected device of the remote network of the client, wherein the machine learning model is configured to generate the inference data based on collected data; and
a fourth indication, from the user, of at least one connected device of a remote network of the client to run the inference application, wherein the first indication, the second indication, the third indication, and the fourth indication are different indications received from the same user via the management interface;
responsive to receipt, from the user via the management interface of the machine learning deployment service, of the set of indications:
generate a package based at least on the inference application, the machine learning framework, and the machine learning model; and
deploy the package from the provider network to the at least one connected device of the remote network of the client based on the fourth indication, from the user, of the at least one connected device of the remote network of the client to run the inference application,
wherein another package is deployed from the provider network to another connected device to run the inference application using a different machine learning model based at least on an indication of the other connected device and a selection of the different machine learning model from among the plurality of machine learning models.