US 11,868,895 B2
Dynamic processing element array expansion
Randy Renfu Huang, Morgan Hill, CA (US); Ron Diamant, Santa Clara, CA (US); and Richard John Heaton, San Jose, CA (US)
Assigned to Amazon Technologies, Inc., Seattle, WA (US)
Filed by Amazon Technologies, Inc., Seattle, WA (US)
Filed on Jan. 13, 2023, as Appl. No. 18/154,576.
Application 18/154,576 is a continuation of application No. 16/456,414, filed on Jun. 28, 2019, granted, now 11,568,238.
Prior Publication US 2023/0153620 A1, May 18, 2023
Int. Cl. G06E 1/00 (2006.01); G06E 3/00 (2006.01); G06T 7/00 (2017.01); G06N 3/08 (2023.01); G06N 3/04 (2023.01)
CPC G06N 3/08 (2013.01) [G06N 3/04 (2013.01)] 20 Claims
OG exemplary drawing
 
1. A computer-implemented method, comprising:
receiving a neural network model that includes a first tensor operation between a first tensor and a second tensor;
dividing the first tensor operation into a set of sub-operations, wherein each sub-operation of the set of sub-operations generates a portion of a final output of the first tensor operation, and wherein dividing the first tensor operation comprises one of:
assigning each sub-operation to a respective portion of the first tensor,
assigning each sub-operation to a respective portion of the second tensor, or
assigning each sub-operation to both a respective portion of the first tensor and a respective portion of the second tensor; and
generating instructions for performing individual sub-operations of the set of sub-operations on respective computing engines of a plurality of computing engines.