US 11,776,247 B2
Classification parallelization architecture
William S. Bowman, Melbourne, FL (US); Sean Wagoner, West Melbourne, FL (US); Andrew D. Falendysz, Grant, FL (US); Matthew D. Summer, Melbourne, FL (US); Kevin Makovy, West Melbourne, FL (US); Jeffrey S. Cooper, Centreville, VA (US); and Brad Truesdell, Indialantic, FL (US)
Assigned to Tomahawk Robotics, Melbourne, FL (US)
Filed by Tomahawk Robotics, Melbourne, FL (US)
Filed on Jan. 7, 2022, as Appl. No. 17/571,081.
Prior Publication US 2023/0222783 A1, Jul. 13, 2023
Int. Cl. G06V 10/70 (2022.01); G06V 10/96 (2022.01); G05B 13/02 (2006.01); G06V 10/94 (2022.01); G06V 20/56 (2022.01); G06V 10/764 (2022.01)
CPC G06V 10/96 (2022.01) [G05B 13/0265 (2013.01); G06V 10/764 (2022.01); G06V 10/87 (2022.01); G06V 10/955 (2022.01); G06V 20/56 (2022.01)] 20 Claims
OG exemplary drawing
 
1. A system for generating composite images that include object identifiers determined based on processing source images through multiple machine learning models, the system comprising:
one or more processors; and
a non-transitory computer-readable storage medium storing instructions, which when executed by the one or more processors cause the one or more processors to perform operations comprising:
capturing, at an unmanned vehicle, a plurality of images corresponding to a video stream;
determining a plurality of object types to be recognized in the plurality of images;
determining, based on the plurality of object types, a plurality of machine learning models for processing the plurality of images, wherein each machine learning model identifies an object type different from object types identified by other machine learning models in the plurality of machine learning models;
sequentially inputting the plurality of images into each of the plurality of machine learning models to obtain a plurality of sets of objects from the plurality of machine learning models, wherein each machine learning model in the plurality of machine learning models identifies one or more object types of the plurality of object types in the plurality of images;
receiving from each machine learning model coordinates within each image of the plurality of images corresponding to each object identified in a corresponding image of the plurality of images;
generating a set of composite images for the plurality of images with a plurality of identifiers corresponding to each identified object;
generating an output video stream comprising the set of composite images to be displayed in chronological order; and
transmitting the output video stream from the unmanned vehicle to a controller.