US 12,240,106 B2
Manual operation of a remote robot assembly
Jonathan Westin Sykes, Gower, MO (US); Timothy J. Mourlam, Shawnee, KS (US); David Lindquist, Cameron, MO (US); William Naber, Saint Joseph, MO (US); and Robert Nichols, Saint Joseph, MO (US)
Assigned to Altec Industries, Inc., Birmingham, AL (US)
Filed by Altec Industries, Inc., Birmingham, AL (US)
Filed on Sep. 14, 2023, as Appl. No. 18/368,155.
Application 18/368,155 is a continuation of application No. 17/875,674, filed on Jul. 28, 2022, granted, now 11,794,359.
Prior Publication US 2024/0033950 A1, Feb. 1, 2024
Int. Cl. B25J 19/02 (2006.01); B25J 5/00 (2006.01); B25J 9/16 (2006.01); B25J 13/02 (2006.01); H04N 7/18 (2006.01); H04N 13/282 (2018.01); H04N 23/57 (2023.01); H04N 23/66 (2023.01)
CPC B25J 19/023 (2013.01) [B25J 5/00 (2013.01); B25J 9/1689 (2013.01); B25J 13/02 (2013.01); H04N 7/181 (2013.01); H04N 13/282 (2018.05); H04N 23/57 (2023.01); H04N 23/66 (2023.01)] 15 Claims
OG exemplary drawing
 
1. A robotic unit comprising:
a base for receiving one or more components of the robotic unit;
at least one utility arm operatively coupled to the base;
a camera unit operatively coupled to the base, the camera unit comprising at least one camera for capturing visual information and at least one three-dimensional camera for capturing three-dimensional depth information;
a transmitter for transmitting a signal including the visual information and the three-dimensional depth information;
at least one processor; and
at least one non-transitory computer-readable media storing computer-executable instructions that, when executed by the at least one processor, causes the robotic unit to carry out actions, comprising:
receiving the visual information and the three-dimensional depth information;
creating a three-dimensional representation based at least in part on the visual information or the three-dimensional depth information;
transmitting the visual information and the three-dimensional depth information to a control system comprising at least one hand-held controller and a head-mounted controller with a display, the at least one hand-held controller and the head-mounted controller configured to capture movement and position of an inputting user;
receiving a user instruction to perform an action,
wherein the user instruction comprises a user movement that is captured by the at least one hand-held controller or by the head-mounted controller, and
wherein the action is a replication of the user movement;
responsive to receiving the user instruction, modifying the user instruction to generate a modified action based on a first comparison of the user instruction to the three-dimensional representation or on a second comparison of the user instruction to the visual information; and
performing the modified action.