| CPC B25J 9/163 (2013.01) [B25J 9/1612 (2013.01); B25J 9/1669 (2013.01); B25J 9/1671 (2013.01); B25J 9/1697 (2013.01); G05B 2219/39316 (2013.01); G05B 2219/39536 (2013.01); G05B 2219/39543 (2013.01); G05B 2219/40514 (2013.01); Y10S 901/02 (2013.01); Y10S 901/31 (2013.01); Y10S 901/47 (2013.01)] | 20 Claims |

|
1. A method implemented by one or more processors of a robot, comprising:
identifying an image captured by a vision component associated with the robot;
identifying a target object captured by the image;
generating a target object input that denotes position of the target object in the image;
generating a candidate end effector motion vector defining motion to move a grasping end effector of the robot from a current pose to an additional pose;
identifying a current image captured by the vision component, the current image capturing the grasping end effector and at least the target object;
processing the current image, the candidate end effector motion vector, and the target object input using a trained instance grasping model;
generating, based on the processing, a prediction of successful grasp of the target object with application of the motion defined by the candidate end effector motion vector;
generating an end effector command based on the prediction of successful grasp of the target object; and
providing the end effector command to one or more actuators of the robot.
|