| CPC G06F 3/0425 (2013.01) [G06V 10/7715 (2022.01); G06V 10/774 (2022.01); G06V 10/806 (2022.01); G06V 10/82 (2022.01); G06V 20/46 (2022.01); G06V 40/28 (2022.01)] | 7 Claims |

|
1. A method for predicting a touch interaction position on a large display based on a binocular camera, comprising the following steps:
S1, separately acquiring arm movement video frames of a user and facial and eye movement video frames of the user by a binocular camera;
S2, extracting a video clip of each tapping action from the arm movement video frames and the facial and eye movement video frames and obtaining a key frame by screening;
S3, marking the key frame of each tapping action with coordinates to indicate coordinates of a finger in a display screen;
S4, inputting the marked key frame to an efficient convolutional network for online video understanding (ECO)-Lite neural network for training to obtain a predictive network model; and
S5, inputting a video frame of a current operation to be predicted to the predictive network model and outputting a touch interaction position predicted for the current operation.
|