US 12,471,818 B2
Systems and methods of capturing eye-gaze data
Jatin Sharma, Sammamish, WA (US); Jonathan T. Campbell, Redmond, WA (US); Jay C. Beavers, Duvall, WA (US); and Peter John Ansell, Renton, WA (US)
Assigned to Microsoft Technology Licensing, LLC, Redmond, WA (US)
Filed by Microsoft Technology Licensing, LLC, Redmond, WA (US)
Filed on Apr. 24, 2024, as Appl. No. 18/644,889.
Application 18/644,889 is a continuation of application No. 17/234,568, filed on Apr. 19, 2021, granted, now 11,998,335.
Prior Publication US 2024/0341648 A1, Oct. 17, 2024
Int. Cl. A61B 5/16 (2006.01); A61B 5/00 (2006.01)
CPC A61B 5/163 (2017.08) [A61B 5/7267 (2013.01); A61B 5/7435 (2013.01)] 20 Claims
OG exemplary drawing
 
1. A computer-implemented method for tracking eye-gaze of an operator using an eye-gaze prediction model, the method comprising:
receiving, from a camera, an input image of the operator of a computing device, wherein the input image comprises an eye in a face of the operator;
generating eye-gaze point data using the eye-gaze prediction model, wherein the eye-gaze prediction is previously trained by executing operations comprising:
selecting a scan path from a set of predetermined scan paths, wherein the scan path passes through a point in a region of a series of regions in a grid on a screen of a computing device, and the point represents an expected value of uniformly distributed random eye-gaze points in the region;
displaying a symbol as an eye-gaze target on the screen, wherein the eye-gaze target moves along the scan path for guiding attention of the operator;
receiving a combination of eye-gaze point data and input images associated with a plurality of points along the scan path as training data for the eye-gaze prediction model;
training the eye-gaze prediction model using the training data, wherein the eye-gaze prediction model includes data associated with parameters in one or more neural networks; and
updating the parameters in the one or more neural networks using the trained eye-gaze prediction model; and
executing a function according to an object associated with a location of the screen corresponding to the generated eye-gaze point data.