US 11,689,877 B2
Immersive augmented reality experiences using spatial audio
Ilteris Canberk, Marina Del Rey, CA (US); Shin Hwun Kang, Playa Del Rey, CA (US); and James Powderly, Venice, CA (US)
Assigned to Snap Inc., Santa Monica, CA (US)
Filed by Ilteris Canberk, Marina Del Rey, CA (US); Shin Hwun Kang, Playa Del Rey, CA (US); and James Powderly, Venice, CA (US)
Filed on Jun. 8, 2021, as Appl. No. 17/342,031.
Application 17/342,031 is a continuation of application No. 16/836,363, filed on Mar. 31, 2020, granted, now 11,089,427.
Prior Publication US 2021/0306791 A1, Sep. 30, 2021
Int. Cl. H04S 7/00 (2006.01); G02B 27/01 (2006.01); G06T 19/00 (2011.01); H04W 4/029 (2018.01)
CPC H04S 7/303 (2013.01) [G02B 27/0176 (2013.01); G06T 19/006 (2013.01); H04W 4/029 (2018.02); G02B 2027/014 (2013.01); H04R 2499/15 (2013.01); H04S 2400/11 (2013.01)] 19 Claims
OG exemplary drawing
 
1. A device configured to be head mounted on a user, comprising:
a frame having a first side and a second side;
a first temple extending from a first side of the frame, the first temple having a proximal end adjacent the first side of the frame and a distal end;
a second temple extending from a second side of the frame, the second temple having a proximal end adjacent the second side of the frame and a distal end;
a processor;
a memory;
at least one image sensor;
at least four speakers that produce at least three directional audio zones, a first speaker of the at least four speakers positioned adjacent the proximal end of the first temple, a second speaker of the at least four speakers positioned adjacent the distal end of the first temple, a third speaker of the at least four speakers positioned adjacent the proximal end of the second temple, and a fourth speaker of the at least four speakers positioned adjacent the distal end of the second temple; and
programming in said memory, wherein execution of said programming by said processor configures the device to perform functions, including functions to:
capture, using the at least one image sensor, images in an environment of the device;
identify at least one of an object or feature within the captured images;
determine a position of the device within the environment with respect to the at least one object or feature;
determine a target location within the environment that may be associated with the object or feature;
determine a current orientation of the device with respect to the target location;
selectively emit audio signals from the at least one speaker in respective directional audio zones responsive to the current orientation to guide the user to the target location; and
adjust a volume of the audio signals to indicate a distance of the device from the target location.