US 12,250,525 B2
One-touch spatial experience with filters for AR/VR applications
Scott Phillip Selfon, Kirkland, WA (US); and Andrew Lovitt, Redmond, WA (US)
Assigned to Meta Platforms Technologies, LLC, Menlo Park, CA (US)
Filed by Meta Platforms Technologies, LLC, Menlo Park, CA (US)
Filed on Jun. 6, 2022, as Appl. No. 17/833,631.
Claims priority of provisional application 63/301,269, filed on Jan. 20, 2022.
Claims priority of provisional application 63/233,143, filed on Aug. 13, 2021.
Prior Publication US 2023/0049175 A1, Feb. 16, 2023
Int. Cl. G06V 20/20 (2022.01); G06F 3/0482 (2013.01); G06V 20/40 (2022.01); G11B 27/031 (2006.01); G11B 27/10 (2006.01); H04N 5/76 (2006.01); H04R 1/40 (2006.01); H04R 3/00 (2006.01)
CPC H04R 3/005 (2013.01) [G06F 3/0482 (2013.01); G06V 20/40 (2022.01); G11B 27/031 (2013.01); G11B 27/10 (2013.01); H04N 5/76 (2013.01); H04R 1/406 (2013.01); H04R 2201/401 (2013.01); H04R 2410/01 (2013.01)] 9 Claims
OG exemplary drawing
 
1. A computer-implemented method, comprising:
receiving, from a user of an immersive reality application, a selection of a first sound source from a recorded video in a display of a client device, the recorded video provided by a headset at an event including a headset user;
identifying a direction of audio for the first sound source relative to the headset user; and
enhancing an audio signal in the recorded video from the first sound source based on the direction of audio,
wherein identifying a direction of audio for the first sound source comprises correlating multiple waveforms from each of multiple soundtracks collected from each of multiple microphones spatially distributed on the headset with a time of arrival of the waveforms to the microphones, and determining a location of the first sound source based on the time of arrival.