| CPC G06T 13/40 (2013.01) [G06F 40/47 (2020.01); G06T 13/205 (2013.01); G06V 40/174 (2022.01); G06V 40/20 (2022.01); G09B 21/009 (2013.01); G10L 15/1815 (2013.01); G10L 15/22 (2013.01); G10L 21/10 (2013.01); G10L 25/63 (2013.01); G10L 2021/065 (2013.01)] | 18 Claims |

|
1. A method comprising:
capturing video data using a camera of a device and capturing audio data using a microphone of the device;
extracting a spoken word from the audio data and an image of a speaker who uttered the spoken word from the video data;
querying a sign language database to determine a translation of the spoken word to a sign language gesture;
identifying visual characteristics of the speaker based on the extracted image of the speaker who uttered the spoken word;
generating an avatar based on the identified visual characteristics of the speaker who uttered the spoken word;
identifying in a model database, a skeleton model representing the sign language gesture; and
generating for display an animation of the avatar that was generated based on the identified visual characteristics of the speaker who uttered the spoken word performing the sign language gesture by applying the skeleton model to the avatar.
|