| CPC G10L 15/14 (2013.01) [G06F 3/167 (2013.01); G10L 15/005 (2013.01); G10L 15/02 (2013.01); G10L 15/1822 (2013.01); G10L 15/183 (2013.01); G10L 15/22 (2013.01); G10L 15/30 (2013.01); G10L 2015/088 (2013.01); G10L 2015/223 (2013.01); G10L 2015/228 (2013.01)] | 16 Claims |

|
1. A method implemented by one or more processors, the method comprising:
receiving audio data corresponding to a spoken utterance of a user, the audio data being based on detection of the spoken utterance by a client device;
processing the audio data using a first speech recognition model corresponding to a first language;
determining, based on processing the audio data using the first speech recognition model, content that is responsive to the spoken utterance;
monitoring for an additional spoken input from the user;
receiving, during the monitoring, additional audio data corresponding to an additional spoken utterance, the additional audio data being based on detection of the additional spoken utterance by the client device;
determining, based on receiving the additional audio data, that the additional spoken utterance is provided by an additional user;
accessing, based on the additional spoken utterance being provided by the additional user, a user profile corresponding to the additional user;
determining, based on accessing the user profile corresponding to the additional user, that the user profile provides a correspondence between the additional user and a second language;
based on determining that the user profile provides the correspondence between the additional user and the second language:
using a second speech recognition model, for the second language, in processing the additional audio data; and
causing the client device to render further responsive content based on the processing of the additional audio data using the second speech recognition model.
|