US 11,735,173 B2
Automatically determining language for speech recognition of spoken utterance received via an automated assistant interface
Pu-sen Chao, Los Altos, CA (US); Diego Melendo Casado, Mountain View, CA (US); Ignacio Lopez Moreno, New York, NY (US); and William Zhang, San Mateo, CA (US)
Assigned to GOOGLE LLC, Mountain View, CA (US)
Filed by Google LLC, Mountain View, CA (US)
Filed on May 24, 2021, as Appl. No. 17/328,400.
Application 17/328,400 is a continuation of application No. 16/163,327, filed on Oct. 17, 2018, granted, now 11,017,766.
Application 16/163,327 is a continuation in part of application No. 15/769,013, granted, now 10,839,793, issued on Nov. 17, 2020, previously published as PCT/US2018/027808, filed on Apr. 16, 2018.
Prior Publication US 2021/0280177 A1, Sep. 9, 2021
Int. Cl. G10L 15/22 (2006.01); G10L 15/197 (2013.01); G10L 15/00 (2013.01); G10L 15/30 (2013.01); G10L 15/08 (2006.01); G10L 15/14 (2006.01); G10L 15/18 (2013.01); G10L 13/00 (2006.01)
CPC G10L 15/197 (2013.01) [G10L 13/00 (2013.01); G10L 15/005 (2013.01); G10L 15/08 (2013.01); G10L 15/14 (2013.01); G10L 15/1822 (2013.01); G10L 15/22 (2013.01); G10L 15/30 (2013.01); G10L 2015/088 (2013.01); G10L 2015/223 (2013.01); G10L 2015/228 (2013.01)] 6 Claims
OG exemplary drawing
 
1. A method implemented by one or more processors, the method comprising:
determining that a spoken utterance was received by a computing device from a user, the computing device comprising an automated assistant that is capable of being invoked in response to the user providing the spoken utterance;
causing audio data, which is based on the spoken utterance, to be processed, by at least a first language model and a second language model, wherein the first language model and the second language model are selected according to a user-specific preference of language models for interpreting spoken utterances from the user;
determining, based on processing of the audio data, a first score that characterizes a probability that the spoken utterance was provided in a first language and a second score that characterizes another probability that the spoken utterance was provided in the second language;
determining, based on a user-specific language profile that is accessible to the automated assistant, that the user has intentionally accessed digital content provided in the first language;
determining, based on determining that the user has intentionally accessed the digital content provided in the first language, another first score to reflect an increase in the probability that the spoken utterance was provided in the first language; and
causing, based on the other first score and the second score, additional audio data to be processed according to a language selected from at least the first language and the second language.