US 12,008,988 B2
Electronic apparatus and controlling method thereof
Hyeontaek Lim, Suwon-si (KR); Sejin Kwak, Suwon-si (KR); and Youngjin Kim, Suwon-si (KR)
Assigned to Samsung Electronics Co., Ltd., Suwon-si (KR)
Filed by Samsung Electronics Co., Ltd., Suwon-si (KR)
Filed on Oct. 7, 2020, as Appl. No. 17/065,027.
Claims priority of application No. 10-2019-0125169 (KR), filed on Oct. 10, 2019.
Prior Publication US 2021/0110824 A1, Apr. 15, 2021
Int. Cl. G10L 15/22 (2006.01); G10L 15/18 (2013.01); G10L 15/24 (2013.01)
CPC G10L 15/22 (2013.01) [G10L 15/18 (2013.01); G10L 15/24 (2013.01)] 16 Claims
OG exemplary drawing
 
1. An electronic apparatus comprising:
a microphone;
a camera;
a memory configured to store at least one command; and
at least one processor configured to be connected to the microphone, the camera, and the memory and control the electronic apparatus,
wherein the at least one processor is further configured, by executing the at least one command, to:
receive a first voice input from a first user through the microphone, using a first natural language understanding model at the electronic apparatus to identify a first category information of the first voice input and provide a response to the first voice input,
receive a second voice input through the microphone while the response to the first voice input is being provided, analyze an image captured by the camera and determine whether the second voice input is uttered by the first user based on the captured image,
in response to determining that the second voice input is uttered by the first user, stop the providing the response to the first voice input and using the first natural language understanding model to identify second category information in the second voice input,
determine whether the first category in the first voice input is being related to the second category in the second voice input,
in response to the first category information in the first voice input is being related to the second category information in the second voice input, use the first natural language understanding model to perform natural language understanding on the second voice input, and
in response to the first category information in the first voice input is being unrelated to the second category information in the second voice input, control a communication interface to transmit the second voice input to a server and using a second natural language understanding model at the server to perform natural language understanding on the second voice input.