G10L15/28

COMPUTING RESOURCE-SAVING VOICE ASSISTANT
20220406312 · 2022-12-22 ·

A voice assistant includes an electronic processor unit connected to at least one microphone and to remote equipment. The electronic processor unit includes both single detection modules for detecting respective single keywords from an audio signal supplied by the microphone, and also a control unit connected to the single detection modules to select predetermined actions as a function of the detected keywords and to perform those actions. The control module is also arranged to detect whether actions are doable and to activate or deactivate the single detection modules as a function of the doability of the actions.

Detecting a trigger of a digital assistant

Systems and processes for operating an intelligent automated assistant are provided. In accordance with one example, a method includes, at an electronic device with one or more processors, memory, and a plurality of microphones, sampling, at each of the plurality of microphones of the electronic device, an audio signal to obtain a plurality of audio signals; processing the plurality of audio signals to obtain a plurality of audio streams; and determining, based on the plurality of audio streams, whether any of the plurality of audio signals corresponds to a spoken trigger. The method further includes, in accordance with a determination that the plurality of audio signals corresponds to the spoken trigger, initiating a session of the digital assistant; and in accordance with a determination that the plurality of audio signals does not correspond to the spoken trigger, foregoing initiating a session of the digital assistant.

Detecting a trigger of a digital assistant

Systems and processes for operating an intelligent automated assistant are provided. In accordance with one example, a method includes, at an electronic device with one or more processors, memory, and a plurality of microphones, sampling, at each of the plurality of microphones of the electronic device, an audio signal to obtain a plurality of audio signals; processing the plurality of audio signals to obtain a plurality of audio streams; and determining, based on the plurality of audio streams, whether any of the plurality of audio signals corresponds to a spoken trigger. The method further includes, in accordance with a determination that the plurality of audio signals corresponds to the spoken trigger, initiating a session of the digital assistant; and in accordance with a determination that the plurality of audio signals does not correspond to the spoken trigger, foregoing initiating a session of the digital assistant.

NATURAL LANGUAGE PROCESSING ROUTING

Devices and techniques are generally described for a speech processing routing architecture. First input data representing an input request may be received. First data including a semantic interpretation of the input request may be determined. Metadata of the first input data may be determined. The metadata may identify an entity associated with the input request. In some examples, a query may be sent to a first component. The query may include the metadata. In some examples, second data that identifies a first skill associated with the entity may be received from the first component. In various examples, the first skill may be selected for processing the first input data based at least in part on the first data and the second data.

NATURAL LANGUAGE PROCESSING ROUTING

Devices and techniques are generally described for a speech processing routing architecture. First input data representing an input request may be received. First data including a semantic interpretation of the input request may be determined. Metadata of the first input data may be determined. The metadata may identify an entity associated with the input request. In some examples, a query may be sent to a first component. The query may include the metadata. In some examples, second data that identifies a first skill associated with the entity may be received from the first component. In various examples, the first skill may be selected for processing the first input data based at least in part on the first data and the second data.

Discrete three-dimensional processor

A discrete 3-D processor comprises first and second dice. The first die comprises three-dimensional memory (3D-M) arrays, whereas the second die comprises logic circuits and at least an off-die peripheral-circuit component of the 3D-M array(s). The first die does not comprise the off-die peripheral-circuit component. The first and second dice are communicatively coupled by a plurality of inter-die connections. The preferred discrete 3-D processor can be applied to mathematical computing, computer simulation, configurable gate array, pattern processing and neural network.

Dialogue processing apparatus, a vehicle including the same, and a dialogue processing method

A dialogue processing apparatus includes: a speech input device configured to receive a speech signal of a user; a first buffer configured to store the received speech signal therein; an output device; and a controller. The controller is configured to: detect an utterance end time point on the basis of the stored speech signal; generate a second speech recognition result corresponding to a speech signal after the utterance end time point on the basis of whether an intention of the user is to be identified from a first speech recognition result corresponding to a speech signal before the utterance end time point; and control the output device to output a response corresponding to the intention of the user determined on the basis of at least one of the first speech recognition result or the second speech recognition result.

Dialogue processing apparatus, a vehicle including the same, and a dialogue processing method

A dialogue processing apparatus includes: a speech input device configured to receive a speech signal of a user; a first buffer configured to store the received speech signal therein; an output device; and a controller. The controller is configured to: detect an utterance end time point on the basis of the stored speech signal; generate a second speech recognition result corresponding to a speech signal after the utterance end time point on the basis of whether an intention of the user is to be identified from a first speech recognition result corresponding to a speech signal before the utterance end time point; and control the output device to output a response corresponding to the intention of the user determined on the basis of at least one of the first speech recognition result or the second speech recognition result.

A METHOD AND AN APPARATUS FOR EXECUTING OPERATION/S ON DEVICE/S
20220392447 · 2022-12-08 ·

Aspects of the invention are directed towards an apparatus and method for executing operation/s on device/s. One or more embodiments of the invention describe the method comprising steps of receiving a voice command from a user for enabling one or more devices to execute an operation and determining validity of the voice command received from the user. The method further describes steps of converting the voice command to a generic command based on the validity of the command and transmitting the generic command for enabling the one or more devices to execute the operation.

A METHOD AND AN APPARATUS FOR EXECUTING OPERATION/S ON DEVICE/S
20220392447 · 2022-12-08 ·

Aspects of the invention are directed towards an apparatus and method for executing operation/s on device/s. One or more embodiments of the invention describe the method comprising steps of receiving a voice command from a user for enabling one or more devices to execute an operation and determining validity of the voice command received from the user. The method further describes steps of converting the voice command to a generic command based on the validity of the command and transmitting the generic command for enabling the one or more devices to execute the operation.