Patent classifications
G10L17/22
APPARATUS FOR CONTROLLING DISPLAY OF VEHICLE AND METHOD THEREOF
The present disclosure relates to an apparatus for controlling a multi-image display provided in a vehicle, and a method thereof. According to the present disclosure, a multi-image display device may simultaneously output two or more images separated depending on an output angle or an optical frequency, and a first reflection member provided in an indoor ceiling of the vehicle may reflect one image among the two or more images output from the multi-image display device.
Voice controlled media playback system
Disclosed herein are systems and methods for receiving a voice command and determining an appropriate action for the media playback system to execute based on user identification. The systems and methods receive a voice command for a media playback system, and determines whether the voice command was received from a registered user of the media playback system. In response to determining that the voice command was received from a registered user, the systems and methods configure an instruction for the media playback system based on content from the voice command and information in a user profile for the registered user.
Voice controlled media playback system
Disclosed herein are systems and methods for receiving a voice command and determining an appropriate action for the media playback system to execute based on user identification. The systems and methods receive a voice command for a media playback system, and determines whether the voice command was received from a registered user of the media playback system. In response to determining that the voice command was received from a registered user, the systems and methods configure an instruction for the media playback system based on content from the voice command and information in a user profile for the registered user.
Information processing system, information processing apparatus including circuitry to store position information of users present in a space and control environment effect production, information processing method, and room
An information processing system includes: an image display apparatus provided in a space and configured to display an image; a sensor apparatus carried by a user who is present in the space and configured to output a signal for detecting position information of the user in the space; and an information processing apparatus. The information processing apparatus includes circuitry configured to store a plurality of pieces of position information of a plurality of users including the user, who are in present in the space, in association with the plurality of users, the plurality of users being detected based on signals output from a plurality of sensor apparatuses including the sensor apparatus, and control environment effect production that supports communication between the plurality of users by the image displayed by the image display apparatus, based on each of the plurality of pieces of position information of the plurality of users.
Information processing system, information processing apparatus including circuitry to store position information of users present in a space and control environment effect production, information processing method, and room
An information processing system includes: an image display apparatus provided in a space and configured to display an image; a sensor apparatus carried by a user who is present in the space and configured to output a signal for detecting position information of the user in the space; and an information processing apparatus. The information processing apparatus includes circuitry configured to store a plurality of pieces of position information of a plurality of users including the user, who are in present in the space, in association with the plurality of users, the plurality of users being detected based on signals output from a plurality of sensor apparatuses including the sensor apparatus, and control environment effect production that supports communication between the plurality of users by the image displayed by the image display apparatus, based on each of the plurality of pieces of position information of the plurality of users.
Digital assistant processing of stacked data structures
Processing stacked data structures is provided. A system receives an input audio signal detected by a sensor of a local computing device, identifies an acoustic signature, and identifies an account corresponding to the signature. The system establishes a session and a profile stack data structure including a first profile layer having policies configured by a third-party device. The system pushes, to the profile stack data structure, a second profile layer retrieved from the account. The system parses the input audio signal to identify a request and a trigger keyword. The system generates, based on the trigger keyword and the second profile layer, a first action data structure compatible with the first profile layer. The system provides the first action data structure for execution. The system disassembles the profile stack data structure to remove the first profile layer or the second profile layer from the profile stack data structure.
AUTOMATIC INTERPRETATION SERVER AND METHOD BASED ON ZERO UI
Provided a method performed by an automatic interpretation server based on a zero user interface (UI), which communicates with a plurality of terminal devices having a microphone function, a speaker function, a communication function, and a wearable function. The method includes connecting terminal devices disposed within a designated automatic interpretation zone, receiving a voice signal of a first user from a first terminal device among the terminal devices within the automatic interpretation zone, matching a plurality of users placed within a speech-receivable distance of the first terminal device, and performing automatic interpretation on the voice signal and transmitting results of the automatic interpretation to a second terminal device of at least one second user corresponding to a result of the matching.
AUTOMATIC INTERPRETATION SERVER AND METHOD BASED ON ZERO UI
Provided a method performed by an automatic interpretation server based on a zero user interface (UI), which communicates with a plurality of terminal devices having a microphone function, a speaker function, a communication function, and a wearable function. The method includes connecting terminal devices disposed within a designated automatic interpretation zone, receiving a voice signal of a first user from a first terminal device among the terminal devices within the automatic interpretation zone, matching a plurality of users placed within a speech-receivable distance of the first terminal device, and performing automatic interpretation on the voice signal and transmitting results of the automatic interpretation to a second terminal device of at least one second user corresponding to a result of the matching.
Speaker verification
Methods, systems, apparatus, including computer programs encoded on computer storage medium, to facilitate language independent-speaker verification. In one aspect, a method includes actions of receiving, by a user device, audio data representing an utterance of a user. Other actions may include providing, to a neural network stored on the user device, input data derived from the audio data and a language identifier. The neural network may be trained using speech data representing speech in different languages or dialects. The method may include additional actions of generating, based on output of the neural network, a speaker representation and determining, based on the speaker representation and a second representation, that the utterance is an utterance of the user. The method may provide the user with access to the user device based on determining that the utterance is an utterance of the user.
Speaker verification
Methods, systems, apparatus, including computer programs encoded on computer storage medium, to facilitate language independent-speaker verification. In one aspect, a method includes actions of receiving, by a user device, audio data representing an utterance of a user. Other actions may include providing, to a neural network stored on the user device, input data derived from the audio data and a language identifier. The neural network may be trained using speech data representing speech in different languages or dialects. The method may include additional actions of generating, based on output of the neural network, a speaker representation and determining, based on the speaker representation and a second representation, that the utterance is an utterance of the user. The method may provide the user with access to the user device based on determining that the utterance is an utterance of the user.