Patent classifications
G10L2015/226
Remote Control Device with Environment Mapping
A remote control device for controlling devices in an environment can utilize an environment map and location information to accurately determine an intended device to provide control for multiple devices in an environment. The environment mapping can be performed using the remote control device including a plurality of sensors. A spatial map can be generated for an environment along with location information for controllable devices within the environment. The spatial map and location information can be stored on the remote control device. The mapping can allow the remote control device to quickly group devices or drag and drop content from one type of device to another type of device. The remote control device can perform search queries based on combinations of image and audio data in some examples.
System and method for recognizing user's speech
Provided is a system and method for recognizing a user's speech. A method, performed by a server, of providing a text string for a speech signal input to a device includes: receiving, from the device, an encoder output value derived from an encoder of an end-to-end automatic speech recognition (ASR) model included in the device; identifying a domain corresponding to the received encoder output value; selecting a decoder corresponding to the identified domain from among a plurality of decoders of an end-to-end ASR model included in the server; obtaining a text string from the received encoder output value using the selected decoder; and providing the obtained text string to the device.
ELECTRONIC APPARATUS, CONTROLLING METHOD OF ELECTRONIC APPARATUS AND SERVER
An electronic apparatus which registers a device to a server by using a voice, and a method therefor are provided. The electronic apparatus includes a communication circuit, a microphone, a memory for storing computer executable instructions, and at least one processor configured to execute the computer executable instructions to acquire, from a voice received through the microphone, information on an external device which a user wishes to register, based on an external device corresponding to the acquired information being searched through the communication circuit, control the communication circuit to transmit information on an access point to the external device to enable the external device to communicate with a server, and control the communication circuit to transmit a registration request with respect to the external device to the server.
EFFICIENT COLLABORATION USING A VIRTUAL ASSISTANT
In an approach to assisting users of a collaborative meeting platform, one or more computer processors detect a start of a collaborative meeting. One or more computer processors monitor one or more activities of the collaborative meeting. Based on the one or more activities of the collaborative meeting, one or more computer processors detect a trigger for assistance with a user interface of the collaborative meeting. One or more computer processors retrieve one or more correlated actions associated with the trigger. One or more computer processors perform at least one of the one or more retrieved correlated action within the user interface of the collaborative meeting.
INTENT RECOGNITION METHOD AND INTENT RECOGNITION SYSTEM HAVING SELF LEARNING CAPABILITY
An intent recognition method having a self-learning capability includes the following steps: acquiring a user expression, and recognizing a voice as a corresponding text; performing preliminary intent recognition on the user expression, and outputting candidate intents; acquiring historical data feature parameters of the candidate intents; on the basis of a pre-set rule strategy, deciding whether to directly output a final recognized intent, and on the basis of the feature parameters of each intent, performing rule computation, and outputting a final recognized intent; submitting prediction data of the final recognized intent and the candidate intents from the intent recognition process to a self-learning system, and performing self learning and indicator parameter data updating. The present disclosure is able to perform self learning on the basis of the feature distribution in historical data of intent recognition and dynamically adjust intent recognition strategies.
DYNAMIC SPEECH RECOGNITION METHODS AND SYSTEMS WITH USER-CONFIGURABLE PERFORMANCE
Methods and systems are provided for assisting operation of a vehicle using speech recognition. One method involves identifying a user-configured speech recognition performance setting value selected from among a plurality of speech recognition performance setting values, selecting a speech recognition model configuration corresponding to the user-configured speech recognition performance setting value from among a plurality of speech recognition model configurations, where each speech recognition model configuration of the plurality of speech recognition model configurations corresponds to a respective one of the plurality of speech recognition performance setting values, and recognizing an audio input as an input state using the speech recognition model configuration corresponding to the user-configured speech recognition performance setting value.
Voice context-aware content manipulation
Techniques performed by a data processing system for processing voice content received from a user herein include receiving a first audio input from a user comprising spoken content, analyzing the first audio input using one or more natural language processing models to produce a first textual output comprising a textual representation of the first audio input, analyzing the first textual output using one or more machine learning models to determine first context information of the first textual output, and processing the first textual output in the application based on the first context information.
Method and apparatus for recognizing a voice
Disclosed are a speech recognition method and a speech recognition device, in which speech recognition is performed by executing an artificial intelligence (AI) algorithm and/or a machine learning algorithm provided therein. According to an embodiment of the present disclosure, the speech recognition method includes buffering an inputted spoken utterance, determining whether a preset wake-up word is present in the spoken utterance by comparing the buffered spoken utterance to the preset wake-up word, and in response to the preset wake-up word in the spoken utterance, activating a speech recognition function and isolating, from the spoken utterance, a spoken sentence as a voice command without the wake-up word, and processing the spoken sentence and outputting a processing result.
SPEECH-ENABLED AUGMENTED REALITY
Methods and systems for implementing an intuitive interaction between the user and the virtual content of augmented reality applications are disclosed. By implementing an augmented reality inquiry mode of a device, the system can enable a user to interact with relevant virtual objects via a speech-enabled interface. The speech-enabled augmented reality system can identify visual objects in images and recognize virtual objects corresponding to the visual objects, determine one or more relevant objects from the virtual objects based on relevance factors. Once the interaction session is established, a user can further interact with the relevant virtual objects, notably through voice commands addressed to the object. Accordingly, the present subject matter can enable a natural and hands-free interaction between the user and any virtual object that the user is interested in.
Interactive voice-control method and apparatus, device and medium
The present disclosure discloses an interactive voice-control method and apparatus, a device and a medium. The method includes: obtaining a sound signal at a voice interaction device and recognized information that is recognized from the sound signal; determining an interaction confidence of the sound signal based at least on at least one of an acoustic feature representation of the sound signal and a semantic feature representation associated with the recognized information; determining a matching status between the recognized information and the sound signal; and providing the interaction confidence and the matching status for controlling a response of the voice interaction device to the sound signal.