G10L2015/226

System and method for federated, context-sensitive, adaptive language models
11430432 · 2022-08-30 · ·

A system and method for federated context-sensitive language models comprising a federated language model server and a plurality of edge devices. The federated language model server may comprise one or more machine learning models trained and developed centrally on the server and distribute these one or more machine learning models to edge devices wherein they may be operated locally on the edge devices. The edge devices may gather or generate context data that can be used by a speech recognition engine, and the local language models contained therein, to develop adaptive, context-sensitive, user-specific language models. Periodically, the federated language model server may select a subset of edge devices from which to receive uploaded local model parameters, that may be aggregated to perform central model updates wherein the updated model parameters may then be sent back to edge devices in order to update the local model parameters.

INFORMATION PROCESSING DEVICE, METHOD OF INFORMATION PROCESSING, AND PROGRAM

There is provided an information processing device technology that enables an improvement in precision of sound recognition processing based on collected sound information, the information processing device including: a recognition controller that causes a speech recognition processing portion to execute sound recognition processing based on collected sound information obtained by a sound collecting portion; and an output controller that generates an output signal to output a recognition result obtained through the sound recognition processing. The output controller causes an output portion to output an evaluation result regarding a type of sound based on the collected sound information prior to the recognition result.

VOICE ASSISTANT-ENABLED CLIENT APPLICATION WITH USER VIEW CONTEXT

Various embodiments discussed herein enable client applications to be heavily integrated with a voice assistant in order to both perform commands associated with voice utterances of users via voice assistant functionality and also seamlessly cause client applications to automatically perform native functions as part of executing the voice utterance. For example, some embodiments can automatically and intelligently cause a switch to a page the user needs and automatically and intelligently cause a population of particular fields of the page the user needs based on a user view context and the voice utterance.

Identifying a location of a voice-input device
11455994 · 2022-09-27 · ·

Techniques for identifying a location of a voice-controlled device within an environment. After identifying a location of the device, the device may receive a voice command from a user within the environment and may determine a response to the command based in part on the location, may determine how to output a response based in part on the location or may determine how to interact with the user based in part on the location.

Information processing device and information processing method
11237794 · 2022-02-01 · ·

An information processing device and information processing method capable of outputting an action based on an intention of the user. The information processing device including an action deciding unit that determines an action for a user on a basis of a distance from the user and an output control unit that outputs the action.

System and method for conversational agent via adaptive caching of dialogue tree
11455986 · 2022-09-27 · ·

The present teaching relates to method, system, medium, and implementations for managing a user machine dialogue. Sensor data is received at a device, including an utterance representing a speech of a user engaged in a dialogue with the device. The speech of the user is determined based on the utterance and a response to the user is searched by a local dialogue manager residing on the device against a sub-dialogue tree stored on the device. The response, if identified from the sub-dialogue tree, is rendered to the user in response to the speech. A request is sent to a server for the response, if the response is not available in the sub-dialogue tree.

Contextual utterance resolution in multimodal systems

A system and method of responding to a vocal utterance may include capturing and converting the utterance to word(s) using a language processing method, such as natural language processing. The context of the utterance and of the system, which may include multimodal inputs, may be used to determine the meaning and intent of the words.

Information processing device and information processing method

[Problem] More effective information is gathered efficiently for the correction of device behavior. [Solution] Provided is an information processing device that includes an output control unit that controls, on the basis of a gathered operation history of a device, an output of a reproduced scene pertaining to a behavior which the device has executed on the basis of context information; and a communication unit that receives feedback input by a judge regarding the reproduced scene thus output. Further provided is an information processing device that comprises a communication unit that receives information on a reproduced scene pertaining to a behavior which the device has executed on the basis of context information; and a playback unit that plays back the reproduced scene, wherein the communication unit transmits feedback input by a judge regarding the reproduced scene.

Electronic apparatus and control method thereof

An electronic apparatus includes a communicator configured to communicate with a plurality of external apparatus. A storage is configured to store situation information. A processor is configured to, based on a first utterance of a user, control a first operation corresponding to the first utterance to be carried out from among a plurality of operations related to the plurality of external apparatuses. Situation information corresponding to each of a plurality of situations where the first operation is carried out based on the first utterance is stored in the storage. Based on a second utterance of the user, a second operation is identified corresponding to the second utterance from among the plurality of operations based on the stored situation information, and the identified second operation is carried out.

MALWARE-PROOF PRIVACY INDICATOR

A voice command device (VCD) has privacy protection. The VCD comprises a processor, first and second input devices, at least one data line to couple the first and second input devices to the processor, a power supply, and a sensor power line to couple the first and second input devices to the power supply. The VCD also comprises a manually operated mechanical switch on the sensor power line, to divide the sensor power line into a first leg comprising the power supply and a second leg comprising the input devices. The VCD also comprises an active sensor indicator light on the second leg of the sensor power line. The indicator light is configured to indicate whether the input devices are operational, based on a power level of the second leg of the sensor power line. Other embodiments are described and claimed.