Patent classifications
G06F3/16
Messaging content data storage
Techniques for determining when to store content and when to refrain from storing content are described herein. In some instances, devices exchange communications that include different types of content, such as text, audio data, video data, image data, or the like. For instance, a first device may receive, from a second device, a communication that includes audio data representing speech of a user of the second device, along with text for display on the first device. The text may comprise a transcription of the audio file, additional commentary provided by the user of the second device, or the like. Upon receiving the communication that includes text and audio data, the first device may determine whether or not to store the audio data. For instance, the first device may determine whether it currently stores this audio content. If so, then the first device may refrain from storing the content again.
Audio analytics and accessibility across applications and platforms
Systems and methods for audio analytics and audio-based accessibility are provided. A user audio profile may be stored in memory for a user. The user audio profile may include a custom prioritization of one or more audio parameters associated with one or more audio modifications. Audio streams associated with a user device of the user may be monitored based on the user audio profile during a current session. The audio parameters may be detected as being present in the monitored audio streams, and the detected audio parameters may be prioritized based on the custom prioritization of the user audio profile. A sound property of at least one of the audio streams may be modified in real-time based on the prioritization of the detected audio parameters by applying the audio modifications of the user audio profile to the at least one audio stream before the at least one audio stream is provided to the user device.
System and method for an augmented reality goal assistant
A method for an augmented reality goal assistant is described. The method includes detecting an object associated with a behavioral goal of a user. The method also includes altering an appearance of the object based on the behavioral goal of the user. The method further includes displaying the altered appearance of a detected object on an augmented reality headset, such that the altered appearance of the detected object is modified based on the behavioral goal of the user.
Task resumption in a natural understanding system
A speech-processing system may provide access to one or more skills via spoken commands and/or responses in the form of synthesized speech. The system may be capable of keeping one or more skills active in the background while a user interacts (e.g., provides inputs to and/or receives outputs from) with a skill running in the foreground. A background skill may receive some trigger data, and determine to request the system to return the background skill to the foreground to, for example, request a user input regarding an action previously requested by the user. In some cases, the user may invoke a background skill to continue a previous interaction. The system may return the background skill to the foreground. The resumed skill may continue a previous interaction to, for example, to query the user for instructions, provide an update or alert, or continue a previous output.
Selective information provision and indoor navigation assistance for the visually impaired
A software application and system may be configured to enable a smartphone or other device to be used by a visually impaired person to receive voice navigation guidance during a directed exploration of an area. Directed exploration uses combinations of location data, directional data, and orientation data from the configured device to determine a direction that user wishes to explore, and only providing narrated results for streets, businesses, and other points of interest in that direction. The system may also utilize sets of wireless indicators positioned within indoor areas to provide accurate positioning to particular locations and floors within buildings.
Selective information provision and indoor navigation assistance for the visually impaired
A software application and system may be configured to enable a smartphone or other device to be used by a visually impaired person to receive voice navigation guidance during a directed exploration of an area. Directed exploration uses combinations of location data, directional data, and orientation data from the configured device to determine a direction that user wishes to explore, and only providing narrated results for streets, businesses, and other points of interest in that direction. The system may also utilize sets of wireless indicators positioned within indoor areas to provide accurate positioning to particular locations and floors within buildings.
Method and apparatus for waking up device, electronic device, and storage medium
A method and apparatus for waking up a device, an electronic device, and a storage medium are provided, which are related to fields of image processing and deep learning. The method includes: acquiring an environment image of a surrounding environment of a target device in real time, and recognizing a face region of a user in the environment image; acquiring a plurality of facial landmarks in the face region, and acquiring a left eye image and a right eye image according to the facial landmarks; acquiring a left eye sight classification result and a right eye sight classification result according to the left eye image and the right eye image; and waking up the target device in a case of determining that the user is looking at the target device according to the left eye sight classification result and the right eye sight classification result.
Whole-body human-computer interface
A human-computer interface system having an exoskeleton including a plurality of structural members coupled to one another by at least one articulation configured to apply a force to a body segment of a user, the exoskeleton comprising a body-borne portion and a point-of-use portion; the body-borne portion configured to be operatively coupled to the point-of-use portion; and at least one locomotor module including at least one actuator configured to actuate the at least one articulation, the at least one actuator being in operative communication with the exoskeleton.
Method and apparatus for evaluating user intention understanding satisfaction, electronic device and storage medium
A method and apparatus for generating a user intention understanding satisfaction evaluation model, a method and apparatus for evaluating a user intention understanding satisfaction, an electronic device and a storage medium are provided, relating to intelligent voice recognition and knowledge graphs. The method for generating a user intention understanding satisfaction evaluation model is: acquiring a plurality of sets of intention understanding data, at least one set of which comprises a plurality of sequences corresponding to multi-round behaviors of an intelligent device in multi-round man-machine interactions; and learning the plurality of sets of intention understanding data through a first machine learning model, to obtain the user intention understanding satisfaction evaluation model after the learning, wherein the user intention understanding satisfaction evaluation model is configured to evaluate user intention understanding satisfactions of the intelligent device in the multi-round man-machine interactions according to the plurality of sequences corresponding to the multi-round man-machine interactions.
System for deliverables versioning in audio mastering
Some implementations of the disclosure relate to using a model trained on mixing console data of sound mixes to automate the process of sound mix creation. In one implementation, a non-transitory computer-readable medium has executable instructions stored thereon that, when executed by a processor, causes the processor to perform operations comprising: obtaining a first version of a sound mix; extracting first audio features from the first version of the sound mix obtaining mixing metadata; automatically calculating with a trained model, using at least the mixing metadata and the first audio features, mixing console features; and deriving a second version of the sound mix using at least the mixing console features calculated by the trained model.