G10L25/51

MULTI-USER VOICE ASSISTANT WITH DISAMBIGUATION

Disambiguating question answering responses by receiving voice command data associated with a first user, determining a first user identity according to the first user voice command data, determining a first user activity context according to the first user voice command data, determining a first response for the first user, receiving voice command data associated with a second user, determining a second user identity according to the second user voice command data, determining a second user activity context according to the second user voice command data, determining a second response for the second user, determining a predicted ambiguity between the first response and the second response, altering the first response according to the predicted ambiguity, and providing the first response and the second response.

Home sound loacalization and identification
11582554 · 2023-02-14 · ·

A system for sound localization can include a first electronic device having a microphone to detect a sound, and a second electronic device. A processor can be in communication with the first electronic device and the second electronic device. The processor can receive a first signal from the first electronic device corresponding to the detected sound, determine a location of origin of the detected sound based at least in part on the first signal, and provide a second signal to the second electronic device based at least in part on the location of origin.

Home sound loacalization and identification
11582554 · 2023-02-14 · ·

A system for sound localization can include a first electronic device having a microphone to detect a sound, and a second electronic device. A processor can be in communication with the first electronic device and the second electronic device. The processor can receive a first signal from the first electronic device corresponding to the detected sound, determine a location of origin of the detected sound based at least in part on the first signal, and provide a second signal to the second electronic device based at least in part on the location of origin.

Altering undesirable communication data for communication sessions

This disclosure describes techniques implemented partly by a communications service for identifying and altering undesirable portions of communication data, such as audio data and video data, from a communication session between computing devices. For example, the communications service may monitor the communications session to alter or remove undesirable audio data, such as a dog barking, a doorbell ringing, etc., and/or video data, such as rude gestures, inappropriate facial expressions, etc. The communications service may stream the communication data for the communication session partly through managed servers and analyze the communication data to detect undesirable portions. The communications service may alter or remove the portions of communication data received from a first user device, such as by filtering, refraining from transmitting, or modifying the undesirable portions. The communications service may send the modified communication data to a second user device engaged in the communication session after removing the undesirable portions.

Drone assisted setup for building specific sound localization model
11581010 · 2023-02-14 · ·

Techniques and systems are described for generating and using a sound localization model. A described technique includes obtaining for a building a sound sensor map indicating locations of first and second sound sensor devices in respective first and second rooms of the building; causing an autonomous device to navigate to the first room and to emit, during a time window, sound patterns at one or more frequencies within the first room; receiving sound data including first and second sound data respectively from the first and second sound sensor devices that are observed during the time window; and generating and storing a sound localization model based on the sound sensor map, autonomous device location information, and the received sound data, the model being configured to compensate for how sounds travels among rooms in at least a portion of the building such that an origin room of a sound source is identifiable.

Drone assisted setup for building specific sound localization model
11581010 · 2023-02-14 · ·

Techniques and systems are described for generating and using a sound localization model. A described technique includes obtaining for a building a sound sensor map indicating locations of first and second sound sensor devices in respective first and second rooms of the building; causing an autonomous device to navigate to the first room and to emit, during a time window, sound patterns at one or more frequencies within the first room; receiving sound data including first and second sound data respectively from the first and second sound sensor devices that are observed during the time window; and generating and storing a sound localization model based on the sound sensor map, autonomous device location information, and the received sound data, the model being configured to compensate for how sounds travels among rooms in at least a portion of the building such that an origin room of a sound source is identifiable.

AR-based supplementary teaching system for guzheng and method thereof

An AR-based supplementary teaching system for guzheng and method thereof, the system includes an AR device, a data processing device and positioning devices for key positions, the data processing device is signal-connected to the AR device, and the positioning devices is installed on the guzheng code of guzheng, the positioning devices corresponds to the guzheng code of guzheng one by one; the AR device is used to obtain real scene data; the data processing device is used to guzheng and the positioning devices identify and generate string distribution data; also used to obtain operation instruction based on user actions, execute the operation instruction and generate virtual data; the AR device is also used to convert all data based on the string distribution data The virtual data and the real scene data are superimposed and displayed.

AR-based supplementary teaching system for guzheng and method thereof

An AR-based supplementary teaching system for guzheng and method thereof, the system includes an AR device, a data processing device and positioning devices for key positions, the data processing device is signal-connected to the AR device, and the positioning devices is installed on the guzheng code of guzheng, the positioning devices corresponds to the guzheng code of guzheng one by one; the AR device is used to obtain real scene data; the data processing device is used to guzheng and the positioning devices identify and generate string distribution data; also used to obtain operation instruction based on user actions, execute the operation instruction and generate virtual data; the AR device is also used to convert all data based on the string distribution data The virtual data and the real scene data are superimposed and displayed.

Audio playout report for ride-sharing session
11581011 · 2023-02-14 · ·

In one aspect, an example method to be performed by a computing device includes (a) determining that a ride-sharing session is active; (b) in response to determining the ride-sharing session is active, using a microphone of the computing device to capture audio content; (c) identifying reference audio content that has at least a threshold extent of similarity with the captured audio content; (d) determining that the ride-sharing session is inactive; and (e) outputting an indication of the identified reference audio content.

Audio playout report for ride-sharing session
11581011 · 2023-02-14 · ·

In one aspect, an example method to be performed by a computing device includes (a) determining that a ride-sharing session is active; (b) in response to determining the ride-sharing session is active, using a microphone of the computing device to capture audio content; (c) identifying reference audio content that has at least a threshold extent of similarity with the captured audio content; (d) determining that the ride-sharing session is inactive; and (e) outputting an indication of the identified reference audio content.