Patent classifications
G06F16/634
Processing Voice Commands
Recorded background noises, and other contextual data, may be used to assist in resolving ambiguity in spoken voice commands. The background noises may comprise sounds from entities in a room other than the user issuing the voice commands. One such entity may be a content item being watched by the user, and the captured background noises may comprise audio of the content item. The content item may be identified based on the captured audio of the content item in the background noises, and the identification may be used to interpret the ambiguous voice command. Additional contextual information associated with the voice commands (e.g., identifications of the users in the room) and/or the content item (e.g., the video quality of the content item, a service outputting the content item, a genre of the content item, etc.) may be used to identify the content item.
Collaborative artificial intelligence method and system
A method and system of audibly broadcasting responses to a user based on user queries about a specific patient molecular report, the method comprising receiving an audible query from the user to a microphone coupled to a collaboration device, identifying at least one intent associated with the audible query, identifying at least one data operation associated with the at least one intent, associating each of the at least one data operations with a first set of data presented on the molecular report, executing each of the at least one data operations on a second set of data to generate response data, generating an audible response file associated with the response data and providing the audible response file for broadcasting via a speaker coupled to the collaboration device.
Intelligent Interactive Voice Recognition System
Systems for performing intelligent interactive voice recognition functions are provided. In some aspects, natural language data may be received from a plurality of users. The natural language data may be used to generate a plurality of user-specific machine learning datasets. Subsequent natural language input data including a user query may be received. The query may be analyzed to identify the user and a user-specific machine learning dataset associated with the user may be identified. The natural language data may be processed to remove noise associated with the data and may be further processed using the identified user-specific machine learning dataset to interpret the query of the user and generate an output. The output may be transmitted to the user and feedback data may be received from the user. The user-specific machine learning dataset may then be validated and/or updated based on the feedback data.
Expansion of a question and answer database
A system and method for expanding a question and answer (Q&A) database. The method includes obtaining a set of Q&A documents and speech recognition results, each Q&A document in the set having an identifier, and each speech recognition result having an identifier common with the identifier of a relevant Q&A document, and adding one or more repetition parts extracted from the speech recognition results to a corresponding Q&A document in the set to generate an expanded set of Q&A documents for increasing Q&A document extraction accuracy.
SYSTEMS AND METHODS FOR REMOTELY INTERACTING WITH PERFORMERS AND INFLUENCING LIVE EVENTS
A computer-implemented method of remotely influencing a performer at a live event via a customer mobile device is disclosed herein. The method includes: displaying a graphical user interface configured to receive user inputs; receiving a first user input including a user request for the performer at the live event; presenting predetermined terms and conditions associated with the user request; receiving a second user input including a user acceptance of the terms and conditions associated with the user request; transmitting the user request to a host server upon receiving the user acceptance of the terms and conditions associated with the user request; receiving a confirmation of the terms and conditions associated with the user request from the host server; and transmitting the user request for receipt by a performer mobile device of the performer during the live event.
AUDIO RECOGNITION METHOD, ELECTRONIC DEVICE AND STORAGE MEDIUM
A method includes obtaining a query content. The query content includes segment information representing a to-be-recognized audio. The method further includes selecting the preset quantity of candidate audios corresponding to the query content from a preset library. Each candidate audio includes a candidate audio segment matched with the segment information. The method further includes inputting the candidate audio segment into a trained detection model so as to obtain target segment information including the segment information and a target audio where the target segment information is located.
Guidance query for cache system
A device may be configured to determine whether an audio file is a first type of audio file that is capable of being processed to recognize the voice query based on a characteristic of the audio file itself or a second type of audio file that may require speech recognition processing in order to recognize the voice query associated with the audio file. In determining whether the audio file is a first type of audio file or a second type of audio file, a query filter associated with the device may be configured to access one or more guidance queries. Using the one or more guidance queries, the device may classify the audio file as a first type of audio file or a second type of audio file based on receiving only a portion of the audio file, thereby improving the speed at which the audio file can be processed.
Music streaming, playlist creation and streaming architecture
A system and method for making categorized music tracks available to end user applications. The tracks may be categorized based on computer-derived rhythm, texture and pitch (RTP) scores for tracks derived from high-level acoustic attributes, which is based on low level data extracted from the tracks. RTP scores are stored in a universal database common to all of the music publishers so that the same track, once RTP scored, does not need to be re-RTP scored by other music publishers. End user applications access an API server to import collections of tracks published by publishers, to create playlists and initiate music streaming. Each end user application is sponsored by a single music publisher so that only tracks capable of being streamed by the music publisher are available to the sponsored end user application.
Audio request interaction system
A person can use a portable electronic device to electronically purchase or otherwise request a product, service or other deliverable related to audio programming to which the person is listening at the time they initiate the request. The request is fulfilled by a service that analyzes the audio content to identify the deliverable the person desires.
Tagging an Image with Audio-Related Metadata
In one aspect, an example method to be performed by a computing device includes (a) receiving a request to use a camera of the computing device; (b) in response to receiving the request, (i) using a microphone of the computing device to capture audio content and (ii) using the camera of the computing device to capture an image; (c) identifying reference audio content that has at least a threshold extent of similarity with the captured audio content; and (d) outputting an indication of the identified reference audio content while displaying the captured image.