Patent classifications
H04N21/2335
METHODS AND SYSTEMS FOR IMPLEMENTING AN ELASTIC CLOUD BASED VOICE SEARCH USING A THIRD-PARTY SEARCH PROVIDER
Systems and methods are described to provide voice search in an elastic cloud environment communicating with a set-top box (STB) by receiving by a voice cloud search server pulse-code modulation (PCM) audio packets transmitted from the STB; sending the PCM audio packets to a natural language processing (NLP) service for converting to text; sending the text sets to an elastic voice cloud search server for querying an electronic program guide (EPG) service, channel and program data associated with the text set wherein the EPG service to at least return identified channel and program data; in response to an identified return of channel and television program data, sending sets of text to a third-party search service for performing an independent search for related data and returning search results of video and image content which is then stripped of dynamic scripts to return to the STB.
METHOD AND SYSTEM FOR NAVIGATING AT A CLIENT DEVICE SELECTED FEATURES ON A NON-DYNAMIC IMAGE PAGE FROM AN ELASTIC VOICE CLOUD SERVER IN COMMUNICATION WITH A THIRD-PARTY SEARCH SERVICE
Systems, and methods are described for navigating non-dynamic pages at a client device served up from voice search requests by selecting a feature on the image page by a user request for navigating the feature viewed on the image page wherein the feature selected has an appearance of a dynamic feature when served up on the display wherein the user request selects a location on the image page; identifying the feature by the voice cloud server by location data of a selected location on the page; executing a feature request on a corresponding page which contains dynamic script at the voice cloud server to receive a linked page from a third-party search service; and removing dynamic scripts, at the voice cloud search server, of the linked page for serving up as an image page of the linked page at the client device.
Simultaneous recording and uploading of multiple audio files of the same conversation and audio drift normalization systems and methods
The invention relates to audio drift normalization, and more particularly to audio drift normalization systems and methods that can normalize audio drift of a plurality of recordings from a source.
Systems and methods for tailoring media encoding to playback environments
The disclosed system may include various modules that are stored in memory and executed by a physical processor. The disclosed system may (1) receive a request to stream presentation data to a computing device, (2) determine at least one characteristic of the computing device in connection with the request, (3) identify a compressed version of the presentation data that corresponds to the characteristic of the computing device, and then (4) stream the compressed version of the presentation data to the computing device to facilitate presentation of the presentation data to a user of the computing device. Various other systems, devices, and methods are also disclosed.
Augmenting Content Items
Systems, apparatuses, and methods are described for adding visual and/or sound effects to a content item to improve user experience. Addition of the visual and/or sound effects may be based on metadata associated with the content item.
Loudness adjustment method and apparatus, and electronic device and storage medium
The present application relates to a loudness adjustment method and apparatus, and an electronic device and a storage medium, mainly relating to the technical field of multimedia. The method comprises: converting a sound signal of a multimedia resource into a first frequency domain signal; acquiring a second frequency domain signal based on the first frequency domain signal and frequency response information of a current electronic device, wherein the second frequency domain signal is used for reflecting the loudness of the first frequency domain signal when same is played on the current electronic device; and adjusting the loudness of the sound signal based on the second frequency domain signal and a target loudness, so as to obtain a target sound signal of the multimedia resource.
SYSTEM AND METHOD FOR PROVIDING INTERACTIVE STORYTELLING
A system for providing interactive storytelling includes an output device configured to output storytelling content to a user, wherein the storytelling content includes one or more of audio data or visual data, a playback controller configured to provide storytelling content to the output device, one or more sensors configured to generate measurement data by capturing an action of the user, an abstraction device configured to generate extracted characteristics by analyzing the measurement data, an action recognition device configured to determine a recognized action by analyzing a time behavior of the measurement data and/or the extracted characteristics. The playback controller is additionally configured to interrupt provision of storytelling content, to trigger the abstraction device and/or the action recognition device to determine a recognized action, and to continue provision of storytelling content based on the recognized action. A corresponding method, a computer program product, and a computer-readable storage medium are also disclosed.
Digital Data Processing
A media system and a method of operating a media system are provided. The media system includes a processing device and a media receiving device. The processing device is configured to transcode a digital data (e.g., at least one of: an audio component and a video component) from a first format to a second format. The processing device is configured to transmit the digital data in the second format. The media receiving device is in communication with the processing device. The media receiving device has at least one of: an obscured mode and a mute mode, and conversely at least one of an un-obscured mode and an un-mute mode. The processing device is configured to stop transcoding and transmitting at least a portion of the digital data to the media receiving device when the media receiving device is in at least one of the obscured mode and the mute mode.
METHOD AND APPARATUS FOR DETERMINING AGGREGATE SENTIMENTS
An example method includes obtaining, from a plurality of audience members that are recipients of media content, sentiment data indicating reactions of the plurality of audience members to the media content. The media content includes at least one of audio and video, and the sentiment data includes commentary from the recipients regarding the media content or physical reactions of the audience members to the media content. At least one aggregate sentiment of the plurality of audience members is determined based on the sentiment data. At least one the following is performed: transmitting an indication of the at least one aggregate sentiment to a computing device, and augmenting the media content to indicate the at least one aggregate sentiment to the audience members.
SYSTEMS AND METHODS FOR APPLYING STYLE TRANSFER FUNCTIONS IN MULTI-CAMERA SYSTEMS AND MULTI-MICROPHONE SYSTEMS
Systems and methods for applying style transfer functions in multi-camera systems and multi-microphone systems are disclosed herein. An example multi-camera style transfer system includes at least one memory, instructions in the system, and processor circuitry to execute the instructions to at least apply a style transfer function to a second image from a second video feed to generate a stylized image based on a reference image. The reference image corresponds to a first image from a first video feed. The first video feed is from a first camera. The second video feed is from a second camera. The processor circuitry also executed the instructions to adjust one or more parameter settings of a video filter based on the stylized image, and filter the second video feed using the video filter with the adjusted parameter settings to generate a filtered version of the second video feed.