Patent classifications
G10H2240/251
Audiovisual content rendering with display animation suggestive of geolocation at which content was previously rendered
Techniques have been developed to facilitate (1) the capture and pitch correction of vocal performances on handheld or other portable computing devices and (2) the mixing of such pitch-corrected vocal performances with backing tracks for audible rendering on targets that include such portable computing devices and as well as desktops, workstations, gaming stations, even telephony targets. Implementations of the described techniques employ signal processing techniques and allocations of system functionality that are suitable given the generally limited capabilities of such handheld or portable computing devices and that facilitate efficient encoding and communication of the pitch-corrected vocal performances (or precursors or derivatives thereof) via wireless and/or wired bandwidth-limited networks for rendering on portable computing devices or other targets.
Template-Based Excerpting and Rendering of Multimedia Performance
Disclosed herein are computer-implemented method, system, and computer-readable storage-medium embodiments for implementing template-based excerpting and rendering of multimedia performances technologies. An embodiment includes at least one computer processor configured to retrieve a first content instance and corresponding first metadata. The first content instance may include a first plurality of structural elements, with at least one structural element corresponding to at least part of the first metadata. The first content instance may be transformed by a rendering engine running on the at least one computer processor and/or transmitted to a content-playback device.
Audio-visual effects system for augmentation of captured performance based on content thereof
Visual effects schedules are applied to audiovisual performances with differing visual effects applied in correspondence with differing elements of musical structure. Segmentation techniques applied to one or more audio tracks (e.g., vocal or backing tracks) are used to compute some of the components of the musical structure. In some cases, applied visual effects schedules are mood-denominated and may be selected by a performer as a component of his or her visual expression or determined from an audiovisual performance using machine learning techniques.
Mobile machine
A system for providing mobile content to a mobile communication device includes a first computing system including one or more servers to cause a graphical user interface to be displayed at a second computing system, the graphical user interface (i) enabling a user of the second computing system to at least one of create, edit, or select the mobile content and (ii) enabling the user of the second computing system to provide a phone number associated with the mobile communication device, the mobile communication device being separate and remote from the first computing system and the second computing system. The first computing system uses the phone number to cause delivery of the mobile content to the mobile communication device via a wireless communications network in a format compatible with one or more operational parameters of the mobile communication device, the one or more operational parameters including at least one of a mobile communication device type and a software platform type, wherein the using of the phone number to cause the delivery of the mobile content to the mobile communication device is performed without the mobile communication device identifying the one or more operational parameters to the first computing system or the second computing system.
COORDINATING AND MIXING AUDIOVISUAL CONTENT CAPTURED FROM GEOGRAPHICALLY DISTRIBUTED PERFORMERS
Audiovisual performances, including vocal music, are captured and coordinated with those of other users in ways that create compelling user experiences. In some cases, the vocal performances of individual users are captured (together with performance synchronized video) on mobile devices, television-type display and/or set-top box equipment in the context of karaoke-style presentations of lyrics in correspondence with audible renderings of a backing track. Contributions of multiple vocalists are coordinated and mixed in a manner that selects for visually prominent presentation performance synchronized video of one or more of the contributors. Prominence of particular performance synchronized video may be based, at least in part, on computationally-defined audio features extracted from (or computed over) captured vocal audio. Over the course of a coordinated audiovisual performance timeline, these computationally-defined audio features are selective for performance synchronized video of one or more of the contributing vocalists.
Social music system and method with continuous, real-time pitch correction of vocal performance and dry vocal capture for subsequent re-rendering based on selectively applicable vocal effect(s) schedule(s)
Embodiments described provide a method for mixing vocal performances from different vocalists. A vocal score temporally synchronized with a corresponding backing track and lyrics is retrieved via a communications interface of a portable computing device. A first vocal performance of a user is captured, via a microphone interface of the portable computing device, and in correspondence with the backing track. An open call indication for soliciting, from a second vocalist, a second vocal performance to be mixed for audible rendering with the first vocal performance is transmitted. A mix to one of the user and the second vocalist is provided by selecting, based on to whom the mix is provided, the mix from alternative mixes each having a different prominent vocal performance.
RECORDING METHOD OF TRUE WIRELESS STEREO EARBUDS AND RECORDING SYSTEM
In this application, by using a method for collaborative recording by using a mobile terminal and the true wireless stereo earbud, recording with more complete data amount can be collected, and cooperative noise reduction is performed on a recording signal collected by the mobile terminal and a recording signal collected by the true wireless stereo earbud, so that accuracy and integrity of a human voice signal obtained through separation are higher. Further, in this application, when the true wireless stereo earbud performs recoding in a mode of single-earbud recording, a data amount of the true wireless stereo earbud is relatively small. Therefore, when the mobile terminal receives, through Bluetooth, data sent by the true wireless stereo earbud, excessive occupation of Bluetooth bandwidth of the mobile terminal is avoided, and power consumption of a mobile phone is reduced.
METHODS AND SYSTEMS FOR SYNCHRONIZING AN AUDIO CLIP EXTRACTED FROM AN ORIGINAL RECORDING WITH CORRESPONDING LYRICS
Methods, systems, and devices for determining an audio portion based on a request received from a consumer or user of an operating device where the requests comprise a set of lyrics, then effect the streaming of the determined audio portion.
Mobile terminal and ringtone playing method thereof
A mobile terminal and a ringtone playing method thereof are provided. The playing method includes: selecting a music file which to be set as a ringtone, and popping up a playing preview interface corresponding to the music film and, meanwhile, playing the music file; sliding a cursor of the playing preview interface to select a start position and a playing end position on a playing progress bar corresponding to the music file so as to acquire and save playing information of the ringtone, wherein the playing information comprises a playing start time point and a playing end time point of the ringtone; and in response to occurrence of a ringing event, reading the saved playing information of the ringtone and playing a music clip between the playing start time point and the playing end time point.
Networks of portable electronic devices that collectively generate sound
Systems and methods for collectively generating sound are provided. A sound composition can include multiple component pieces and those component pieces can be divided amongst devices so that the devices can collectively generate the sound composition. For example, each component piece may be assigned to a different device and the each device may provide its assigned component pieces in a synchronized manner so that the overall sound composition is collectively generated. By generating a sound composition in this manner, the network may create new and exciting spatial effects.