SYSTEMS AND METHODS FOR DYNAMICALLY SYNTHESIZING AUDIO FILES ON A MOBILE DEVICE
20220326906 · 2022-10-13
Inventors
Cpc classification
G10L15/22
PHYSICS
G10L19/018
PHYSICS
G10H2240/131
PHYSICS
G10H1/0025
PHYSICS
G06F3/167
PHYSICS
G10H2230/015
PHYSICS
G10H2240/325
PHYSICS
G06F3/165
PHYSICS
International classification
G10L15/22
PHYSICS
G10L19/018
PHYSICS
Abstract
Embodiments of the present invention provide for systems and methods for dynamically synthesizing audio files on a mobile device. Embodiments of the present invention are directed to an exemplary audio file recorder with unique music-related features, such as synchronization of multiple pieces of content by the beats per minute, with functionality for content creation, editing, layering of multiple tracks, and sharing. Further, embodiments of the present invention are also related to an exemplary computer software platform which allows users to record any audio into their mobile device and organize the audio files in a manner that allows the files to be easily retrieved and shared.
Claims
1. An audio synthesization system, comprising: a mobile device, the mobile device being configured to: record a first audio file and a second audio file; overlay the second audio file over the first audio file to create a combined audio file; assign at least one textual note to at least one of the first audio file, the second audio file, and the combined audio file; and compress the combined audio file.
2. The system of claim 1, further including: a computing device in communication with the mobile device, the computing device being configured to: receive the compressed combined audio file from the mobile device; and decompress the compressed combined audio file.
3. The system of claim 2, wherein the computing device is further configured to: generate a copy of the decompressed combined audio file; and edit the copy based on a user-provided input to the computing device.
4. The system of claim 3, wherein the computing device includes processing circuitry, the processing circuitry having: a processor; and a memory in communication with the processor.
5. The system of claim 1, wherein the first audio file is recorded during a first period of time and the second audio file is recorded during a second period of time different than the first period of time.
6. The system of claim 1, wherein the at least one textual note is assigned to at least one point in time in the combined audio file.
7. The system of claim 1, wherein the mobile device is further configured to assign at least one textual note to at least one of the first audio file and the second audio file prior to creating the combined audio file.
8. The system of claim 1, wherein the mobile device is further configured to: record a user voice input; correlate the user voice input to a predetermined command; and perform at least one action based on the correlation.
9. A system for dynamically synthesizing audio files on a mobile device, the system comprising: a computing device; and a mobile device in communication with the computing device, the mobile device being configured to: record a first audio file and a second audio file; overlay the second audio file over the first audio file to create a combined audio file; assign at least one textual note to at least one point in time in the combined audio file; compress the combined audio file; and transmit the compressed combined audio file to the computing device.
10. The system of claim 9, wherein the computing device includes processing circuitry configured to: receive the compressed combined audio file from the mobile device; decompress the compressed combined audio file; generate a copy of the decompressed combined audio file; and edit the copy based on a user-provided input to the computing device.
11. The system of claim 10, wherein the computing device is further configured to: assign at least one textual note to at least one point in time in the copy; compress the copy; and transmit the compressed copy to a second computing device configured to receive the compressed copy.
12. The system of claim 9, wherein the first audio file is recorded during a first period of time and the second audio file is recorded during a second period of time different than the first period of time.
13. The system of claim 9, wherein the mobile device is further configured to assign at least one textual note to at least one of the first audio file and the second audio file prior to creating the combined audio file.
14. The system of claim 9, wherein the mobile device further includes a microphone configured to record a user voice input.
15. The system of claim 14, wherein the mobile device is further configured to: correlate the user voice input to a predetermined command; and perform at least one action based on the correlation.
16. The system of claim 15, wherein the at least one action includes at least one of: (a) assigning a desired time signature to at least one of the first audio file, the second audio file, and the combined audio file; and (b) assigning a desired beats per minute to at least one of the first audio file, the second audio file, and the combined audio file.
17. A method for dynamically synthesizing audio files on a mobile device, the method comprising: recording, with a mobile device, a first audio file and a second audio file, the first audio file is recorded during a first period of time and the second audio file is recorded during a second period of time, wherein the first period of time and the second period of time are different; overlaying, with the mobile device, the second audio file over the first audio file to create a combined audio file; assigning, with the mobile device, at least one textual note to at least one point in time in the combined audio file; compressing, with the mobile device, the combined audio file; and transmitting, with the mobile device, the compressed combined audio file to a computing device.
18. The method of claim 17, further including: decompressing, with the computing device, the compressed combined audio file; generating, with the computing device, a copy of the decompressed combined audio file; and editing, with the computing device, the copy based on a user-provided input to the computing device.
19. The method of claim 18, further including: recording, with the mobile device, a user voice input; correlating, with the mobile device, the user voice input to a predetermined command; and performing, with mobile device, at least one action based on the correlation.
20. The method of claim 19, wherein the at least one action includes at least one of: (a) assigning a desired time signature to at least one of the first audio file, the second audio file, and the combined audio file; and (b) assigning a desired beats per minute to at least one of the first audio file, the second audio file, and the combined audio file.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0034] Further objects, features, and advantages of the present disclosure will become apparent from the following detailed description taken in conjunction with the accompanying Figure showing illustrative embodiments of the present disclosure.
[0035]
[0036]
[0037]
[0038]
[0039]
[0040]
DETAILED DESCRIPTION
[0041] The following description of embodiments provides non-limiting representative examples referencing numerals to particularly describe features and teachings of different aspects of the invention. The embodiments described should be recognized as capable of implementation separately, or in combination, with other embodiments from the description of the embodiments. A person of ordinary skill in the art reviewing the description of embodiments should be able to learn and understand the different described aspects of the invention. The description of embodiments should facilitate understanding of the invention to such an extent that other implementations, not specifically covered but within the knowledge of a person of skill in the art having read the description of embodiments, would be understood to be consistent with an application of the invention.
[0042]
[0043] As described herein, it is to be understood that the first note 20 and second note 22 may be attached to either first audio file 13, the second audio file 15, and/or the combined audio file 18 at any stage. In other words, the first and second notes 20, 22 can be attached to the combined audio file 18 prior to, or after, the first audio file 13 has been overlaid by the second audio file 15.
[0044]
[0045] Continuing to refer to
[0046] It is to be understood that the server 14 is configured to store or archive audio files to the session folder 24, thereby allowing users to retrieve, edit, and layer the related audio files as the users see fit from a common portal. In some embodiments, any edits to the session folder 24, e.g., by the user who created the recording (user 1) or another user (user 2), can result in the creation, e.g., by the exemplary software platform, of the duplicate patch session folder 26, thereby leaving original session folder 24 in its original form. In other words, after a user edits the session folder 24, there can be at least two versions of the session folder, e.g., the original session folder 24 and at least one edited duplicate patch session folder 26. Further, the server 14 can allow a user to hold the session folder 24 in manner that would not occupy storage space on the user's device because the session folder 24 can be stored and accessed in the “cloud.” Search capabilities for session folders 24, 26 may include the name of any user who has worked on and/or shared the folder. The result is a virtual studio environment with all content stored in the “cloud” and accessible by authorized users. Each access and edit shall be documented with a keystroke-by-keystroke, time-stamped audit trail.
[0047] Accordingly, users are able to record, layer, organize, notate, and share session folders with any other desired user in the world.
[0048] According to one or more embodiments, the files on the server 14 may also be dragged and dropped using a touch screen, a mouse, or any other means or user interface to any destination on a computer, laptop, or other device. When creating music or transferring audio files people have run into the issue of not being able to easily revert to the original folder where the first content was stored. This allows for users to easily keep track of their versions and edits, including who has contributed to each version and at which point in the process, and also provides a simplified way to revert to the original session folder.
[0049]
[0050] As mentioned above, in one or more embodiments, once an audio file 13, 15, and/or 18 is created the user can add text in the form of a note to specific points of the audio file and have the text stored as time stamps. For example, once the user has created a combined audio file 18 and then writes text in a first note 20 at a first point of the audio file 18 (e.g., 1.45 second mark), a second note 22 at a second point 22 of the audio file 18 (e.g., 4.75 second mark), and a third note (not shown) at a third point of the audio file 18 (e.g., 24.44 second mark). The user may also zoom in and out of the audio file 18 to choose the exact location for the text by pinching in and out with their fingers on a touch screen interface of the mobile or secondary devices 12, 16, or by pressing zoom in and out buttons on a non-touch screen mobile or secondary device 12, 16. The second marks written on the audio file 18 can be stored in chronological order as time stamps in a log 28. The user may then press on the time stamps listed and whichever stamp the user presses will bring the user to that exact location in the audio file 18 where the text of the corresponding note is. The user can then initiate the playing of the audio file 18 from that exact text spot after pressing on the specified time stamp 28. The users may also rename the time stamps to specific names rather than the marked time to make it easier to keep track of their text.
[0051] According to one or more embodiments, the time stamps can be renamed by pressing on the “time stamp” and selecting an option to rename the time stamp. The user can type in whatever it wants that particular time stamp to be called. The user can also color code the time stamps if they desire that for an easier retrieval process. The time stamps can be color coded by pressing on the time stamp and selecting the option to color code. Upon the selection of this option, the user can be presented with an array of colors to choose from for that particular time stamp. These time stamps can be stored in the user's session folder 24 that contains all content for that recording. This allows users to resort back to their recordings and easily retrieve their notes in this efficient new process using time stamps.
[0052]
[0053] As described herein, the voice command from the user may be initiated by activating a Voice Time Signature button (not shown) (or any other similar option) built into the software platform of the exemplary system, or by initiating a built-in, voice-controlled personal assistant (e.g., Siri®, Cortana®, etc.) of the mobile device that is configured and programmed to sense and react to the user's voice command. Once the user's voice is sensed by the mobile device 12, the mobile device 12 can record the user's voice with the device's microphone (microphones 33 and/or 43 discussed in more detail below) and analyze the user's voice using the exemplary software platform of the present invention. Once the voice command is analyzed, the time signature can be displayed on the screen. Additionally, the user may engage the Voice Time Signature button or the built-in, voice-controlled personal assistant of the mobile device to specify the BPM using the user's voice. Once the user's voice is sensed by the mobile device 12, the mobile device 12 can record the user's voice with the device's microphone and analyze the user's voice using the exemplary software platform of the mobile device 12. Once the voice command is analyzed, the mobile device can determine the average distance in milliseconds between the audio's amplitude using the peak indicator illustrated in
[0054] According to one or more embodiments, the mobile device 12 and/or secondary device 16 are each configured to enable a user to create audio that corresponds to a set Time Signature and BPM, or allows the user to upload their own audio to the device. For example, the mobile device is configured to generate a randomly-generated “mashup song” using these created or uploaded audio files. The mobile device and/or secondary device are configured to enable the user to select the desired length of the song by typing in the length parameters in minutes and seconds.
[0055] As described above, in one or more embodiments, a user can record or upload to the mobile device, which can randomly generate unique songs out of original content. In one example, the user can upload an 8-bar loop of a drumbeat, bassline, chord progression, guitar pattern, and/or piano solo. The user can select to make a 3-minute song on the mobile device. The user can then select the number of versions they want to generate by typing in the number of versions. In this example, the user enters 10 versions. The mobile device can generate 10 unique versions of the 3-minute song using the original loops. The mobile device is configured such that the user can change the sections of the song to their liking by enabling the user to move, delete, and mute the loops to make the randomly generated versions achieve their desired sound. Audio files, such as vocals, can have the option of staying the same throughout the song and unedited by the mobile device or the user can opt for the mobile device to generate new patterns to create unique content. Thus, the mobile device provides music creators the ability to randomly generate any number of desired versions, of specified length, using their own original uploaded content. This allows creators to revision their music creation abilities. The BPM of the song can be determined using the processes described herein and can also be notated as described herein. Additionally, in some embodiments, the mobile device is configured to enable the user to attach notes and text to edit each generated song using the process described herein. The user (such as, for example, a DJ, music producer, audio engineer, or entertainer) can use text to “Grade the Versions” to express which versions of the song sound the best and which can be more valuable to potential consumers. For example, a DJ could play songs at events, but every event could hear a different version of each song, which means no audience will have exactly the same overall experience, making each event truly unique. Songs can be randomly generated from a centralized set of content using the systems described herein.
[0056]
[0057] The mobile device 12 is also configured to include software (SW) 40. The software 40 is stored internally in, for example, the memory 38, or stored in external memory (e.g., database, storage array, network storage device, etc.) and may be accessible by the mobile device 12 via an external connection. The software 40 may be executable by the processing circuitry 34.
[0058] The processing circuitry 34 may be configured to control any of the methods and/or processes described herein and/or to cause such methods, and/or processes to be performed, e.g., by the mobile device 12. The processor 36 corresponds to one or more processors for performing mobile device 12 functions described herein. In some embodiments, the software 40 may include instructions that, when executed by the processor and/or processing circuitry, causes the processor 36 and/or processing circuitry 34 to perform the processes described herein with respect to the mobile device 12.
[0059]
[0060] The secondary device 16 is also configured to include software 50. The software is stored internally in, for example, the memory 48, or stored in external memory (e.g., database, storage array, network storage device, etc.) and may be accessible by the secondary device 16 via an external connection. The software 50 may be executable by the processing circuitry 44.
[0061] The processing circuitry 44 may be configured to control any of the methods and/or processes described herein and/or to cause such methods, and/or processes to be performed, e.g., by the secondary device 16. The processor 46 corresponds to one or more processors for performing secondary device functions described herein. In some embodiments, the software 50 may include instructions that, when executed by the processor 46 and/or processing circuitry 44, causes the processor 46 and/or processing circuitry 44 to perform the processes described herein with respect to the secondary device 16.
[0062] The foregoing merely illustrates the principles of the disclosure. Various modifications and alterations to the described embodiments will be apparent to those skilled in the art in view of the teachings herein. It will thus be appreciated that those skilled in the art will be able to devise numerous systems, arrangements, and procedures which, although not explicitly shown or described herein, embody the principles of the disclosure and can be thus within the spirit and scope of the disclosure. Various different exemplary embodiments can be used together with one another, as well as interchangeably therewith, as should be understood by those having ordinary skill in the art. In addition, certain terms used in the present disclosure, including the specification and drawings, can be used synonymously in certain instances, including, but not limited to, for example, data and information.
[0063] It should be understood that, while these words, and/or other words that can be synonymous to one another, can be used synonymously herein, that there can be instances when such words can be intended to not be used synonymously. Further, to the extent that the prior art knowledge has not been explicitly incorporated by reference herein above, it is explicitly incorporated herein in its entirety. All publications referenced are incorporated herein by reference in their entireties.