Display control method and editing apparatus for voice synthesis
10497359 ยท 2019-12-03
Assignee
Inventors
Cpc classification
G06F3/04847
PHYSICS
G10H2220/126
PHYSICS
G10L13/02
PHYSICS
G06F3/048
PHYSICS
G10H2220/116
PHYSICS
G10L13/033
PHYSICS
G10H2250/025
PHYSICS
International classification
G10L13/033
PHYSICS
G10L13/04
PHYSICS
G10L13/02
PHYSICS
G06F3/048
PHYSICS
G06F3/0484
PHYSICS
Abstract
A display control method executed by a processor, the method includes the steps of: displaying, on a display device, a note icon that represents a note of a voice to be synthesized and an indicator that is moved in accordance with an operation received from a user; displaying, on the display device, first options that belong to a first layer among layers in a hierarchical structure, for the user to select a singing expression to be applied to the note from among a plurality of singing expressions; and displaying, on the display device, when the indicator is moved into an area corresponding to a particular option selected from among the first options, second options that correspond to the particular option and belong to a second layer that is below the first layer in the hierarchical structure.
Claims
1. A display control method executed by a processor, the method comprising the steps of: displaying, on a display device, a note icon that represents a note of a voice to be synthesized and an indicator that is moved in accordance with an operation received from a user; displaying, on the display device, first options that belong to a first layer among layers in a hierarchical structure while not displaying other layers among layers in the hierarchical structure when the indicator is not moved into an area corresponding to an option of the first options, for the user to select a singing expression to be applied to the note from among a plurality of singing expressions; and displaying, on the display device, when the indicator is moved into an area corresponding to a particular option selected from among the first options and without requiring any further operation, second options that correspond to the particular option and belong to a second layer that is below the first layer in the hierarchical structure, the second options being displayed in a balloon image, the balloon image being positioned on the display so as not to overlap with the previously displayed options.
2. The display control method according to claim 1, further comprising the step of selecting a singing expression to be applied to the note from among the plurality of singing expressions, after options belonging to at least two layers including the first layer, which is the highest layer among the layers in the hierarchical structure, are selected by the user, wherein the selected singing expression is specified by a combination of the selected options.
3. The display control method according to claim 2, wherein options that belong to the highest layer among the layers represent categories of singing expressions, and one of which options is selected to select one of the categories by the user.
4. The display control method according to claim 3, wherein options that belong to the second highest layer among the layers and correspond to the selected category represent features of singing expressions of the category, and one of which options is selected to select one of the features by the user.
5. The display control method according to claim 2, further comprising the steps of: arranging, within a score region having a time axis, the note icon that extends along the time axis over an utterance period of the note; and displaying, on the display device, a first operation icon when the indicator is moved into a range of the utterance period, wherein the first options are displayed on the display device when the indicator is moved to a position corresponding to the first operation icon.
6. The display control method according to claim 5, wherein options of each layer for selecting a singing expression for a beginning section of the utterance period of the note are displayed on a start side of the note icon, and options of each layer for selecting a singing expression for an ending section of the utterance period of the note are displayed on an end side of the note icon.
7. The display control method according to claim 5, wherein the utterance period of the note includes multiple sections to each of which a singing expression can be applied, wherein an operation icon is displayed on the display device for each section of the multiple sections when the indicator is moved into the range of the utterance period, and wherein a singing expression is selected with using the displayed operation icon of one section among the sections and applied to the one section.
8. The display control method according to claim 5, further comprising the steps of: changing the first operation icon, after the singing expression is selected by the user, into a second operation icon that represents a time length during which the singing expression is applied to the note; and changing a display length of the second operation icon along the time axis in accordance with an operation received from the user.
9. The display control method according to claim 8, further comprising the step of displaying a character string that indicates the selected singing expression on the second operation icon.
10. The display control method according to claim 1, further comprising the step of displaying, on the display device, control information that relates to a process of applying a singing expression selected by the user to the voice of the note to be synthesized, wherein the control information includes at least one of the following variables: a duration ratio of a period, within the utterance period of the note, to which the singing expression is to be applied; a speed variable indicating, in a case where a voice component of the singing expression is to be at least either extended or contracted in the direction of the time axis, a degree of extension or contraction, the speed variable also indicating which of a start side or an end side of the voice component is to be extended or contracted; a maximum value of an application ratio of the singing expression to be applied to the voice of the note to be synthesized; a temporal change of the application ratio from a value at a start of the singing expression to the maximum value; or a temporal change of the application ratio from the maximum value to a value at an end of the singing expression.
11. The display control method according to claim 10, further comprising the step of setting each variable in the control information in accordance with an operation received from the user.
12. The display control method according to claim 11, wherein each variable in the control information is set in accordance with an operation received from the user with respect to each of a plurality of feature amounts.
13. The display control method according to claim 1, wherein the voice of the note is synthesized by mixing of a first sound unit and a second sound unit, the second sound unit having a singing expression selected by the user from among the plurality of singing expressions.
14. The display control method according to claim 1, further comprising the steps of: arranging, within a score region having a time axis, the note icon that extends along the time axis over an utterance period of the note; and displaying, on the display device, a first operation icon when the indicator is moved into a range of the utterance period, wherein the first options are displayed on the display device when the indicator is moved to a position corresponding to the first operation icon.
15. An editing apparatus for voice synthesis comprising: a display device; an input device configured to receive operations from a user; and a processor configured to: control the display device to display a note icon that represents a note of a voice to be synthesized and an indicator that is moved in accordance with an operation received by the input device; control the display device to display first options that belong to a first layer among layers in a hierarchical structure while not displaying other layers among layers in the hierarchical structure when the indicator is not moved into an area corresponding to an option of the first options, and when the indicator is moved into an area corresponding to a particular option from among the first options and without requiring any further operation, to control the display device to display second options that correspond to the particular option and belong to a second layer that is below the first layer in the hierarchical structure, the second options being displayed in a balloon image, the balloon image being positioned on the display so as not to overlap with the previously displayed options; and select a singing expression to be applied to the note, from among a plurality of singing expressions, according to selection operations received from the user by the input device, wherein the selection operations are performed in a layer-by-layer manner for the layers in the hierarchical structure.
16. The editing apparatus according to claim 15, wherein the processor is further configure to: arrange, within a score region having a time axis, the note icon that extends along the time axis over an utterance period of the note; and and control the display device to display a first operation icon when the indicator is moved into a range of the utterance period, wherein the first options are displayed on the display device when the indicator is moved to a position corresponding to the first operation icon.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
DESCRIPTION OF THE EMBODIMENTS
(26) First Embodiment
(27)
(28) The display device 14 (e.g., a liquid crystal display panel) displays an image instructed by the arithmetic processing device 10. The input device 16 is an operating device that is operated by a user for the purpose of providing the voice synthesis apparatus 100 with different instructions. In other words, the input device 16 receives an operation from a user. A pointing device, such as a mouse, is assumed as the input device 16 in the first embodiment. The sound output device 18 (e.g., a loudspeaker or headphones) reproduces sounds corresponding to voice signals V.
(29) The storage device 12 is, for example, a non-transitory recording medium. Any known storage medium, such as a semiconductor recording medium or a magnetic recording medium, or a combination of recording media may be used as the storage device 12. In this specification, a non-transitory recording medium encompasses all computer-readable storage media excluding a transitory, propagating signal, and does not exclude volatile storage media. The storage device 12 stores a program PGM that is executed by the arithmetic processing device 10 together with a variety of data used by the arithmetic processing device 10. The storage device 12 in the first embodiment stores sound unit groups L and synthesis information S as described below.
(30) A sound unit group L is a collection of sound units (a voice synthesis library) obtained in advance from a recorded voice of a particular singer. Each sound unit is a phoneme corresponding to the smallest linguistic unit of a voice or a phoneme chain (e.g., a diphone or a triphone) consisting of inter-connected phonemes. A sound unit is expressed as a sound waveform in a time domain or a spectrum in a frequency domain. The sound unit groups L in the first embodiment contain a sound unit group L.sub.1 and a sound unit group L.sub.2. The sound unit group L.sub.1 contains sound units P obtained from general voice utterances of the singer. The sound unit group L.sub.2 contains sound units Q obtained from the voice uttered by the same singer as that for the sound units P. The sound units Q, however, differ from the sound units P in vocal characteristics. Each sound unit Q is obtained from the voice of the singer uttered at a certain pitch in a certain singing expression. More specifically, a sound unit Q is obtained from the voice of the singer uttered with special (non-modal) vocal characteristics in a particular singing expression, which is produced by movement of at least either the vocal chords or the mouth cavity in a way that differs to that in producing a general voice utterance. The sound unit group L.sub.2 contains sound units Q to which singing expressions of a rough voice, a husky voice, a growling voice, or other characteristic singing expressions have been applied. Alternatively, the singer whose voice is used for sound units Q of the sound unit group L.sub.2 may be different from a singer whose voice is used for sound units P of the sound unit group L.sub.1.
(31) As shown in
(32) The user may apply a desired singing expression to each note in a synthesized song. Expression information X4 that defines a singing expression is appended to the data unit U, within the synthesis information S, of a note to which the singing expression is applied. The expression information X4 includes identification information E representing a type of the singing expression and control information Z including variables related to the process of applying the singing expression to the voice to be synthesized. In the first embodiment, the user may specify a singing expression for each of a beginning section and an ending section of a given note, the beginning section being at the start side within the utterance period X2 of the note (e.g., an attack part) and the ending section being at the end side within the utterance period X2 of the note (e.g., a release part). That is, for each of the beginning section and the ending section of a note, identification information E and control information Z of a singing expression are set. In other words, the expression information X4 is set for each section, within a note, to which a singing expression is applied.
(33) The arithmetic processing device 10 (e.g., a Central Processing Unit (CPU) or simply a processor) in
(34) The instruction receiver (instruction receiving means) 22 receives instructions from the user, which instructions correspond to an operation made by the user; or more specifically, correspond to an operation made by the user at the input device 16. The display controller (display control means) 24 causes different images to be displayed on the display device 14. More specifically, the display controller 24 in the first embodiment causes an editing screen to be displayed on the display device 14, the editing screen allowing the user to review a content of the synthesized song specified by synthesis information S. An example of an editing screen 40 is shown in
(35) For each note that is specified by the synthesis information S, the display controller 24 positions a note icon 42 on the editing screen 40. The note icon 42 is an image used to represent each note in the synthesized song. More specifically, the position of a note icon 42 relative to the pitch-axis direction is set according to the pitch X1 specified by the synthesis information S with respect to a note corresponding to the note icon 42, and the position and display length of the note icon 42 relative to the time-axis direction is set according to the utterance period X2 specified by the synthesis information S with respect to the note. To the note icon 42 of each note, the phonetic symbol X3 (an utterance character/letter(s) and a phoneme sign) is applied.
(36) The user may instruct the voice synthesis apparatus 100 (editing screen 40) of at least either an addition/shift of the note icon 42 or an addition/change of the phonetic symbol X3 by appropriately operating the input device 16 while referring to the editing screen 40 in
(37) The information manager 26 in
(38) The selection handler (selection handling means) 25 in
(39)
(40) For example, in the first layer (the highest layer) of the singing expressions that can be applied to the beginning section, three options C.sub.1 are included: namely, fry (vocal fry), growl, and rough. Each option C.sub.1 (first option) in the first layer represents a category of one or more singing expressions corresponding to the option C.sub.1. Accordingly, by selecting one of the options C.sub.1, the user selects one of the categories of singing expressions.
(41) The second layer is a layer below the first layer in the hierarchical structure. That is, the second layer is the second highest layer among the layers in the hierarchical structure. The following options are included in the second layer: two options C.sub.2, slow and creaky, that correspond to fry in the first layer; three options C.sub.2, hard, sharp, and slow, that correspond to growl in the first layer; and two options C.sub.2, slow and sharp, that correspond to rough in the first layer. Each option C.sub.2 (second option) in the second layer represents a feature of a singing expression(s) that corresponds to the option C.sub.1 corresponding to each option C.sub.2. For example, a singing expression with slow means that this singing expression is to make the beginning (or attack) of a sound relatively gradual, and a singing expression with sharp means that this singing expression is to make the beginning of a sound relatively rapid. A singing expression with hard means that this singing expression makes the strength of the main feature of this singing expression represented by the corresponding option C.sub.1 (in this case, the category growl) is notable. In the same way, options C.sub.2 corresponding to a category (an option C.sub.1) represent subsidiary features of singing expressions of the category. By selecting one of the options C.sub.2 corresponding to the option C.sub.1 selected in the first layer, the user selects one of the subsidiary features of the singing expressions corresponding to the selected option C.sub.1.
(42) Two supplemental options C.sub.3 in the third layer, namely type 1 and type 2 are corresponded with sharp in the second layer, where the sharp in the second layer corresponds to rough in the first layer. The type 1 and type 2 have different vocal characteristics from one another. A combination of options C belonging to different layers from the first to the last layer specifies a single type of a singing expression. More specifically, in
(43) In the above description, examples are given of singing expressions that apply to the beginning section of a note. A similar hierarchical structure defines the singing expressions for the ending section of a note. More specifically, the first layer includes one option C.sub.1 (first option), soft, which means that a voice sound slowly decays, as in a sigh; and the second layer includes two options C.sub.2 (second options), long (taking a long time) and short (taking a short time), each of which corresponds to the option C.sub.1. In other words,
(44) A sound unit Q uttered using one type of a singing expression is included in the sound unit group L.sub.2 for each of multiple vowels (a, i, u, e, and o). More specifically, from a voice steadily uttering a given vowel with a particular singing expression, a sound unit Q of the vowel corresponding to the singing expression is collected. The types of singing expressions are not limited to the examples shown in
(45) The selection handler 25 in the first embodiment causes the display device 14 to display, for each of layers in the above-described hierarchical structure, options C of the layer so that the user can select any one of the options C in a step-by-step manner. In this way, the selection handler 25 selects, as the singing expression to be applied to a voice to be synthesized (i.e., to be applied to a given note), a singing expression that corresponds to a combination of options C across the multiple layers, according to selection operations received from the user by the input device, the selection operations being performed in a layer-by-layer manner for the layers in the hierarchical structure. Referring to
(46) In the following, referring to
(47) The user moves the indicator 44 close to a desired note for which a singing expression is to be selected (hereinafter, a target note) by appropriately operating the input device 16. More specifically, when the indicator 44 moves into the range Y of the note icon 42 of the target note relative to the time-axis direction (St1), the selection handler 25 causes the display device 14 to display an operation icon 46.sub.A (Atk) and an operation icon 46.sub.B (Rls) as in the example shown in
(48) If the user wishes to select a singing expression for the beginning section, he/she moves the indicator 44 close to the operation icon 46.sub.A by appropriately operating the input device 16. When the indicator 44 is moved to a position corresponding to the operation icon 46.sub.A in accordance with the operation of the user (St3), the selection handler 25 controls the display device 14 to display, as in the example shown in
(49) The user moves the indicator 44 close to a desired option C.sub.1 among the options C.sub.1 arranged in the balloon image N.sub.1 by operating the input device 16. When the indicator 44 moves into an area corresponding to a given option C.sub.1 belonging to the first layer (e.g., an area overlapping this option C.sub.1) (St6), the selection handler 25 controls the display device 14 to display, as in an example shown in
(50) The user moves the indicator 44 close to a desired option C.sub.2 among the options C.sub.2 arranged in the balloon image N.sub.2 by operating the input device 16. When the indicator 44 moves into an area corresponding to a given option C.sub.2 belonging to the second layer (e.g., an area overlapping this option C.sub.2) (St6), the selection handler 25, as in an example shown in
(51) The selection handler 25 deletes both the balloon image N.sub.1 and the balloon image N.sub.2 from the editing screen 40 when the indicator 44 moves to a position that does not correspond to any of the following positions (e.g., outside the right edge of the balloon image N.sub.2 in
(52)
(53) After the singing expression for the beginning section has been determined by the above-steps, the information manager 26 appends, to the data unit U of the target note, expression information X4 that represents the selected singing expression. More specifically, the expression information X4 includes identification information E of the selected singing expression (e.g., information indicating the combination of the options C across the different layers) and control information Z that is set to an initial value.
(54) In addition, after the selection of the singing expression for the beginning section has been finalized, the selection handler 25 changes the operation icon 46.sub.A (an example of a first operation icon) of the target note to an operation icon 48.sub.A (an example of a second operation icon) as in the example shown in
(55) An operation point W.sub.A that indicates the end of the expression applied period T.sub.A is displayed at the rear end (right-hand side end), along the time axis, of the operation icon 48.sub.A. As in the example shown in
(56) In the above description, selection of a singing expression and adjustment of the control information Z are described with respect to the beginning section of the utterance period X2 of the target note. The operations required for selection of a singing expression and adjustment of the control information Z with respect to the ending section of the utterance period X2 are similar to those for the beginning section. In other words, as in the example shown in
(57) As in the example shown in
(58) After the singing expression for the ending section has been determined by the above steps, the selection handler 25 changes the operation icon 46.sub.B (an example of a first operation icon) to an operation icon 48.sub.B (an example of a second operation icon) as shown in
(59) As will be understood from the above description, in accordance with an instruction that the instruction receiver 22 receives from the user via the input device 16, the information manager 26 sets variables in the control information Z (the expression applied period T.sub.A and the expression applied period T.sub.B), the control information Z relating to the process of applying a sound component (sound unit Q) to the voice of a certain note to be synthesized, the sound component having a singing expression selected by the user with respect to the note.
(60) The voice synthesizer 28 in
(61) As shown in
(62) After the above adjustments are complete, the voice synthesizer 28 mixes the adjusted beginning-section sound unit Q.sub.A with the start side of the sound unit P (the beginning section of the sound unit P), and also mixes the adjusted ending-section sound unit Q.sub.B with the end side of the sound unit P (the ending section of the sound unit P). In detail, the sound unit Q.sub.A is applied to the sound unit P at an application ratio R.sub.A, and the sound unit Q.sub.B is applied to the sound unit P at an application ratio R.sub.B. The application ratios R.sub.A and R.sub.B are each a degree to which the sound unit Q.sub.A or Q.sub.B, respectively, is applied to the sound unit P, and are each set to a predetermined value. In other words, the application ratio R.sub.A is a weighted value for the sound unit Q.sub.A, and the application ratio R.sub.B is a weighted value for the sound unit Q.sub.B. A voice signal V is generated by interconnecting, along the time axis, sound units P that have been generated by the above steps sequentially for the notes included in the synthesis information S.
(63) Since the vocal characteristics differ between a sound unit Q and a sound unit P to which the sound unit Q is applied, the user may notice discontinuity in sound at least either at the start of or at the end of the sound unit Q. To minimize such discontinuity in sound, the sound unit Q may be cross-faded into the sound unit P.
(64) The singing voice of the synthesized song is reproduced by voice signals V generated by the voice synthesizer 28 being supplied to the sound output device 18.
(65) The above-described configuration allows the user to select one of multiple singing expressions as a result of allowing the user to select an option C for each layer in a layer-by-layer manner. Therefore, the user is able to easily find and select a desired singing expression, as compared to a configuration in which all possible types of singing expressions selectable by the user are displayed, for the user to select a single desired singing expression from among them. Furthermore, in the above configuration, compared to a configuration in which all possible types of singing expressions are displayed, the number of types of singing expressions displayed at a time can be reduced, and thus an area required to display singing expressions also can be reduced. Thus, this configuration is preferably suited for use on a display device (e.g., a display device of a mobile terminal) that has a relatively small display area.
(66) In addition, when the indicator 44 is moved into an area that corresponds to an option C belonging to a layer of the hierarchical structure, options C that are situated below the subject option C are displayed on the display device 14. Accordingly, the user can select an option C for each layer without need to perform any operation (e.g., a mouse-click) other than moving the indicator 44. Therefore, compared to a configuration in which the user selects a desired option by performing an operation additional to moving the indicator 44, less of a difficulty is imposed on the user in selecting one of multiple singing expressions.
(67) Moreover, in the above-described embodiment, when as a result of an operation performed by the user the indicator 44 moves into a range Y between the start and the end of a note icon 42, operation icons (46.sub.A and 46.sub.B) are displayed, and when the indicator 44 moves to a position corresponding to an operation icon (46.sub.A or 46.sub.B), options C belonging to the highest layer of the corresponding hierarchical structure are displayed. Thus, in contrast to a configuration in which the operation icons 46.sub.A and 46.sub.B are continuously shown, the editing screen 40 is relatively simple. Moreover, when the indicator 44 moves to a position corresponding to an operation icon (46.sub.A or 46.sub.B), options C belonging to the highest layer are displayed (i.e., selection of an option for each layer by the user is initiated). As a result, less of a difficulty is imposed on the user as compared to a configuration in which the options C belonging to the highest layer are displayed when an operation (e.g., a mouse-click) other than moving the indicator 44 is applied to the operation icon (46.sub.A or 46.sub.B).
(68) Furthermore, the options C belonging to the layers for selecting a singing expression for the beginning section of a note are displayed on the start side of the note icon 42, and the options C belonging to the layers for selecting a singing expression for the ending section of the note are displayed on the end side of the note icon 42. Accordingly, the user is able to understand both visually and intuitively the relationship between the portion (start or end) and the singing expression to be applied to the portion.
(69) In addition, the voice signal V that represents the synthesized voice (the singing voice) is generated by mixing multiple sound units (a sound unit P and a sound unit Q) having different vocal characteristics. More specifically, the voice signal V is generated by mixing a sound unit Q (the second sound unit) that has a singing expression selected by the user with the sound unit P (the first sound unit). Thus, compared to a configuration in which a voice signal V is generated by modifying a single sound unit (e.g., a sound unit P), less deterioration of the voice occurs, and a more natural-sounding voice is synthesized.
(70) Second Embodiment
(71) A second embodiment of the present invention will now be explained. In the first embodiment, an expression applied period T.sub.A and an expression applied period T.sub.B are given as examples of variables in the control information Z. In the second embodiment, the user may adjust, as the control information Z, a variety of variables that relate to the application of singing expressions. In the following modes, which are described as examples, elements that are the same as those in the first embodiment regarding their effect and function are denoted by the same reference symbols as used in the first embodiment, and detailed explanation thereof will be omitted as appropriate. Description of the selection of singing expressions also will be omitted, since the selection is substantially the same as in the first embodiment.
(72)
(73) When the user inputs into the input device 16 an instruction to perform detailed settings of the control information Z after selecting a desired note (target note), the display controller 24 causes the display device 14 to display the control information setting screen 50 of
(74) In the region D.sub.A, an operation icon 48.sub.A and an operator 49.sub.A are displayed. In a manner similar to that described above where the user performs an operation on the editing screen 40, the information manager 26 sets the duration ratio .sub.A of the expression applied period T.sub.A in accordance with an operation performed by the user to move the operation point W.sub.A of the operation icon 48.sub.A. By performing an operation on the control information setting screen 50 via the input device 16, the user can move the operator 49.sub.A in the time-axis direction. The information manager 26 sets a speed variable S.sub.A in accordance with a position of the operator 49.sub.A. The speed variable S.sub.A defines a degree of extension or contraction in the process of adjusting the time length of a sound unit Q such that it becomes equivalent to the expression applied period T.sub.A. The speed variable S.sub.A also defines which of the start side or the end side of the sound unit Q (sound component) is to be extended or contracted. The speed variable S.sub.A may be variably set from an initial value (typically, zero) towards both the positive and negative sides (e.g., in a range from 50 to +50).
(75)
(76) The above examples concern the designation of the speed variable S.sub.A for the beginning section of the target note. In a similar manner, with regard to the ending section, the information manager 26 sets a speed variable S.sub.B in accordance with an operation performed by the user on the control information setting screen 50 (an operator 49.sub.B in the region D.sub.B). Regarding the ending section, a relationship between whether the speed variable S.sub.B is a positive or a negative number and the position of the contraction in the sound unit Q (either at the start side or the end side) is reversed as compared to that regarding the speed variable S.sub.A for the beginning section. More specifically, when the speed variable S.sub.B is above an initial value (S.sub.B>0), the closer the sections of the sound unit Q are to the end of the sound unit Q, the more contracted are the sections; and conversely, the closer the sections of the sound unit Q are to the start of the sound unit Q, the more extended are the sections. When the speed variable S.sub.B is below the initial value (S.sub.B<0), the closer the sections of the sound unit Q are to the start of the sound unit Q, the more contracted are the sections; and conversely, the closer the sections of the sound unit Q are to the end of the sound unit Q, the more extended are the sections. The speed variable S.sub.B has a tendency similar to the speed variable S.sub.A such that the greater the absolute value of the speed variable S.sub.B, the higher the degree of extension and contraction. As will be understood from the above description, the greater the speed variable S.sub.B, the more contraction there is at the end side of the sound unit Q for the ending section of the target note. As a result, the release phase of the sound is accelerated immediately before the end of the sound unit Q. Thus, the speed variable S.sub.B can be expressed as a variable that defines the speed of the release phase of the sound of the sound unit Q for the ending section.
(77) The region D.sub.C in
(78) As shown in
(79) The transition image 50.sub.B is a trapezoid that extends across the expression applied period T.sub.B of the singing expression for the ending section within the utterance period X2 of the target note (T.sub.B=.sub.BX2). More specifically, as shown in
(80) The display controller 24 updates the transition image 50.sub.A and the transition image 50.sub.B after the control information Z is modified. The user may change the control information Z by performing operations on the different operators (51.sub.A, 51.sub.B, 52.sub.A1, 52.sub.A2, 52.sub.B1, 52.sub.B2, 53.sub.A1, 53.sub.A2, 53.sub.B1, and 53.sub.B2) of the control information setting screen 50, or by directly inputting values of the control information Z into value input boxes shown as examples in
(81) Maximum Values of Application Ratios R.sub.Amax, R.sub.Bmax
(82) The user can move the operator 51.sub.A vertically by operating the input device 16. In the example shown in
(83) Duration Ratios .sub.A1, .sub.A2, .sub.B1, .sub.B2
(84) By operating the input device 16, the user can move the operator 52.sub.A1 that represents the end of the leading edge period T.sub.A1 along the time-axis direction. The information manager 26 sets the duration ratio .sub.A1 to be variable in accordance with a ratio (%) of the leading edge period T.sub.A1 to the expression applied period T.sub.A, the leading edge period T.sub.A1 being from the start of the expression applied period T.sub.A of the beginning section to the point at which the operator 52.sub.A1 is located. In other words, the application ratio R.sub.A increases from zero to the maximum value R.sub.Amax over the leading edge period T.sub.A1, which is a time length equivalent to the product of the expression applied period T.sub.A and the duration ratio .sub.A1.
(85) Similarly, the user can move the operator 52.sub.A2 that represents the start of the trailing edge period T.sub.A2 along the time-axis direction. The information manager 26 sets the duration ratio .sub.A2 to be a ratio of a certain period to the expression applied period T.sub.A, the certain period being from the start of the expression applied period T.sub.A of the beginning section to a point at which the operator 52.sub.A2 is located. In other words, the application ratio R.sub.A decreases from the maximum value R.sub.Amax to zero over the trailing edge period T.sub.A2, which is from a certain time point to the end of the expression applied period T.sub.A, the certain time point being a time point at which the time length equivalent to the product of the expression applied period T.sub.A and the duration ratio .sub.A2 has passed from the start of the expression applied period T.sub.A. The same applies for the ending section, and the following variables are set according to the operation performed by the user: the ratio (duration ratio .sub.B1) of the leading edge period T.sub.B1 to the expression applied period T.sub.B, the leading edge period T.sub.B1 being from the start of the expression applied period T.sub.B to the point at which the operator 52.sub.B1 is located; and the ratio (duration ratio .sub.B2) of a certain period to the expression applied period T.sub.B, the certain period being from the start of the expression applied period T.sub.B to the point at which the operator 52.sub.B2 is located. In an alternative configuration, the duration ratio .sub.A2 may be the ratio of the trailing edge period T.sub.A2 to the expression applied period T.sub.A; or the duration ratio .sub.B2 may be the ratio of the trailing edge period T.sub.B2 to the expression applied period T.sub.B.
(86) Transition Coefficients .sub.A1, .sub.A2, .sub.B1, .sub.B2
(87) The operator 53.sub.A1 in
(88) When an upward operation is performed on the operator 53.sub.A1 (e.g., mouse-dragging the operator 53.sub.A1 upward), the information manager 26 increases the value of the transition coefficient .sub.A1 by an amount corresponding to the amount of operation from the initial value. As in the example shown in
(89) As in the example shown in
(90) The method to change the form of the temporal change of the application ratio R.sub.A (locus r.sub.A1) according to the transition coefficient .sub.A1 may be freely selected. One possible method is to control the locus r.sub.A1 of the application ratio R.sub.A within the leading edge period T.sub.A1 according to the transition coefficient .sub.A1 by using the following equation (1) that includes a variable that depends on the transition coefficient .sub.A1. The variable in the equation (1) is a value obtained by converting the value of the transition coefficient .sub.A1, which is within a range of 0.sub.A1100, to a value within a range of from 0.9 to +0.9. The symbol t in the equation (1) denotes a time elapsed from a start of the leading edge period T.sub.A1 treated as an origin (t=0).
(91)
(92) In the foregoing, description is given of the temporal change of the application ratio R.sub.A in the leading edge period T.sub.A1 of the expression applied period T.sub.A (i.e., the transition coefficient .sub.A1). The temporal change of the application ratio R.sub.A in the trailing edge period T.sub.A2 of the expression applied period T.sub.A (i.e., the transition coefficient .sub.A2) similarly is set according to the operation performed on the operator 53.sub.A2. More specifically, the information manager 26 sets the transition coefficient .sub.A2 according to the amount of operation performed by the user on the operator 53.sub.A2. In a case where the transition coefficient .sub.A2 is set to an initial value (e.g., 50), the application ratio R.sub.A decreases linearly over the trailing edge period T.sub.A2. When the transition coefficient .sub.A2 is above the initial value 50, the application ratio R.sub.A decreases from the maximum value R.sub.Amax to zero along an upward convex locus r.sub.A2 over the trailing edge period T.sub.A2. When the transition coefficient .sub.A2 is below the initial value 50, the application ratio R.sub.A decreases from the maximum value R.sub.Amax to zero along a downward convex locus r.sub.A2 over the trailing edge period T.sub.A2. The change in the pointer applied to the operator 53.sub.A2 and the display of the interim locus r.sub.A2 while the operator 53.sub.A2 is being operated may be similar to those described with respect to the operations performed on the operator 53.sub.A1.
(93) The locus r.sub.A2 for the decrease of the application ratio R.sub.A in the trailing edge period T.sub.A2 is expressed by the following equation (2). A variable in the equation (2) is, similarly to that in the equation (1), a value obtained by converting the transition coefficient .sub.A2 that is in the range of 0.sub.A2100 to a value in the range between 0.9 and +0.9. The symbol t in the equation (2) denotes the time elapsed from the start of the trailing edge period T.sub.A2 treated as an origin (t=0).
(94)
(95) Similarly to the example described above, the transition coefficient .sub.B1 that defines the form (locus r.sub.B1) of the increase of the application ratio R.sub.B in the leading edge period T.sub.B1 of the expression applied period T.sub.B is set in accordance with the operation performed by the user on the operator 53.sub.B1 shown in
(96) As described in the above example, in the second embodiment, the information manager 26 sets each variable in the control information Z in accordance with an operation received from the user. Such control information Z includes the following:
(97) the duration ratios (.sub.A and .sub.B) of the periods to which singing expressions are to be applied within the utterance period X2 of the target note;
(98) the speed variables (S.sub.A and S.sub.B) indicating, when the time length of a sound unit Q having the corresponding singing expression is to be extended and/or contracted, the degree of the extension or contraction, the speed variables also indicating which of the start side or end side of the sound unit Q is to be extended or contracted;
(99) the maximum values of the application ratios R (R.sub.Amax and R.sub.Bmax);
(100) the temporal changes of the application ratios R on the start side of the singing expressions (.sub.A1, .sub.B1, .sub.A, and .sub.B1); and
(101) the temporal changes of the application ratios R on the end side of the singing expressions (.sub.A2, .sub.B2, .sub.A2, and .sub.B2).
(102) Accordingly, in comparison with a configuration in which fixed values are set for the different variables included in the control information Z, it is possible to generate a voice signal V with various characteristics that match the preference or intention of the user. One or more of the multiple variables given as examples in the above description may be omitted, as appropriate. In other words, the information manager 26 can be understood as an element that sets at least one of the multiple variables given above as examples.
(103) Third Embodiment
(104) In the first embodiment, an example configuration is shown in which a sound unit P and a sound unit Q are each expressed as a sound waveform in the time domain or a spectrum in the frequency domain. In the third embodiment, each of a sound unit P and a sound unit Q is expressed with a plurality of feature amounts that approximately express the audio characteristics of the sound unit. In the third embodiment, parameters representative of the audio characteristics of different frames of a sound unit, such as a spectrum curve, excitation source characteristics, or a fundamental frequency, are used as examples of the feature amounts . A frame is a time length that serves as a unit of analysis of a sound unit. The voice synthesizer 28 calculates each feature amount as the weighted sum of the feature amount for the sound unit P and the feature amount for the sound unit Q of the singing expression, using the corresponding application ratio R as the weighted value. The voice synthesizer 28 inter-connects, along the time axis, sound units having frequency characteristics approximated by the calculated feature amounts to generate a voice signal V.
(105)
(106) By appropriately operating the input device 16, the user can select a feature amount , from among the feature amounts , the control information Z of which is to be modified (hereinafter, a target feature amount). With respect to the target feature amount selected by the user, as in the example shown in the second embodiment, the display controller 24 in the third embodiment displays the transition image 50.sub.A, the transition image 50.sub.B, and the operators (51.sub.A, 51.sub.B, 52.sub.A1, 52.sub.A2, 52.sub.B1, 52.sub.B2, 53.sub.A1, 53.sub.A2, 53.sub.B1, and 53.sub.B2) as shown in the region D.sub.C1 in
(107) In the third embodiment, for each of the feature amounts that relate to a singing expression, each variable in the corresponding control information Z is set in accordance with instructions from the user. Thus, it is possible to synthesize a variety of voices that reflect the user's intention or preference. Another configuration is also preferable in which the user can select between an operation mode in which the user sets the control information Z separately for each feature amount , as in the above example, and an operation mode in which the user sets the control information Z collectively for the feature amounts .
(108) Modifications
(109) The above-described modes may be modified in a variety of ways. Specific modes of modification are described in the following. Two or more modes freely selected from the modes described above and the modes described below may be appropriately combined so long as they do not contradict one another.
(110) (1) In the above-described modes, a sound unit Q uttered at a particular pitch is adjusted to the pitch X1 specified by the synthesis information S. However, sound units Q that are obtained by uttering the same content (i.e., same text) at different pitches (e.g., at three pitches: a high pitch; a middle pitch; and a low pitch) may be recorded for each singing expression for inclusion in the sound unit group L.sub.2; and a sound unit Q may be adjusted to the target pitch X1 by interpolating between sound units Q with different pitches.
(111) (2) In the above-described modes, the operation icons (46.sub.A and 46.sub.B) are displayed when the indicator 44 moves into the range Y, which is between the start and the end of a note icon 42 along the time axis. However, a timing of when to display the operation icons (46.sub.A and 46.sub.B) is not limited to this example. For example, the operation icons (46.sub.A and 46.sub.B) may be displayed when the indicator 44 moves into a rectangular region where a predetermined range that covers the pitch X1 of the note icon 42 along the pitch axis overlaps the range Y of the note icon 42 along the time axis.
(112) (3) In the third embodiment, an example configuration is described in which the sound unit P and the sound unit Q are each expressed with a plurality of feature amounts . In this case, the voice synthesizer 28 may extract a plurality of feature amounts from the sound unit P and the sound unit Q that are expressed each by, for example, a sound waveform or a frequency spectrum. In other words, the storage device 12 does not need to retain a plurality of feature amounts for each sound unit Q.
(113) (4) In the above-described modes, examples are shown in which a singing expression is applied to each of the beginning section and the ending section of a note. However, sections to which singing expressions are applied are not limited to the beginning section and the ending section. For example, any of the above modes may be adopted when a singing expression is applied to a section in the middle of the utterance period X2 of a note. The total number of sections to each of which a singing expression is applied within the utterance period X2 of a note may be any number.
(114) When the indicator 44 moves close to a note icon 42, an operation icon 46 (e.g., 46.sub.A, 46.sub.B, etc.) is displayed for each of all the sections to which singing expressions can be applied within the utterance period X2 of the note. A singing expression for a particular section within the utterance period X2 is selected using the displayed operation icon 46 corresponding to that section and is applied to that section. In this configuration, the user is able to visually recognize the sections to which singing expressions can be applied with respect to the note. Alternatively, an operation icon 46 may be displayed selectively for a particular section (e.g., the section that is closest to the indicator 44) from among the multiple sections, within the note, to which singing expressions can be applied. In the above-described modes, the operation icon 46.sub.A and options C relating to a singing expression for the beginning section are arranged on the start side of the note icon 42, and the operation icon 46.sub.B and options C relating to a singing expression for the ending section are arranged on the end side of the note icon 42. In view of a configuration in which there are sections other than the beginning section and the ending section, each configuration in the above-described modes may be comprehensively expressed as a configuration in which the operation icon 46 and options C relating to a singing expression for a particular section within a note are displayed in a position corresponding to this section, close to the note icon 42 of the note.
(115) In a configuration in which a singing expression may be applied to a section other than the beginning section and the ending section, information that specifies this section (e.g., information that indicates the start and the end of the section, or information that indicates the range of the section) is set separately for each section in the note and is included in the expression information X4. In the sound unit group L.sub.2, multiple sound units Q, each corresponding to a different singing expression, are registered for each section, within a note, to which a singing expression can be applied.
(116) (5) In the above-described modes, the voice synthesis apparatus 100 including the voice synthesizer 28 is described as an example. However, the present invention may be exploited as an apparatus (an editing apparatus for voice synthesis) for editing a voice to be synthesized by the voice synthesizer 28. In other words, the voice synthesizer 28 is not an essential element in the editing apparatus for voice synthesis.
(117) (6) The editing apparatus for voice synthesis, a preferable mode of which is the voice synthesis apparatus 100 shown above, is realized by a general arithmetic processing device, such as a CPU, and a program functioning in cooperation with each other, as shown in the above-mentioned embodiments. Alternatively, the editing apparatus may be realized by hardware (electric circuitry), such as a DSP, that is used exclusively for editing a voice to be synthesized. The above-mentioned program may be provided in a format stored in a computer-readable recording medium and be installed in a computer. The recording medium is, for example, a non-transitory recording medium and may preferably be an optical recording medium (e.g., an optical disc), such as a CD-ROM; but it may be a recording medium of any publicly-known format, such as a semiconductor recording medium or a magnetic recording medium. The program described above may be provided in a format distributed via a communication network and be installed in a computer. The present invention may also be understood as the operation method (editing method for voice synthesis) or the display control method for the editing apparatus in each of the above-described modes.
(118) At least one of the following modes are understood from the above-mentioned embodiments and modifications.
(119) A display control method according to the present invention is executed by a processor and includes the steps of: displaying, on a display device, a note icon that represents a note of a voice to be synthesized and an indicator that is moved in accordance with an operation received from a user; displaying, on the display device, first options that belong to a first layer among layers in a hierarchical structure, for the user to select a singing expression to be applied to the note from among a plurality of singing expressions; and displaying, on the display device, when the indicator is moved into an area corresponding to a particular option selected from among the first options, second options that correspond to the particular option and belong to a second layer that is below the first layer in the hierarchical structure. Since options are displayed in a layer-by-layer manner in the above configuration, a user can more easily find and select a desired singing expression compared with a configuration in which all types of singing expressions that are selectable by the user are shown all at once and the user selects a desired type of singing expression from them. In the above configuration, when the indicator is moved (e.g., mouse-over) to an area that corresponds to a first option in a first layer in the hierarchical structure, second options belonging to a second layer below the first layer to which the subject first option belongs are displayed on the display device. Thus, in the above configuration, the user is able to select an option in a layer without being required to perform any operation other than to move an indicator (e.g., a mouse-click is not required). Therefore, compared with a configuration in which the user selects a desired option by performing an operation additional to the moving of the indicator, the difficulty placed on the user in selecting any one of singing expressions is reduced. Here, an area that corresponds to an option typically is an area that overlaps the subject option, but it may be an area close to the subject option.
(120) The display control method according to a preferred mode of the present invention further includes the step of selecting a singing expression to be applied to the note from among the plurality of singing expressions, after options belonging to at least two layers including the first layer, which is the highest layer among the layers in the hierarchical structure, are selected by the user, wherein the selected singing expression is specified by a combination of the selected options.
(121) The display control method according to a preferred mode of the present invention further includes the steps of: arranging, within a score region having a time axis, the note icon that extends along the time axis over an utterance period of the note; and displaying, on the display device, a first operation icon when the indicator is moved into a range of the utterance period. The first options are displayed on the display device when the indicator is moved to a position corresponding to the first operation icon. In this configuration, when the indicator moves into the range of the utterance period shown with a note icon along the time axis, an operation icon is displayed, and when the indicator moves to a position corresponding to the operation icon, options that belong to the highest layer of the hierarchical structure are displayed. Thus, compared with a configuration in which operation icons are constantly shown, the display image is simplified. Furthermore, because the options belonging to the highest layer are displayed (i.e., the selection of an option for each layer by the user is initiated) when the indicator moves to the position corresponding to the operation icon, a difficulty placed on the user during operation can be reduced as compared to a configuration in which the options belonging to the highest layer are displayed when an operation other than moving the indicator (e.g., the clicking on the operation icon) is applied to the operation icon.
(122) The display control method according to a preferred mode of the present invention further includes the steps of: changing the first operation icon, after the singing expression is selected by the user, into a second operation icon that represents a time length during which the singing expression is applied to the note; and changing a display length of the second operation icon along the time axis in accordance with an operation received from the user. In this configuration, the first operation icon, which serves as a trigger for displaying options of the highest layer, is changed into the second operation icon, which represents the time length during which the selected singing expression is applied to the note. Accordingly, compared with a configuration in which a first operation icon and a second operation icon are displayed separately at positions far apart, the user can more easily select a desired singing expression and adjust the time length of the duration to which the selected singing expression is to be applied, without shifting his/her view far from the vicinity of the note icon.
(123) The display control method according to a preferred mode of the present invention further includes the step of: displaying a character string that indicates the selected singing expression on the second operation icon. In this configuration, while changing a display length of the second operation icon, the user can review for which singing expression he/she is adjusting the time length of application.
(124) Preferably, options that belong to the highest layer among the layers represent categories of singing expressions, and one of which options is selected to select one of the categories by the user. Preferably, options that belong to the second highest layer among the layers and correspond to the selected category represent features of singing expressions of the category, and one of which options is selected to select one of the features by the user.
(125) Preferably, the display control method causes options of each layer in the hierarchical structure for selecting a singing expression for a particular section within the note to be displayed at a position corresponding to the subject section, close to the note icon. By this configuration, the user is able to both visually and intuitively understand the relationship between the section and a singing expression to be applied to the section, since the options for singing expressions for the particular section of the note are displayed in a position that is close to the note icon and corresponds to the subject section. More specifically, options of each layer for selecting a singing expression for a beginning section of the utterance period of the note are displayed on a start side of the note icon, and options of each layer for selecting a singing expression for an ending section of the utterance period of the note are displayed on an end side of the note icon. In this configuration, the options for singing expressions for the beginning section of the note (e.g., the attack part) are displayed on the start side of the note icon, and the options for singing expressions for the ending section of the note (e.g., the release part) are displayed on the end side of the note icon. Therefore, the user is able to visually and intuitively understand the relationship between a portion (the start or the end) and a singing expression to be applied to the portion.
(126) Preferably, the utterance period of the note includes multiple sections to each of which a singing expression can be applied. An operation icon is displayed on the display device for each section of the multiple sections when the indicator is moved into the range of the utterance period, and a singing expression is selected using the displayed operation icon of one section among the sections and applied to the one section. In this configuration, when the indicator moves close to the note icon, an operation icon is displayed for each of all the sections, within the note, to which a singing expression can be applied. Therefore, the user can visually recognize with ease the multiple sections to which the user can apply singing expressions with respect to the subject note.
(127) Preferably, the display control method further includes the step of displaying, on the display device, control information that relates to a process of applying a singing expression selected by the user to the voice of the note to be synthesized. Preferably, the display control method further includes the step of setting each variable in the control information in accordance with an instruction received from the user. In this configuration, because variables in the control information relating to the application of singing expressions are set in accordance with instructions from the user, a variety of voices that reflect a preference or intention of the user can be synthesized. The control information includes at least one of the following variables: a duration ratio of a period, within the utterance period of the note, to which the singing expression is to be applied; a speed variable indicating, in a case where a voice component of the singing expression is to be at least either extended or contracted in the direction of the time axis, a degree of extension or contraction, the speed variable also indicating which of a start side or an end side of the voice component is to be extended or contracted; a maximum value of an application ratio of the singing expression to be applied to the voice of the note to be synthesized; a temporal change of the application ratio from a value at a start of the singing expression to the maximum value; and a temporal change of the application ratio from the maximum value to a value at an end of the singing expression.
(128) Preferably, each variable in the control information is set in accordance with an operation received from the user with respect to each of a plurality of feature amounts. In this configuration, because variables in the control information are set in accordance with instructions received from the user for each of the plurality of feature amounts relating to singing expressions, a variety of voices that reflect the intention and/or the preferences of the user can be synthesized.
(129) Preferably, the voice of the note is synthesized by mixing of a first sound unit and a second sound unit, the second sound unit having a singing expression selected by the user from among the plurality of singing expressions. In this configuration, because multiple sound units are mixed to synthesize a voice, there will be less deterioration in the voice whereby a more natural-sounding voice can be synthesized as compared to a configuration in which a single sound unit is modified to synthesize a voice.
(130) An editing apparatus for voice synthesis according to the present invention includes: a display device; an input device configured to receive operations from a user; and a processor configured: to control the display device to display a note icon that represents a note of a voice to be synthesized and an indicator that is moved in accordance with an operation received by the input device; to control the display device to display first options that belong to a first layer among layers in a hierarchical structure, and when the indicator is moved into an area corresponding to a particular option from among the first options, to control the display device to display second options that correspond to the particular option and belong to a second layer that is below the first layer in the hierarchical structure; and to select a singing expression to be applied to the note, from among a plurality of singing expressions, according to selection operations received from the user by the input device, wherein the selection operations are performed in a layer-by-layer manner for the layers in the hierarchical structure. In the above configuration, the user is allowed to select one of the singing expressions by selecting an option for each layer in a step-by-step manner. Therefore, the user can more easily find and select a desired singing expression as compared to a configuration in which all possible singing expressions selectable by the user are shown at once and the user has to select a desired singing expression from them.
(131) The present invention is described herein with reference to the above embodiments, but the present invention is not limited to the above embodiments. The configurations and details of the present invention may be modified in a variety of ways that a skilled person in the art would understand as remaining within the scope of the present invention. This application claims priority from Japanese Patent Application No. 2015-164279, filed on Aug. 21, 2015, and the disclosures of which are hereby incorporated by reference in their entirety.
DESCRIPTION OF REFERENCE SIGNS
(132) 100 . . . voice synthesis apparatus, 10 . . . arithmetic processing device, 12 . . . storage device, 14 . . . display device, 16 . . . input device, 18 . . . sound output device, 22 . . . instruction receiver, 24 . . . display controller, 25 . . . selection handler, 26 . . . information manager, 28 . . . voice synthesizer, 40 . . . editing screen, 42 . . . note icon.