Method for converting an image and corresponding device
11202032 · 2021-12-14
Assignee
Inventors
Cpc classification
H04N7/0125
ELECTRICITY
H04N9/77
ELECTRICITY
International classification
H04N9/77
ELECTRICITY
Abstract
A method is described for converting an input image into an output image, the output image including an output luminance component made of elements. The method includes: obtaining an input luminance component from the input image; determining the output luminance component, the respective ranges of the output luminance component element values and input luminance component element values being of different range extension, the determining step including: —determining a first intermediate luminance component from the input luminance component and an exponent, —obtaining a mapping profile allowing for mapping a luminance component based on the input luminance component into the output luminance component, —determining a second intermediate luminance component from the input luminance component and the obtained mapping profile, —determining the output luminance component from the first and second intermediate luminance components; and converting the input image into the output image.
Claims
1. A method for converting an input image into an output image, the output image including an output luminance component made of elements, the method comprising: obtaining an input luminance component from the input image, determining the output luminance component, wherein a range of the output luminance component element values differs from a range of the input luminance component element values, said determining step including: determining a first intermediate luminance component from the input luminance component and an exponent, obtaining a mapping profile that maps a luminance component based on the input luminance component into the output luminance component, determining a second intermediate luminance component from the input luminance component and the obtained mapping profile, determining the output luminance component from the first and second intermediate luminance components, and converting the input image into the output image; wherein the step of determining a first intermediate luminance component is performed either before or after the step of determining a second intermediate luminance component.
2. The method of claim 1, wherein the step of determining a second intermediate luminance component is performed before the step of determining a first intermediate luminance component, the first intermediate luminance component being determined from the second intermediate luminance component and the exponent.
3. The method of claim 1, wherein the step of determining a first intermediate luminance component is performed before the step of determining a second intermediate luminance component, the second intermediate luminance component being determined from the first intermediate luminance component and the obtained mapping profile.
4. The method of claim 1, wherein the second intermediate luminance component is determined by using an invertible formulation.
5. The method of claim 1, wherein the mapping profile comprises extrinsic parameters.
6. The method of claim 5, further comprising displaying the output image on a display device, wherein said extrinsic parameters characterize the display device.
7. The method of claim 1, further comprising a calculating step in which the exponent is calculated as a function of a logarithm of a global luminosity level of the input image perceived by an observer.
8. The method of claim 7, wherein the exponent is calculated in order to monotonically decrease when the median of the input image luminance component increases.
9. The method of claim 8, wherein the exponent γ is calculated according to the following function:
10. The method of claim 1, wherein the mapping profile is an inverse tone mapping profile.
11. The method of claim 10, wherein the second intermediate luminance component is determined as follows:
12. The method of claim 1, wherein the mapping profile is a tone mapping profile.
13. The method of claim 12, wherein the second intermediate luminance component is determined as follows:
14. The method of claim 1, further comprising obtaining selection information, and obtaining the mapping profile by selecting a mapping profile from among a plurality of predetermined mapping profiles according to said selection information.
15. A device for converting an input image into an output image, the output image including an output luminance component made of elements, the device being configured to execute the following steps: obtaining an input luminance component from the input image, determining the output luminance component, wherein a range of the output luminance component element values differs from a range of the input luminance component element values, said determining step including: determining a first intermediate luminance component from the input luminance component and an exponent, obtaining a mapping profile that maps a luminance component based on the input luminance component into the output luminance component, determining a second intermediate luminance component from the input luminance component and the obtained mapping profile, determining the output luminance component from the first and second intermediate luminance components, and converting the input image into the output image; wherein the step of determining a first intermediate luminance component is performed either before or after the step of determining a second intermediate luminance component.
16. The device of claim 15, wherein the step of determining a second intermediate luminance component is performed before the step of determining a first intermediate luminance component, the first intermediate luminance component being determined from the second intermediate luminance component and the exponent.
17. A method for converting an input image into an output image, the output image including an output luminance component made of elements, the method comprising: obtaining an input luminance component from the input image, determining the output luminance component, wherein a range of the output luminance component element values differs from a range of the input luminance component element values, converting the input image into the output image, wherein the step of determining the output luminance component includes: obtaining a mapping profile, determining an intermediate luminance component from the input luminance component and the obtained mapping profile, and determining the output luminance component from the intermediate luminance component and an exponent.
18. The method of claim 17, wherein the exponent is calculated from the intermediate luminance component.
19. A method for converting an input image into an output image, the output image including an output luminance component made of elements, the method comprising: obtaining an input luminance component from the input image, determining the output luminance component, wherein a range of the output luminance component element values differs from a range of input luminance component element values, converting the input image into the output image, wherein the step of determining the output luminance component includes: determining an intermediate luminance component from the input luminance component and an exponent, obtaining a mapping profile, and determining the output luminance component from the intermediate luminance component and the obtained mapping profile.
20. The method of claim 19, wherein the exponent is calculated from the input luminance component.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) Further advantages of the present invention will become apparent to those skilled in the art upon examination of the drawings and detailed description. Embodiments of the invention will now be described, by way of example only, and with reference to the following drawings.
(2)
(3)
(4)
(5)
(6)
(7)
(8) The invention will now be described by means of specific non-limiting exemplary embodiments and by reference to the figures.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
(9)
(10) Alternatively, M2 could be mixing means like a video switcher (Sony MVS-8000X for example) configured for mixing signals in different formats. Then the other means M2 provides images resulting from the mix of different input signals including the one delivered by the converting means M1.
(11) The converting means M1 are able to process the luminance component of the input image I.sub.in in order to provide the luminance component of the output image I.sub.out. Both luminance components have different range extensions. For example, the input image I.sub.in can be a SDR image and the output image I.sub.out an HDR image. In this case, it is referred to tone expansion or inverse (or reverse) tone mapping (ITMO). On the contrary the input image 6 can be a HDR image and the output image I.sub.out an SDR image. It is then referred to tone mapping or tone compression (TMO). For example, ITMO means function is to extend the range of values of a first interval [0:1] at a second range of values [0:L.sub.max], where L.sub.max represents the length of the second interval, L.sub.max being an integer greater than 1. For example, with an HDR screen standard, if the peak luminance of the screen is 1000 nits or cd/m2, then L.sub.max is 1000 (see Report ITU-R BT.2446-0 (April/2019) “Methods for conversion of high dynamic range content to standard dynamic range content and vice-versa”.)
(12) The converting means M1 may be configured to perform round-tripping from SDR image to HDR image and back to SDR image or vice versa from HDR image to SDR image and back to HDR image.
(13)
(14) The static mapping allows for adapting the luminance component of the delivered image to features not related to the image content, such as a capturing camera, a display device, a video switcher, information provided by other channels or preferences of a director of photography for instance. A set of parameters defines a mapping profile for the display device or stylistic choices for example. If the display device is switched by another one, another mapping profile is obtained.
(15) Means M12 is able to provide a suitable profile to the means M10 for performing the dynamic mapping. The appropriate profile may be selected among a plurality of profiles, each one corresponding to a type of display device, for example a Sony KD-55X9305C or a LG Oled 55EG960V.
(16) The dynamic mapping allows for adapting the luminance component of the delivered image to the image content. It aims at reflecting at the best, the image characteristics and style (landscape, portrait, neutral, standard . . . ). To perform the dynamic mapping, the means M11 is linked to other means M13 which is able to provide information about the input image content I.sub.in. Such information can be for example the computed median value of the input luminance component.
(17) Having distinct first and second mapping steps, allows for separating the image-adaptive aspect of luminance expansion or compression from the overall behavior of the luminance expansion or compression operator, overall behavior induced par external factors mentioned above. This separation of both mapping steps according to the invention, involves a more robust process. Indeed changing of display device does not require anymore to perform a new optimization of the dynamic mapping, which is a complex operation. Moreover, the inventors have observed that performing the static mapping on a luminance component instead of a color component prevent for artefacts, such as shifts in image hues.
(18)
(19) A first step S11 comprises obtaining an input image I.sub.sdr, for instance an SDR image in a nonlinear R′G′B′ (for “Red Green Blue”) space, as specified in ITU-R BT.709 standard. Of course, other standard may be considered like BT.2020 standard.
(20) A second step S21 consists in obtaining an input luminance component L.sub.lin, which is preferably here a linear luminance component (nonlinear luminance components may be considered for certain standards). For example, this luminance component L.sub.lin is determined by first converting the input image I.sub.sdr by applying first an inverse Opto-Electrical Transfer Function (“OETF.sup.−1”). This Transfer Function represents the action of a sensor, converting brightness to data. The conversion allows for obtaining an image in a linear RGB space. Then the RGB image is converted in another linear space where the converted image comprises three components X, Y and Z, Y being a linear luminance component, called L.sub.lin in the following description
(21) In a third step S31, a static mapping is performed from the input linear luminance component L.sub.in into an intermediate luminance component L.sub.m. The mapping is formulated here as follows:
(22)
(23) Where m.sub.a, m.sub.b, m.sub.c are parameters, whose predetermined and set values defining a mapping profile. This mapping profile as been obtained by setting the value of the parameters among plurality of possible values, to define the overall behavior of the luminance expansion operator, induced par external factors mentioned above.
(24) For example in one implementation, the parameters m.sub.a, m.sub.b, m.sub.c may be set as follows to get a profile that is well adapted for commercial HDR television:
m.sub.a=1.5284
m.sub.b=0.5279
m.sub.c=0.7997.
(25) For example, the parameters are determined according to visual evaluation. In another example, the parameters may be determined based on the features defining the display device.
(26) The formulation (1) is an example allowing for carrying out a stable static mapping, with a same behavior for different parameter values. Advantageously, this formulation is invertible, making it useable for round-tripping SDR to HDR and back to SDR.
(27) Using the proposed formulation (1), a luminance mapping profile can be defined achieving a desired, overall behavior or style (e.g. an overall brighter expansion for a particular client wanting a strong ‘HDR effect’, or a more classic expansion or compression when addressing consumer televisions with mid-level specs for instance).
(28) This is particularly useful as it means that the formulation used to compute the image-adaptive exponent can remain unchanged, only the parameter values for the static mapping are replaced.
(29) Other formulations may be used, as the one below:
(30)
(31) The parameters m.sub.a, m.sub.b, m.sub.c and m.sub.d, whose predetermined and set values define another predetermined mapping profile. Their value may be set as follows to get a profile that is well adapted for commercial HDR television:
m.sub.a=1.3475
m.sub.b=0.0070
m.sub.c=0.2844
m.sub.a=−0.3285.
(32) One skilled in the art could propose other parameterizable formulations adapted for computing intermediate luminance representation L.sub.m from the input linear luminance component L.sub.lin, regarding the parameters on the selected mapping profile.
(33) Preferably, if values of the input linear luminance component L.sub.lin belong to the range [0,1], the values of the intermediate luminance values L.sub.m belong to the same range.
(34) Preferably, the formulation is invertible for enabling round-trippings.
(35) In another embodiment, different mapping profiles may be available, for instance loaded in a dedicated memory. A first profile can be well adapted for displaying RGB images for commercial HDR television and a second profile can be well adapted for displaying images containing additional information defining a mask to apply on pictures. The mask could define a logo or provide statistics information to be displayed superimposed on background images capturing a sport game for instance. The mask could be delivered via an input alpha channel or, if a predefined mask-pattern like a logo is used, pre-loaded in a memory. The definition of such a second mapping profile allows for reducing uncomfortable effects which affect certain colors (like the whites) of superimposed pattern, when displayed. According to an information (for example it is detected that data related to a mask to superimpose is provided via the alpha channel), the first or the second mapping profile is selected.
(36) Steps S21 and S31 may be implemented by the means M10 performing the static mapping.
(37) In a step S41, an exponent γ is calculated from the intermediate luminance component values L.sub.m.
(38) Preferably but not exclusively, the exponent γ is computed through a monotonically decreasing function depending of the relevant luminance component based on the linear input luminance component, and more specifically from the median of this relevant luminance component values (here the intermediate luminance component). Alternatively, the mean or the geometric mean of this relevant luminance component values can be considered.
(39) The patent application WO/2017/103399 provides an embodiment for computing a value of the exponent used for the expansion.
(40) Alternatively, the exponent may be calculated as follows:
(41)
(42) where g is a gain parameter, o is an offset and L.sub.med is based on the median of the intermediate luminance component. Alternatively, L.sub.med is obtained as follows:
L.sub.med=max(median(L.sub.m),τ.sub.med) (4)
(43) where τ.sub.med is set to a minimum threshold, for example τ.sub.med=0.00001.
(44) Advantageously, the formulation (4) for obtaining L.sub.med allows for a more stable performance for very dark images where the median value may be very low.
(45) The gain parameter g may be set as g=0.06 and the offset parameter may be set as o=1.0. The resulting expansion exponent γ may be preferably limited within a range of 1.0 and 2.0 to ensure that the resulting image to display will not lead to results that are too extreme in outlier cases (e.g. very dark or very bright images). Limiting the range of values that the exponent γ can take, has also the additional benefit of a more stable performance may be obtained in cases of quickly varying video content, where the tone expansion is adapted to each frame.
(46) Then a step S51 comprises obtaining an expanded luminance component L.sub.hdr, as follows:
L.sub.hdr=(L.sub.m).sup.γ. (5)
(47) The steps S41 and S51 may be implemented by the means M11 performing the dynamic mapping.
(48) Finally in this example, a step S61 allows for rendering the image on a display device. It comprises scaling the expanded luminance component L.sub.hdr by using the corresponding RGB components (for “Red Green Blue”), the linear luminance component L.sub.lin, and possibly the peak of luminance L.sub.max accordingly. Scaled RGB.sub.hdr components are obtained and converted in the nonlinear R′G′B′ space by applying an OETF. Optionally the saturation of each component may be adjusted, for instance according to one the methods described in R. Fattal, D. Lischinski et M. Werman, «Gradient domain high dynamic range compression» ACM transactions on graphics (TOG), vol. 21, no 13, pp. 249-256, 2002, or R. Mantiuk, R. Mantiuk, A. Tomaszewska et W. Heidrich, «Color correction for tone mapping» Computer Graphics Forum, vol. 28, n° 12, pp. 193-202, 2009.
(49) A final conversion provides colour intensities in the Y′CbCr format which decomposes the colour intensities into a luminance component Y′ separated from the chrominance components Cb, Cr. This format Y′CbCr is a way of representing the colour space in video that is well suited to the transmission problematics. These components are encoded on 10 bits. Embodiments of the conversion step S61 is described in the literature, for example in Poynton, Charles “Digital video and HD: Algorithms and Interfaces.” Elsevier, 2012.)
(50) The step S61 may be implemented by the means M2 in case it is displaying means.
(51)
(52) The converting means M1 comprises in this order, the second mapping means M11′ for performing a second mapping step, which is referred to be the dynamic mapping and the first mapping means M10′ for performing the first mapping step, the static mapping. Both means M11′ and M10′ are arranged in series.
(53) As for the first embodiment, means M12′ is able to provide a suitable profile to the means M10′ for performing the dynamic mapping. To perform the dynamic mapping, the means M11′ is linked to other means M13′ which is able to provide information about the input image content I.sub.in.
(54)
(55) A first step S12 comprises obtaining an input image I.sub.sdr, for instance an SDR image.
(56) A second step S22 consists in computing an input linear luminance component L.sub.lin. Both first and second steps S21 and S22 may be similar to the first and second step S11 and S21 of the first embodiment.
(57) Then in a third step S32 the exponent γ is calculated from the input linear luminance component L.sub.lin. If applying the formulation (3) mentioned above, the exponent is calculated as follows:
(58)
(59) where g is a gain parameter, o is an offset and L.sub.med is based on the median of the input linear luminance component L.sub.lin. Alternatively, L.sub.med is obtained as follows:
L.sub.med=max(median(L.sub.lin),τ.sub.med) (6)
(60) where τ.sub.med is set to a minimum threshold, for example τ.sub.med=0.00001.
(61) Then a step S42 comprises obtaining an intermediate luminance component L.sub.m, as follows:
L.sub.m=(L.sub.lin).sup.γ (7)
(62) In a fifth step S51, the static mapping is performed from the intermediate luminance component L.sub.m into an expanded luminance component L.sub.hdr. As for the first embodiment, the static mapping is realized from appropriate predetermined and set parameter values and a suitable formulation, like one of the formulations (1) or (2) proposed above.
(63) Finally a step S62 comprises processing the expanded luminance component L.sub.hdr to obtain Y′CbCr components, as described by reference to the step S61 for instance.
(64)
(65) A first step S13 comprises obtaining an input image I.sub.hdr, which is a HDR image here.
(66) A second step S23 consists in computing an input linear luminance component L.sub.lin using the appropriate known conversion process.
(67) A third steps S33 comprises computing an exponent γ for luminance compression from the input linear luminance component L.sub.lin, following the same formulation (3) mentioned above, as follows:
(68)
(69) where g is a gain parameter, o is an offset and L.sub.med is based on the median of the input linear luminance component L.sub.lin. Alternatively, L.sub.med is obtained as follows:
L.sub.med=max(median(L.sub.lin),τ.sub.med)
(70) where τ.sub.med is set to a minimum threshold, for example τ.sub.med=0.00001.
(71) In a fourth step S43, an intermediate compressed luminance component L.sub.m is obtained as follows:
(72)
(73) In a fifth step S53, a static mapping is performed from the intermediate compressed luminance component L.sub.m into a compressed SDR luminance component L.sub.sdr. The mapping is formulated as follows:
(74)
(75) where m.sub.a, m.sub.b, m.sub.c are parameters whose predetermined and set values define another predetermined mapping profile. In one implementation, the parameters m.sub.a, m.sub.b, m.sub.c may be set as follows to obtain a profile that is well adapted for typical commercial television, symmetrically to the luminance expansion implementation described previously:
m.sub.a=1.5284
m.sub.b=0.5279
m.sub.c=0.7997.
(76) Finally a step S63 comprises processing the compressed luminance component L.sub.sdr to obtain Y′CbCr components according to the appropriate standard, as described by reference to the step S61 for instance.
(77) Although the present invention has been described hereinabove with reference to specific embodiments, the present invention is not limited to the specific embodiments, and modifications will be apparent to a skilled person in the art which lie within the scope of the present invention.
(78) Many further modifications and variations will suggest themselves to those versed in the art upon making reference to the foregoing illustrative embodiments, which are given by way of example only and which are not intended to limit the scope of the invention, that being determined solely by the appended claims. In particular the different features from different embodiments may be interchanged, where appropriate.
(79) In the claims, the word “comprising” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality. The mere fact that different features are recited in mutually different dependent claims does not indicate that a combination of these features cannot be advantageously used.