Full-color visibility model using CSF which varies spatially with local luminance
10282801 ยท 2019-05-07
Assignee
Inventors
- Alastair M. Reed (Lake Oswego, OR)
- Kristyn R. Falkenstern (Portland, OR)
- David Berfanger (Vancouver, WA, US)
- Yang Bai (Beaverton, OR)
Cpc classification
G06T1/0028
PHYSICS
H04N2201/3233
ELECTRICITY
H04N1/6005
ELECTRICITY
G06T2201/0083
PHYSICS
International classification
H04N1/32
ELECTRICITY
H04N7/167
ELECTRICITY
Abstract
The present disclosure relate generally to image signal processing, color science and signal encoding. Signal encoding can be applied to color image data through use of a luminance contrast sensitivity function and a chrominance contrast sensitive function. Of course, other features, combinations and claims are disclosed as well.
Claims
1. An apparatus comprising: memory storing: i) a luminance contrast sensitivity function (CSF1), ii) a chrominance contrast sensitivity function (CSF2), and iii) data representing color imagery; means for estimating degradation of image areas associated with an application of signal encoding in the data representing color imagery by applying the CSF1 and the CSF2 to the data representing color imagery, in which the CSF1 varies depending on luminance values associated with local regions of the data representing color imagery, and in which the CSF1 is used for processing luminance data and the CSF2 is used for processing chrominance data; and means for changing the data representing color imagery with signal encoding, in which the signal encoding is guided based on results obtained from said means for estimating including estimated degradation of image areas.
2. The apparatus of claim 1 in which the CSF1 varies spatially.
3. The apparatus of claim 2 in which the CSF2 varies spatially in terms of spatial width.
4. The apparatus of claim 1 in which the CSF1 varies spatially in terms of spatial width.
5. The apparatus of claim 1 in which the means for estimating degradation produces image blurring as the estimated degradation, in which the CSF1 varies so that relatively more blurring occurs as luminance of a local image region decreases.
6. The apparatus of claim 1 in which the means for changing utilizes results obtained from the means for estimating by varying signal encoding strength across different image areas of the data representing color imagery based on estimated degradation of the different image areas.
7. The apparatus of claim 6 in which estimated degradation of the signal encoding across the different image areas comprises uniform estimated degradation.
8. The apparatus of claim 1 further comprising means for applying an attention model to the data representing color imagery to predict visual traffic areas.
9. The apparatus of claim 8 in which the means for changing utilizes predicted visual traffic areas and the estimated degradation of image areas.
10. The apparatus of claim 1 in which the chrominance contrast sensitivity function (CSF2) comprises a blue-yellow contrast sensitivity function and a red-green contrast sensitivity function.
11. The apparatus of claim 1 in which the CSF2 varies depending on luminance values associated with local regions of the obtained color image data.
12. The apparatus of claim 1 in which said means for changing the data representing color imagery with signal encoding encodes a payload into the data representing color imagery.
13. The apparatus of claim 1 in which the color imagery comprises video.
14. A method comprising: obtaining color image data; changing the color image data with signal encoding, the signal encoding comprising a payload, said changing yielding encoded color image data; comparing the encoded color image data to the color image data to determine a visibility map, the visibility map comprising a luminance contrast sensitivity function (CSF1) and a chrominance contrast sensitivity function (CSF2); weighting the signal encoding per the visibility map so that local image areas within the color image data are weighted differently, said weighting yielding weighted signal encoding; encoding the color image data with the weighted signal encoding to yield locally varied encoded color image data.
15. The method of claim 14 in which the CSF1 introduces image blurring, and in which the CSF1 varies so that relatively more blurring occurs as luminance of a local image region decreases.
16. The method of claim 15 in which said weighting varies signal encoding strength across different local image regions of the color image data to yield uniform visibility of the signal encoding across the color image data.
17. The method of claim 16 in which the color image data represents video data.
18. The method of claim 14 in which the CSF1 varies spatially in terms of spatial width.
19. The method of claim 14 in which the CSF2 varies spatially in terms of spatial width.
20. The method of claim 14 in which the CSF2 comprises a blue-yellow contrast sensitivity function and a red-green contrast sensitivity function.
21. A non-transitory computer readable medium comprising instructions, which when executed configure one or more processors to: access color image data; change the color image data with signal encoding, the signal encoding comprising a payload, said changing yielding encoded color image data; compare the encoded color image data to the color image data to determine a visibility map, the visibility map comprising a luminance contrast sensitivity function (CSF1) and a chrominance contrast sensitivity function (CSF2); weight the signal encoding per the visibility map so that local image areas within the color image data are weighted differently to yield weighted signal encoding; encode the color image data with the weighted signal encoding to yield locally varied encoded color image data.
22. The non-transitory computer readable medium of claim 21 in which the CSF1 introduces image blurring, and in which the CSF1 varies so that relatively more blurring occurs as luminance of a local image region decreases.
23. The non-transitory computer readable medium of claim 22 in which the signal encoding strength varies across different local image regions of the color image data to yield uniform visibility of the signal encoding across the color image data.
24. The non-transitory computer readable medium of claim 21 in which the CSF1 varies spatially in terms of spatial width.
25. The non-transitory computer readable medium of claim 21 in which the CSF2 varies spatially in terms of spatial width.
26. The non-transitory computer readable medium of claim 21 in which the CSF2 comprises a blue-yellow contrast sensitivity function and a red-green contrast sensitivity function.
27. The non-transitory computer readable medium of claim 21 in which the color image data represents video data.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
DETAILED DESCRIPTION
(27) Portions of the following disclosure discusses a digital watermarking technique that utilizes at least two chrominance channels (also called color planes, color channels and/or color direction). Chrominance is generally understood to include information, data or signals representing color components of an image or video. In contrast to a color image or video, a grayscale (monochrome) image or video has a chrominance value of zero.
(28) Media content that includes a color image (or color video) is represented in
(29) Let's first discuss the additive and subtractive effects on
(30) Now let's consider watermarking in the context of
(31) In a case where a media signal includes (or may be broken into) at least two chrominance channels, a watermark embedder may insert digital watermarking in both the a color direction (
WMa=a(channel)+wm(1)
WMb=b(channel)wm(2)
(32) WMa is a watermarked a channel, WMb is a watermarked b channel, and wm represents a watermark signal. A watermarked color image (including L and WMb and WMa) can be provided, e.g., for printing, digital transfer or viewing.
(33) An embedded color image is obtained (from optical scan data, memory, transmission channel, etc.), and data representing the color image is communicated to a watermark detector for analysis. The detector (or a process, processor or electronic processing circuitry used in conjunction with the detector) subtracts WMb from WMa resulting in WMres as shown below:
WMres=WMaWMb(3)
WMres=(a+wm)(bwm)(4)
WMres=(ab)+2*wm(5)
(34) This subtraction operation yields reduced image content (e.g.,
(35)
(36) A watermark detector may extract or utilize characteristics associated with a synchronization signal (if present) from a frequency domain representation of WMres. The detector may then use this synchronization signal to resolve scale, orientation, and origin of the watermark signal. The detector may then detect the watermark signal and obtain any message or payload carried thereby.
(37) To even further illustrate the effects of improving the watermark signal-to-media content ratio with our inventive processes and systems, we provide some additive and subtractive examples in the content of watermarking.
(38) For the following example, a watermark signal with the same polarity is embedded in each of the a color channel and the b color channel. The same signal polarity is represented by a plus (+) sign in equations 6 and 7.
WMa=a+wm(6)
WMb=b+wm(7)
(39) WMa is a watermarked a channel, WMb is a watermarked b channel, and wm represents a watermark signal. A watermarked color image (including L and WMb and WMa) can be provided, e.g., for printing, digital transfer or viewing.
(40) An embedded color image is obtained, and data representing the color image is communicated to a watermarked detector for analysis. The detector (or a process, processor, or electronic processing circuitry used in conjunction with the detector) adds the a and b color channels to one another (resulting in WMres) as shown below:
WMres=WMa+WMb(8)
WMres=(a+wm)+(b+wm)(9)
WMres=(a+b)+2*wm(10)
(41) This addition operation results in increased image content (e.g.,
(42) By way of further example, if WMb is subtracted from WMa (with watermark signals having the same polarity), the following results:
WMres=WMaWMb(11)
WMres=(a+wm)(b+wm)(12)
WMres=(ab)+0*wm(13)
(43) A subtraction or inverting operation in a case where a watermark signal includes the same polarity decreases image content (e.g.,
(44)
(45) With reference to
(46) With reference to
(47) In addition to the Lab color scheme discussed above, a watermark signal may be embedded in color image (or video) data represented by RGB, Yuv, Ycc, CMYK or other color schemes, with, e.g., a watermark signal inserted in a first chrominance direction (e.g., red/green direction, similar to that discussed above for the a channel) and a second chrominance direction (e.g., a blue/yellow direction, similar to that discussed above for the b channel). For watermark signal detection with an alternative color space, e.g., an RGB or CMYK color space, an image can be converted to Lab (or other color space), or appropriate weights of, e.g., RGB or CMY channels, can be used. For example, the following RGB weights may be used to calculate ab: Chrominance Difference=0.35*R1.05*G+0.70*B+128, where R, G and B are 8-bit integers.
(48) Further Considerations of Video
(49) The human contrast sensitivity function curve shape with temporal frequency (e.g., relative to time) has a very similar shape to the contrast sensitivity with spatial frequency.
(50) Successive frames in a video are typically cycled at about at least 60 Hz to avoid objectionable visual flicker. So-called flicker is due to the high sensitivity of the human visual system (HVS) to high temporal frequency changes in luminance. The human eye is about ten (10) times less sensitive to high temporal frequency chrominance changes.
(51) Consider a video sequence with frames as shown in
(52) In order to recover the watermark, pairs of frames are processed by a watermark detector, and the a channels are subtracted from each other as shown below.
Det_a=(a1+wm)(a2wm)=(a1a2)+2*wm(14)
(53) Det_a refers to watermark detection processing of the a channel. Because of the temporal correlation between frames, the image content in equation 14 is reduced while the watermark signal is reinforced.
(54) In a similar way the b channels are also subtracted from each other
Det_b=(b1wm)(b2+wm)=(b1b2)2*wm(15)
(55) Det_a refers to watermark detection processing of the b channel. Equation 14 and 15 are then subtracted from each other as shown below in equation 16.
Det_aDet_b=(a1a2+2*wm)(b1b22*wm)=(a1a2)(b1b2)+4*wm(16)
(56) In generally, related (but not necessarily immediately adjacent) frames will have spatially correlated content. Because of the spatial correlation between the a and b frames, the image content is reduced while the watermark signal is reinforced. See equation 16.
(57) For any one pair of frames selected by a watermark detector, the polarity of the watermark could be either positive or negative. To allow for this, the watermark detector may examine both polarities.
(58) Watermark Embedding for Spot Colors
(59) Product packaging is usually printed in one of two ways:
(60) 1. Process color printing using cyan, magenta yellow and/or black (CMYK)
(61) 2. Spot color printing (e.g., using special Pantone color or other ink sets)
(62) The majority of packaging is printed using spot colors mainly for reasons of cost and color consistency, and to achieve a wide color gamut over various packaging. Some conventional watermarking techniques embed digital watermarks in either CMYK for printed images or RGB for digital images that are being displayed. But how to embed a watermark with a spot color?
(63) An improvement addresses problem associated with watermarking spot color images. Preferably, packaging contains two (2) or more spot colors (e.g., printed cooperatively to achieve a certain color consistency). Each different color is altered to collectively carry a watermark signal. A maximum signal strength within a user selectable visibility constraint with watermark in at least two (2) of the spot.
(64) A maximized watermark signal is embedded preferably by modulating the spot color inks within a certain visibility constraint across the image. The approach models a color (ink) in terms of CIE Lab values. Lab is a uniform perceptual color space where a unit difference in any color direction corresponds to an equal perceptual difference.
(65) The Lab axes are then scaled for the spatial frequency of the watermark being added to the image, in a similar manner to the Spatial CieLab model by X. Zhang and B. A. Wandell, e.g., A spatial extension of CIELAB for digital color image reproduction, in Proceedings of the Society of Information Display Sumposium (SID '96), vol. 27, pp. 731-734, San Jose, Calif., USA, June 1996. This is a uniform perceptual color space which we will call SLAB, where a unit difference in any color direction corresponds to an equal perceptual difference due to the addition of a watermark signal at that spatial frequency.
(66) The allowable visibility magnitude in SLAB is scaled by spatial masking of the cover image. Spatial masking of the cover image can include the techniques described by Watson in US Published Patent Application No. US 2006-0165311 A1, which is hereby incorporated by reference in its entirety, and can be used to scale the allowable visibility across the image. This is a uniform perceptual color space which we will call VLAB, where the visibility circle is scaled to correspond to an equal perceptual difference due to the addition of a watermark signal at that spatial frequency for that particular image.
(67) The chrominance embedding techniques discussed above forms the foundation for the present watermark embedding techniques. A related discussion is found in U.S. patent application Ser. No. 13/975,919, filed Aug. 26, 2013, now U.S. Pat. No. 9,449,357, under the section Chrominance watermark to embed using a full color visibility model, which uses an iterative embed technique to insert a maximum watermark signal into CMYK images.
(68) The spot color technique described extends this work to embedding that supports special color inks (e.g., spot colors) used in packaging and uses a full color visibility model with spatial masking. A geometric enumerated embed approach can be used to evaluate a range of possible ink changes, which meet the user selected visibility constraint and press constraints. The set of allowable ink changes are evaluated to choose the pair of ink changes which result in the maximum signal strength while meeting the visibility and press constraints.
(69)
(70) A user can insert a maximum watermark signal, while meeting any pre-required visibility constraint. The method has been applied to the case of two spot colors and images have been produced which are more than twice as robust to Gaussian noise as a single color image which is embedded using a luminance only watermark to the same visibility.
(71) A method has been described which allows an image containing 2 or more spot colors to be embedded with a watermark in 2 of the spot colors, with the maximum signal strength within a user selectable visibility constraint.
(72) A look-up table based approach can be used for given colors at given locations, and can easily be extended to 3 or more dimensions while still being computationally reasonable.
(73) Additional related disclosure is found in U.S. patent application Ser. No. 13/975,919, now U.S. Pat. No. 9,449,357, under the heading sections Geometric Enumerated Chrominance Watermark Embed for Spot Colors and Watermarking Embedding in Optimal Color Direction.
(74) Full-Color Visibility Model
(75) A full color visibility model has been developed that uses separate contrast sensitivity functions (CSFs) for contrast variations in luminance and chrominance (red-green and blue-yellow) channels. The width of the CSF in each channel can be varied spatially depending on the luminance of the local image content. The CSF can be adjusted so that relatively more blurring occurs as the luminance of the local region decreases. The difference between the contrast of the blurred original and marked image can be measured using a color difference metric.
(76) This spatially varying CSF performed better than a fixed CSF in the visibility model, approximating subjective measurements of a set of test color patches ranked by human observers for watermark visibility.
(77) A full color visibility model can be a powerful tool to measure visibility of an image watermark. Watermarks used for packaging can be inserted in the chrominance domain to obtain the best robustness per unit visibility. A chrominance image watermark is preferably embedded in a way that the color component in the cover image is minimally altered and is hardly noticeable, due to human vision system's low sensitivity to color changes.
(78) One example of a color visibility model is discussed relative to Spatial CIELAB (S-CIELAB). The accuracy of this model was tested by comparing it to human subjective tests on a set of watermarked color patches. The model was found to significantly overestimate the visibility of some dark color patches. A correction can be applied to the model for the variation of the human contrast sensitivity function (CSF) with luminance. After luminance correction, better correlation was obtained with the subjective tests.
(79) The luminance and chrominance CSF of the human visual system has been measured for various retinal illumination levels. The luminance CSF variation was measured by Van Nes (1967) and the chrominance CSF variation by van der Horst (1969). These measurements show a variation in peak sensitivity of about a factor of 8 for luminance and 5 for chrominance over retinal illumination levels which change by about a factor of 100.
(80) Since the retinal illumination can change by about a factor of 100 between the lightest to darkest area on a page, the CSF peak sensitivity and shape can change significantly. The function is estimated by the average local luminance on the page, and a spatially dependent CSF is applied to the image. This correction is similar to the luminance masking in adaptive image dependent compression.
(81) The luminance dependent CSF performed better than a fixed CSF in the visibility model, when compared to subjective measurements of a set of test color patches ranked by human observers for watermark visibility. In some cases, we use a method of applying a spatially dependent CSF which depends on local image luminance.
(82) The visibility model can be used to embed watermark into images with equal visibility. During the embedding stage, the visibility model can predict the visibility of the watermark signal and then adjust the embedding strength. The result will be an embedded image with a uniform watermark signal visibility, with the embedding strength varying depending on the cover image's content.
(83) The following documents are hereby incorporated herein by reference: Lyons, et al. Geometric chrominance watermark embed for spot color, Proc. Of SPIE, vol. 8664, Imaging and Printing in a Web 2.0 World IV, 2013; Zhang et al. A spatial extension of CIELAB for digital color-image reproduction Journal of the Society for Information Display 5.1 (1997): 61-63; Van Nes et al. Spatial modulation transfer in the human eye, Journal of Optical Society of America, vol. 57, issue 3, pp. 401-406, 1967; Van der Horst et al. Spatiotemporal chromaticity discrimination, Journal of Optical Society of America, vol. 59, issue 11, 1969; and Watson, DCTune, Society for information display digest of technical papers XXIV, pp. 946-949, 1993.
(84) In some cases, even better results can be achieved by combining an attention model with our above visibility model when embedding watermarks in color image data. An attention model generally predicts where the human eye is drawn to when viewing an image. For example, the eye may seek out flesh tone colors and sharp contrast areas. One example attention model is described in Itti et al., A Model of Saliency-Based Visual Attention for Rapid Scene Analysis, IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 20, NO. 11, NOVEMBER 1998, pgs. 1254-1259, which is hereby incorporated herein by reference.
(85) High visual traffic areas identified by the attention model, which would otherwise be embedded with a relatively strong or equal watermark signal, can be avoided or minimized by a digital watermark embedder.
(86) Additional related disclosure is found in Appendix D, attached and included as part of this specification, and which is hereby incorporated herein by reference in its entirety.
(87) Disclosure from Appendix D is also provided below:
(88) Full-Color Visibility Model Using CSF Which Varies Spatially with Local Luminance
(89) Abstract:
(90) A full color visibility model has been developed that uses separate contrast sensitivity functions (CSFs) for contrast variations in luminance and chrominance (red-green and blue-yellow) channels. The width of the CSF in each channel is varied spatially depending on the luminance of the local image content. The CSF is adjusted so that more blurring occurs as the luminance of the local region decreases. The difference between the contrast of the blurred original and marked image is measured using a color difference metric.
(91) This spatially varying CSF performed better than a fixed CSF in the visibility model, approximating subjective measurements of a set of test color patches ranked by human observers for watermark visibility. The effect of using the CIEDE2000 color difference metric compared to CIEDE1976 (i.e., a Euclidean distance in CIELAB) was also compared.
(92) Introduction
(93) A full color visibility model is a powerful tool to measure the visibility of the image watermark. Image watermarking is a technique that covertly embeds additional information in a cover image, such that the ownership, copyright and other details about the cover image can be communicated. Watermarks used for packaging are inserted in the chrominance domain to obtain the best robustness per unit visibility. See Robert Lyons, Alastair Reed and John Stach, Geometric chrominance watermark embed for spot color, Proc. Of SPIE, vol. 8664, Imaging and Printing in a Web 2.0 World IV, 2013. The chrominance image watermark is embedded in a way that the color component in the cover image is minimally altered and is hardly noticeable, due to human vision system's low sensitivity to color changes.
(94) This visibility model is similar to Spatial CIELAB (S-CIELAB). See Xuemei Zhang and Brian A. Wandell, A spatial extension of CIELAB for digital color-image reproduction Journal of the Society for Information Display 5.1 (1997): 61-63. The accuracy of this model was tested by comparing it to subjective tests on a set of watermarked color patches. The model was found to significantly overestimate the visibility of some dark color patches. A correction was applied to the model for the variation of the human contrast sensitivity function (CSF) with luminance as described below. After luminance correction, good correlation was obtained with the subjective tests.
(95) The luminance and chrominance CSF of the human visual system has been measured for various retinal illumination levels. The luminance CSF variation was measured by Floris L. Van Nes and Maarten Bouman, Spatial modulation transfer in the human eye, Journal of Optical Society of America, vol. 57, issue 3, pp. 401-406, 1967 and the chrominance CSF variation by G J Van der Horst and Maarten Bouman, Spatiotemporal chromaticity discrimination, Journal of Optical Society of America, vol. 59, issue 11, 1969. These measurements show a variation in peak sensitivity of about a factor of 8 for luminance and 5 for chrominance over retinal illumination levels which change by about a factor of 100.
(96) Since the retinal illumination can change by about a factor of 100 between the lightest to darkest area on a page, the CSF peak sensitivity and shape can change significantly. The function is estimated by the average local luminance on the page, and a spatially dependent CSF is applied to the image. This correction is similar to the luminance masking in adaptive image dependent compression. See G J Van der Horst and Maarten Bouman, Spatiotemporal chromaticity discrimination, Journal of Optical Society of America, vol. 59, issue 11, 1969.
(97) The luminance dependent CSF performed better than a fixed CSF in the visibility model, when compared to subjective measurements of a set of test color patches ranked by human observers for watermark visibility. Results of our model with and without luminance correction are compared to S-CIELAB in Section 2, Visual Model Comparison. The method of applying a spatially dependent CSF which depends on local image luminance is described in Section 3, Pyramid Processing Method.
(98) The visibility model is then used to embed watermark into images with equal visibility. During the embedding stage, the visibility model can predict the visibility of the watermark signal and then adjust the embedding strength. The result will be an embedded image with a uniform watermark signal visibility, with the embedding strength varying depending on the cover image's content. This method was compared to a uniform strength embed in terms of both visibility and robustness, and the results are shown in Section 4, Watermark Equal Visibility Embed.
(99) Visual Model Comparison
(100) Psychophysical Experiment
(101) To test the full-color visibility model a psychophysical experiment was conducted. The percept of degradation caused by the watermark was compared to the results of the visibility model, as well as to the S-CIELAB metric.
(102) A set of observers were asked to rate their perception of the image degradation of 20 color patch samples using a quality ruler. The quality ruler (illustrated in
(103)
(104) All 22 participants passed the Ishihara color test. There were eight female and 14 male participants, with an average age of 43. Their professions and experience varied. Four people had never participated in a visibility experiment, 12 had some experience and six had participated on several occasions.
(105) Thumbnails of the 20 color patches are illustrated in
(106)
(107) The mean observer scores for the 20 color samples are plotted in
(108)
(109) Validation of the Visibility Model
(110) The motivation for the psychophysical experiment is to test how well the proposed full-color visibility model correlates to the perception of the degradation caused by the watermark signal. The model without and with the luminance adjustment are plotted in
(111)
(112)
(113) The addition of the luminance adjustment primarily affected the darker color patches, darkgreen, foliage and darkblue1. CIEDE94 and CIEDE2000 color difference models were also considered, however there was not a clear advantage to using the more complex formulas.
(114)
(115) The S-CIELAB values are also plotted against the mean observer response
(116) Two different methods were used to compare the different metrics to the observer data, Pearson's correlation and the coefficient of determination (R.sup.2). Both correlation techniques describe the relationship between the metric and observer scores. The coefficient indicates the relationship between two variables on a scale of +/1, the closer the values are to 1 the stronger the correlation is between the objective metric and subjective observer results. The correlations are summarized in Table 1.
(117) TABLE-US-00001 TABLE 1 Table 1: Pearson and R.sup.2 correlation between the observers' mean responses and the objective metrics. For both tests, the proposed full-color visibility model with the luminance adjustment shows the highest correlation. Visibility model using CIE E.sub.76 No Adjust With Adjust S-CIELAB Pearson 0.81 0.86 0.61 R.sup.2 0.70 0.85 0.38
(118) As shown in Table 1, all three objective methods have a positive correlation to the subjective results with both correlation methods. The full-color visibility model with the luminance adjustment had the highest correlation with both the Pearson and R.sup.2 correlation tests, while S-CIELAB had the lowest.
(119) Pyramid Processing Method
(120) In image fidelity measures, the CSF is commonly used as a linear filter to normalize spatial frequencies such that they have perceptually equal contrast thresholds. This can be described by the following shift invariant convolution:
(121)
where f(x,y) is an input image, h(x,y) is the spatial domain CSF, and {tilde over (f)}(x,y) is the frequency normalized output image.
(122) For our luminance dependent CSF model, we allow the CSF to vary spatially according to the local luminance of the image, i.e.:
(123)
(124) Since evaluating this shift variant convolution directly can be computationally expensive, we seek an approximation that is more efficient.
(125) The use of image pyramids for fast image filtering is well-established. An image pyramid can be constructed as a set of low-pass filtered and down-sampled images f.sub.l(x,y), typically defined recursively as follows:
(126)
for l>0 and generating kernel h.sub.0(m, n). It is easily shown from this definition that each level f.sub.l(x,y) of an image pyramid can also be constructed iteratively by convolving the input image with a corresponding effective kernel h.sub.l(m,n) and down-sampling directly to the resolution of the level, as follows:
(127)
where h.sub.l(m,n) is an l-repeated convolution of h.sub.0(m,n) with itself.
(128) For image filtering, the various levels of an image pyramid are used to construct basis images of a linear decomposition representing the point-spread response of the desired filtering, i.e.:
(129)
where a.sub.l is the coefficient of the basis function {tilde over (f)}.sub.l(x, y) obtained by up-sampling the corresponding pyramid level f.sub.l(x,y) back to the base resolution.
(130) We use the effective convolution kernel h.sub.l(x,y) as an interpolating kernel, i.e.,
(131)
such that each basis function {tilde over (f)}.sub.l(x, y) can be described by a simple shift-invariant convolution of the input image with a composite kernel {tilde over (h)}.sub.l(x,y):
{tilde over (f)}.sub.l(x,y)={tilde over (h)}.sub.l(x,y)*f(x,y),(8)
where {tilde over (h)}.sub.l(x,y)=h.sub.l(x,y)*h.sub.l(x,y). Thus, considering Eq. (6), we assert that the optimal representation is obtained by minimizing the sum of the squared error between the desired CSF and the Gaussian representation; i.e.,
(132)
and a=[a.sub.1, a.sub.2, . . . ]. This is a standard linear least-squares problem and can be solved using standard software packages, like Matlab or GNU Octave. Further, the optimization can be pre-calculated for each local luminance of interest and stored in a look-up table, noting that for our application each coefficient a.sub.1 is spatially varying according to the local luminance level L.sub.f=L.sub.f(x,y) of f(x,y), i.e.,
a.sub.l=a.sub.l(L.sub.f)=a.sub.l(L.sub.f(x,y)).
(133) While the development of our approach has been conducted for basis image at the resolution of the input image, the procedure can be conducted within a multi-resolution scheme, reducing the calculation of the spatially variant convolution in Eq. (3.2) into a pyramid reconstruction with spatially variant analysis coefficients.
(134) Watermark Equal Visibility Embed
(135)
(136)
(137) In terms of watermark detection, the embedding scheme with visibility model based adjustment can accommodate more watermark signal without creating a very noticeable degradation, thus making the detection more robust. To demonstrate the powerfulness of applying the visibility model, we performed a stress test with captures of 4 images from the two embedding schemes at various distances and perspectives. The other 3 images from the uniform visibility embedding are shown in
(138) These two tables show that the equal visibility embedding showed a significant visibility improvement over the uniform strength embedding scheme, together with robustness that was about the same or better.
(139)
(140)
(141)
(142) Table 2 shows standard deviation of the visibility maps on the 4 images from the two embedding schemes.
(143) TABLE-US-00002 TABLE 2 Test image Uniform strength embedding Equal visibility embedding Granola 18.32 9.71 Apple Tart 8.19 4.96 Giraffe Stack 16.89 11.91 Pizza Puff 11.81 8.27
(144) Table 3 shows detection rate on 4 images from the two embedding schemes, out of 1000 captures each image/embedding.
(145) TABLE-US-00003 TABLE 3 Test image Uniform strength embedding Equal visibility embedding Granola 18% 47% Apple Tart 50% 58% Giraffe Stack 47% 49% Pizza Puff 63% 61%
Conclusions
(146) A full color visibility model has been developed which has good correlation to subjective visibility tests for color patches degraded with a watermark. The best correlation was achieved with a model that applied a luminance correction to the CSF.
(147) The model was applied during the watermark embed process, using a pyramid based method, to obtain equal visibility. Better robustness and visibility was obtained with equal visibility embed than uniform strength embed.
(148) Discussion
(149) One goal of a color visibility model is to create an objective visual degradation model due to digital watermarking of an image. For example, a model may predict how noticeable or visible image changes will be due to watermark insertion. Highly noticeable changes can be reduced or modified to reduce watermark visibility, and/or to create equal watermark visibility (or lack thereof) across an image. For example, an error metric above or relative to the standard Just Noticeable Difference (JND) can be used to determine noticeable changes.
(150) In a first implementation, with reference to
(151) Contrast between the original image and the marked image can be determined, and then contrast sensitivity functions (CSFs) can be applied to each of the L*, a* and b* channels. For example, the L* CSFs discussed in Daly, Visible differences predictor: an algorithm for the assessment of image fidelity, F. L. van Nes et al. Spatial Modulation Transfer in the Human Eye, J. Opt. Soc. Am., Vol. 57, Issue 3, pp. 401-406 (1967), or Johnson et al, On Contrast Sensitivity in an Image Difference Model, PICS 2002: Image Processing, Image Quality, Image Capture Systems Conference, Portland, Oreg., April 2002; p. 18-23 (which is herein incorporated herein in its entirety), can be used. In other cases a bandpass filter, with a drop off toward low-frequencies, can be applied to the L*. The processed or blurred L* channel (from the original image) can be used to determine visibility masking. For example, areas of high contrast, edges, features, high variance areas, can be identified for inclusion of more or less watermarking strength. Some areas (e.g., flat area, edges, etc.) can be entirely masked out to avoid watermarking all together.
(152) For the a* and b* channels, chrominance CSFs can be applied to the respective channels, e.g., such CSFs as discussed in Johnson et al, Darwinism of Color Image Difference Models; G. J. C. van der Horst et al., Spatiotemporal chromaticity discrimination, J. Opt. Soc. Am., 59(11), 1482-1488, 1969; E. M. Granger et al., Visual chromaticity modulation transfer function, J. Opt. Soc. Am., 63(9), 73-74, 1973; K. T. Mullen, The contrast sensitivity of human colour vision to red-green and blue-yellow chromatic gratings, J. Physiol., 359, 381-400, 1985; each of which are hereby incorporated herein by reference in their entirety. In other cases, a low-pass filter is used which has a lower cut-off frequency relative to the CSF of luminance.
(153) Channel error difference can then be determined or calculated. For example, on a per pixel basis, L*, a* and b* data from the original image are compared to the blurred (e.g., processed with respective CSFs) L*, a* and b*channels from the watermarked image. One comparison utilizes E.sub.76: Using (L*.sub.1, a*.sub.1, b*.sub.1) and (L*.sub.2, a*.sub.2, b*.sub.2), two colors in L*a*b*, the error between two corresponding pixel values is:
(154) E*.sub.ab={square root over (L*.sub.2L*.sub.1).sup.2+(a*.sub.2a*.sub.1).sup.2+(b*.sub.2b*.sub.1).sup.2)}, where E*.sub.ab2.3 corresponds to a JND (just noticeable difference). Other comparisons may utilize, e.g., E.sub.94 or E.sub.2000.
(155) Of course, and more preferably used, is an error determination for the blurred (CSF processed) L*a*b* from the original image and the CSF blurred L*a*b* from the watermarked image.
(156) The output of the Calculate Channel Difference module identifies error metrics. The error metrics can be used to identify image areas likely to include high visibility due to the inserted digital watermark signal. We sometimes refer to this output as an error map. Typically, the lower the error, the less visible the watermark is at a particular area, image blocks or even down to a signal pixel.
(157) The visibility mask and the error map can be cooperatively utilized to guide digital watermarking. For example, watermark signal gain can be varied locally according to the error map, and areas not conducive to receive digital watermark, as identified in the visibility mask, can altogether be avoided or receive a further signal reduction.
(158) One limitation of the
(159) The luminance content of the original image provides potential masking of changes due to watermarking in chrominance as well as luminance. For example, where a watermark signal comprises mostly high frequency components, the masking potential of the original image is greater at regions with high frequency content. We observe that most high frequency content in a typical host image is in the luminance channel. Thus, the luminance content of the host is the dominant contributor to masking potential for luminance changes and chrominance changes for high frequency components of the watermark signal.
(160) Returning to
(161) With reference to
(162) Some visibility advantages of EVE vs. uniform strength embedding (USE) are shown in
CONCLUDING REMARKS
(163) Having described and illustrated the principles of the technology with reference to specific implementations, it will be recognized that the technology can be implemented in many other, different, forms. To provide a comprehensive disclosure without unduly lengthening the specification, applicant hereby incorporates by reference each of the above referenced patent documents in its entirety.
(164) The methods, processes, components, apparatus and systems described above may be implemented in hardware, software or a combination of hardware and software. For example, the watermark encoding processes and embedders may be implemented in software, firmware, hardware, combinations of software, firmware and hardware, a programmable computer, electronic processing circuitry, with a processor, parallel processors or other multi-processor configurations, and/or by executing software or instructions with one or more processors or dedicated circuitry. Similarly, watermark data decoding or decoders may be implemented in software, firmware, hardware, combinations of software, firmware and hardware, a programmable computer, electronic processing circuitry, and/or by executing software or instructions with a processor, parallel processors or other multi-processor configurations.
(165) The methods and processes described above (e.g., watermark embedders and detectors) also may be implemented in software programs (e.g., written in C, C++, Visual Basic, Java, Python, Tcl, Perl, Scheme, Ruby, executable binary files, etc.) stored in memory (e.g., a computer readable medium, such as an electronic, optical or magnetic storage device) and executed by a processor (or electronic processing circuitry, hardware, digital circuit, etc.).
(166) While one embodiment discusses inverting the polarity in a second color channel (e.g., a b channel), one could also invert the polarity in the first color channel (e.g., an a channel) instead. In such a case, the first color channel is then preferably subtracted from the second color channel.
(167) The particular combinations of elements and features in the above-detailed embodiments (including Appendix D) are exemplary only; the interchanging and substitution of these teachings with other teachings in this and the incorporated-by-reference patent documents are also contemplated.