Method and electronic device for producing composite image
10523856 ยท 2019-12-31
Assignee
Inventors
- Alok Shankarlal Shukla (Bangalore, IN)
- Prajit Sivasankaran NAIR (Bangalore, IN)
- Lokesh Rayasandra BOREGOWDA (Bangalore, IN)
Cpc classification
H04N23/81
ELECTRICITY
H04N23/11
ELECTRICITY
H04N9/646
ELECTRICITY
H04N23/45
ELECTRICITY
International classification
Abstract
A method and an electronic device for producing a composite image are provided. The method includes receiving visible image data and near infrared (NIR) image data from a composite sensor, determining whether at least one portion of the NIR image data having a level of detail greater than or equal to a threshold, and generating a composite image by fusing the visible image data with the at least one portion of the NIR image data based on the determination and storing the composite image in a memory.
Claims
1. A method of producing a composite image by an electronic device, the method comprising: receiving, by the electronic device, visible image data and near infrared (NIR) image data from a composite sensor; determining, by the electronic device, whether at least one portion of the NIR image data has a level of detail greater than or equal to a threshold, the level of detail comprising at least one of high frequency components of the NIR image data or brightness of the at least one portion of the NIR image data; generating, by the electronic device, the composite image with the visible image data based on the determining; and storing, by the electronic device, the composite image in a memory.
2. The method of claim 1, wherein the generating of the composite image further comprises generating the composite image by fusing the visible image data with the at least one portion of the NIR image data based on the determining of the at least one portion of the NIR image data having the level of detail greater than or equal to the threshold.
3. The method of claim 1, wherein the level of detail further comprises at least one of: noise reduction of the NIR image data, contrast adjustment of the visible image data, normalization of the visible image data, normalization of the NIR image data, high texture components of the NIR image, and decomposition of the visible image, or decomposition of the NIR image.
4. The method of claim 2, wherein the fusing of the visible image data with the at least one portion of the NIR image data comprises: detecting a contextual event, identifying the at least one portion of the NIR image data based on the contextual event, and transferring a detail layer of the at least one portion of the NIR image to a base layer of the visible image.
5. The method of claim 4, wherein the contextual event is one of a lighting condition or a weather condition.
6. The method of claim 4, wherein the transferring of the detail layer comprises transferring, to the base layer of the visible image, the level of detail of the at least one portion of the NIR image data based on a weight mask obtained using cues from the visible image data and the NIR image data.
7. The method of claim 6, wherein the transferring of the level of detail comprises transferring one of contrast detail or texture detail of the at least one portion of the NIR image data.
8. The method of claim 6, wherein the fusing of the visible image data with the at least one portion of the NIR image data comprises: selecting an amount of the at least one portion of the NIR image data to fuse with the visible image, based on the level of detail of the at least one portion of the NIR image data, and fusing the visible image data with the at least one portion of the NIR image data, based on the selected amount.
9. The method of claim 1, wherein the composite image comprises at least one of: an improved level of realistic skin smoothing compared to the visible image, an improved level of range compared to the visible image, an improved level of visibility compared to the visible image, or an improved level of rich information outside a visible wavelength compared to the visible image.
10. An electronic device for producing a composite image, the electronic device comprising: a composite sensor configured to capture a visible image and near infrared (NIR) image; and at least one processor configured to: receive visible image data associated with the visible image and NIR image data associated with the NIR image from the composite sensor, determine whether at least one portion of the NIR image data has a level of detail greater than or equal to a threshold, the level of detail comprising at least one of high frequency components of the NIR image data or brightness of the at least one portion of the NIR image data, generate the composite image with the visible image data based on the determining, and store the composite image in a memory coupled to the at least one processor.
11. The electronic device of claim 10, wherein the level of detail further comprises at least one of: noise reduction of the NIR image data, contrast adjustment of the visible image data, normalization of the visible image data, normalization of the NIR image data, high texture components of the NIR image, decomposition of the visible image, or decomposition of the NIR image.
12. The electronic device of claim 10, wherein in the generating of the composite image with the visible image, the at least one processor is configured to generate the composite image by fusing the visible image data with the at least one portion of the NIR image data based on the determining of the at least one portion of the NIR image data having the level of detail greater than or equal to the threshold.
13. The electronic device of claim 12, wherein in fusing the visible image data with the at least one portion of the NIR image data, the at least one processor is configured to: detect a contextual event, identify the at least one portion of the NIR image data based on the contextual event, and transfer a detail layer of the at least one portion of the NIR image to a base layer of the visible image.
14. The electronic device of claim 13, wherein the contextual event is one of a lighting condition or a weather condition.
15. The electronic device of claim 13, wherein in transferring the detail layer of the at least one portion of the NIR image data, the at least one processor is configured to: transfer, to the base layer of the visible image, the level of the detail of the at least one portion of the NIR image data based on a weight mask obtained using cues from the visible image data and the NIR image data.
16. The electronic device of claim 15, wherein in transferring the level of detail, the at least one processor is configured to transfer one of contrast detail or texture detail of the at least one portion of the NIR image data.
17. The electronic device of claim 15, wherein in fusing of the visible image data with the at least one portion of the NIR image data, the at least one processor is configured to: select an amount of the at least one portion of the NIR image data to fuse with the visible image, based on the level of detail of the at least one portion of the NIR image data, and fuse the visible image data with the at least one portion of the NIR image data, based on the selected amount.
18. The electronic device of claim 10, wherein the composite image comprises at least one of: an improved level of realistic skin smoothing compared to the visible image, an improved level of range compared to the visible image, an improved level of visibility compared to the visible image, or an improved level of rich information outside a visible wavelength when compared to the visible image.
19. The electronic device of claim 10, the at least one processor is further configured to improve an intensity associated with at least one of visible image of the visible image data or NIR image of the NIR image data.
20. A non-transitory computer readable recording medium having recorded thereon a program for executing a method of producing a composite image by an electronic device, the method comprising: receiving visible image data and near infrared (NIR) image data from a composite sensor; determining whether at least one portion of the NIR image data having a level of detail greater than or equal to a threshold, the level of detail comprising at least one of high frequency components of the NIR image data or brightness of the at least one portion of the NIR image data; generating the composite image with the visible image data based on the determining; and storing, by the electronic device, the composite image in a memory.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The above and other aspects, features, and advantage of certain embodiments of the present disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13) Throughout the drawings, like reference numerals will be understood to refer to like parts, components, and structures.
DETAILED DESCRIPTION
(14) The following description with reference to the accompanying drawings is provided to assist in a comprehensive understanding of various embodiments of the present disclosure as defined by the claims and their equivalents. It includes various specific details to assist in that understanding but these are to be regarded as merely exemplary. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the various embodiments described herein can be made without departing from the scope and spirit of the present disclosure. In addition, descriptions of well-known functions and constructions may be omitted for clarity and conciseness.
(15) The terms and words used in the following description and claims are not limited to the bibliographical meanings, but, are merely used by the inventor to enable a clear and consistent understanding of the present disclosure. Accordingly, it should be apparent to those skilled in the art that the following description of various embodiments of the present disclosure is provided for illustration purpose only and not for the purpose of limiting the present disclosure as defined by the appended claims and their equivalents.
(16) It is to be understood that the singular forms a, an, and the include plural referents unless the context clearly dictates otherwise. Thus, for example, reference to a component surface includes reference to one or more of such surfaces.
(17) Also, the various embodiments described herein are not necessarily mutually exclusive, as some embodiments can be combined with one or more other embodiments to form new embodiments.
(18) Herein, the term or as used herein, refers to a non-exclusive or, unless otherwise indicated. The examples used herein are intended merely to facilitate an understanding of ways in which the embodiments herein can be practiced and to further enable those skilled in the art to practice the embodiments herein. Accordingly, the examples should not be construed as limiting the scope of the embodiments herein.
(19) In accordance with an embodiment of the present disclosure, a method is provided for producing a composite image by an electronic device. The method includes receiving visible image data and near infrared (NIR) image data from a composite sensor, identifying at least one portion of the NIR image data having a level of detail greater than or equal to a threshold, generating a composite image by fusing the visible image data with the at least one portion of the NIR image data, and storing the composite image in a memory.
(20) In accordance with an embodiment of the present disclosure, an electronic device is configured to produce a composite image. The electronic device includes: a memory, a composite sensor capable of capturing a visible image and NIR image, a processor, coupled to the memory and the composite sensor, configured to: receive visible image data associated with the visible image and NIR image data associated with the NIR image from the composite sensor, identify and/or determine whether at least one portion of the NIR image data having a level of detail greater than or equal to a threshold, generate the composite image by fusing the visible image data with the at least one portion of the NIR image data if the at least one portion of the NIR image data has a level of detail greater than or equal to a threshold and store the composite image in the memory.
(21) In a mechanism of the related art, an infrared (IR) filter mounted between the lens and the NIR sensor(s) of an electronic device (i.e., mobile phone, camera) minimizes interference with the NIR sensor(s). However, NIR waves, which have a wavelength of 700-1100 nm, can travel through haze & fog and better penetrate human skin, thus serving to automatically smoothen skin tone of a subject in a captured image. The visible spectrum wavelength can facilitate the user with images which contain desired color information and are visible by the user. However, under low light conditions, visible images obtained through visible image sensor imaging (i.e., visible light-wavelengths of 400-700 nm) may be distorted, and is may be difficult to obtain imperative details in the images.
(22) Further, the international commission on illumination (CIE) standard utilizes a D65 light source (standard daylight), which does not contain NIR content. For NIR images captured with a composite red, green, and blue (RGB)-IR image sensor under D65 light, no extra information is present, only noise. A light source, on the other hand, contains good amount of NIR content. The fusion of visible (e.g., RGB) and NIR images will improve the quality of a composite RGB-IR image in the case of a light source. However, with D65 light it will have little impact on or may degrade the quality of the NIR image due to sensor noise. Generating composite images by blindly fusing visible (e.g., RGB) and NIR images may result in degraded quality. Thus, combining or not combining visible (e.g., RGB) and NIR images depends on the NIR content in the light source (i.e., NIR image quality).
(23) Unlike the mechanism of the related art, according to the proposed mechanism, a user may differentiate details from noise in NIR images obtained from a composite RGB-IR image sensor which can be called as a composite sensor alternately in this disclosure.
(24) Unlike the mechanism of the related art, the proposed mechanisms increase the contrast of visible images and normalization of NIR images, as the obtained NIR images (i.e., images obtained under low light conditions) lack significant contrast therein.
(25) Unlike the mechanism of the related art, the proposed mechanisms perform seamless fusion of visual and NIR images.
(26) Referring now to the drawings, and more particularly to
(27)
(28) Referring to
(29) The electronic device 100 includes a composite sensor 102, a processor 104 (for example, a hardware unit, an apparatus, a central processing unit (CPU), or a graphics processing unit (GPU)), memory 106 (e.g., a volatile memory and/or a non-volatile memory), a display 108 (e.g., a cathode ray tube (CRT) display, a liquid crystal display (LCD), a light-emitting diode (LED)), and a communicator 110.
(30) The memory 106 includes storage locations accessible via the processor 104. The electronic device 100 may interface with the display 108 via the processor 104, and the display 108 may be integrated with the electronic device 100 or communicatively coupled thereto. The processor 104 may be communicatively coupled to the memory 106. Further, the processor 104 is configured to generate an enhanced image (e.g., a visible image and an NIR image captured by the composite sensor 102) and display the same on the display 108.
(31) The composite sensor 102 may include a RGB sensor for capturing visible images (e.g., RGB images) and an NIR sensor for capturing NIR images.
(32) The processor 104 may separate visible images (e.g., RGB images) and NIR images received from the composite sensor 102. The processor 104 may determine whether any portions of an NIR image (e.g., any portions of data of the NIR image) have a level of detail greater than or equal to a threshold. For example, the processor 104 may identify a portion(s) of the NIR image having a level of detail greater than or equal to the threshold. The detail may be, for example, high frequency components of the NIR image, high texture components of the NIR image, or contrast of the NIR image obtained by a contrast adjuster (to be described in greater detail below). The threshold may be configured by the processor 104 but is not limited thereto. For example, the threshold may be preset or adjustable by the processor 104, a user, or a manufacturer. The processor 104 may generate a composite image (e.g., visible image+NIR image) by fusing visible image data (e.g., RGB image data) with the identified portion(s) of the NIR image data, thereby producing an enhanced image if the portion(s) of the NIR image have a level of detail greater than or equal to the threshold.
(33) The composite image may be rendered on the display 108 and stored on the memory 106.
(34) Unlike the mechanism of the related art, according to the proposed method, the processor 104 may analyze the availability of additional visual details in the NIR image data and intelligently perform fusion of the visible image data with one or more portions of the NIR image data having the additional visual details, thereby obtaining an enhanced image (i.e., composite image).
(35) The electronic device 100 of the present disclosure is not limited to the units (i.e., components, devices, or elements) illustrated in
(36)
(37) Referring to
(38) If the processor 104 identifies a portion(s) of the NIR image having a level of detail greater than or equal to the threshold in operation 204, then, in operation 206, the electronic device 100 may generate the composite image by fusing the visible image data with the identified portion(s) of the NIR image data. For example, the processor 104 of the electronic device 100 illustrated in
(39) Once the processor 104 generates the composite image, the electronic device 100 may store the composite image in the memory 106 in operation 210. For example, the processor 104 of the electronic device 100 illustrated in
(40) If the processor 104 does not identify a portion(s) of the NIR image having a level of detail greater than or equal to the threshold in operation 204, then, in operation 208, the electronic device 100 may output a visible image.
(41) The various actions, acts, blocks, operations, etc., as illustrated in
(42)
(43) Referring to
(44) The processor 104 may transfer the details of the visible image 302 and the NIR image 304 to the color temperature detector 306. The color temperature detector 306 may transfer the visible image 302 and the NIR image 304, along with respective details thereof, to a noise suppressor 308.
(45) The Color Temperature Detector 306:
(46) The color temperature of a light source is the temperature of an ideal black-body radiator capable of emitting light comparable in hue to the light source. In practice, color temperature is only meaningful for light sources that may correspond, closely, to the light emitted by the black-body radiator. Low color temperature light refers to warmer light (e.g., a larger amount of yellow/red light), while high color temperature light refers to colder light (e.g., a larger amount of blue light). Daylight has a relatively low color temperature near dawn, and a higher one during the day. The color temperature detector 306 translates RGB values into tristimulus values (e.g., mixture of the three CIE primaries, X, Y, Z) and may obtain further correlated color temperature (CCT) using standard CCT formulae.
(47) In the Gaussian approach of the related art using a bilateral filter for noise reduction in low light scenarios, preserving details of the visible image 302 and the NIR image 304 may result in a loss of details (e.g., noise information), which can affect image enhancement for low light conditions. Unlike the Gaussian approach of the related art using a bilateral filter, the noise suppressor 308 of the proposed mechanism provides for denoising (e.g., reducing noise) of the visible image 302 and the NIR image 304, which will be described in greater detail below.
(48) Noise Suppressor 308:
(49) The sharpness of the visible image 302 and the NIR image 304 is enhanced by modifying the range function of the bilateral filter according to the related art. The bilateral filter of the related art uses a Gaussian function in which the domain function is realized as:
(50)
(51) and the range function is realized as:
(52)
(53) where i, j represents the spatial location of the image, x denotes the difference of the pixel values, .sub.D.sup.2 denotes the variance of the domain function, and .sub.R.sup.2 denotes the variance of the range function. The range function is modified by altering the variance of the Gaussian function as follows:
(54)
(55) where x represents the difference of the pixel values and represents the amount of smoothening of the image.
(56) The noise suppressor 308 may reduce noise from the visible image 302 (i.e., visual image) and output a corresponding denoised visible image. Further, the noise suppressor 308 may reduce noise from the NIR image 304 and output a corresponding denoised NIR image. For purposes of the description corresponding to
(57) However, the embodiment is not limited thereto. The noise suppressor 308 may selectively apply noise filtering to the visible image 302 and the NIR image 304 and/or vary the amount of noise filtering applied thereto. For example, the noise suppressor 308 may directly output the visible image 302 and/or the NIR image 304 to the contrast adjuster 310 without applying noise filtering thereto (i.e., output the visible image 302, as opposed to outputting a corresponding denoised visible image, and output the NIR image 304, as opposed to outputting a corresponding denoised NIR image).
(58) Contrast Adjuster 310:
(59) The contrast adjuster 310 may enhance the denoised visible image to improve visibility thereof.
(60) For example, visible images captured under low light conditions may suffer from deteriorated visual quality in overexposed and underexposed regions. The contrast adjuster 310 may increase the visual quality of such images (e.g., visible images, denoised visible images), thereby enhancing details (e.g., high frequency components, high texture components, contrast, amount of noise, etc.) thereof in both overexposed and underexposed regions while normalizing the same. The contrast adjuster 310 may perform any of the following (A-C) to improve the visibility (e.g., visual properties) of denoised visible images:
(61) A) Gamma Correction:
(62) Gamma correction is performed on the denoised visible image. However, gamma correction results in losing the natural appearance of the image and creates some unnatural artifacts.
(63) B) Histogram Matching:
(64) The probability distribution of the visible image 302 is estimated and the probability distribution of the NIR image 304 is adjusted to that of the visible image 302 by replacing pixel values in the corresponding bins.
(65) C) Brightness Adjustment Using Black and White Points:
(66) In this approach, the intensity of the NIR image 304 is adjusted to that of the luminance channel of the visible image 302 such that detail layers of the visible image 302 and the NIR image 304 are comparable. For example, the processor 104 may shift the mean of the luminance of the NIR image 304 towards the mean of the luminance of the visible image 302 and then normalize the shifted mean of the luminance to the range of visible image 302. Such an approach shows promising results as it deals with the normalization of the intensity values of the visible image 302 and the NIR image 304.
(67) IR Normalizer 312:
(68) The IR normalizer 312 may enhance denoised visible images output from the noise suppressor 308. Since visible images captured under low light conditions may suffer from low intensity levels, the IR normalizer 312 may enhance the same for better visual perception. For example, the IR normalizer 312 may obtain a denoised visible image and enhance the maximum value (I) of R, G, B channels using the following formulae:
(69)
(70) where I represents the input image, r, c represents the spatial location, .sub.R represents the variance of the range function, and D.sub.s represents the enhanced image. R, G, B channels of the denoised visible image are enhanced as below:
R(r,c)=R(r,c)*D.sub.s(r,c)/I(r,c)Equation 5
G(r,c)=G(r,c)*D.sub.s(r,c)/I(r,c)Equation 6
B(r,c)=B(r,c)*D.sub.s(r,c)/I(r,c)Equation 7
(71) The IR normalizer 312 may match intensity levels of denoised NIR images to those of the denoised visible image having enhanced R, G, B channels (e.g., enhanced RGB image), thereby preventing mismatch in intensity levels from occurring during fusion. For example, the IR normalizer 312 may shift the mean of the histogram of a denoised NIR image to match that of the denoised RGB image and normalize the range of the pixels of the denoised NIR image to that of the denoised visible image.
(72) Decomposition Device 314:
(73) Most details of NIR images captured under low light scenarios will be preserved. Hence the base layer of visible image 302 and the detail layer of NIR image 304 are obtained from the bilateral filter of the related art as follows:
Vis_base=bfVis)Equation 8
NIR.sub.base=bf(Nir)Equation 9
NIR_detail=NIRNIR_baseEquation 10
(74) where bf represents the bilateral filter of the related art.
(75) The final enhanced image is obtained using the linear combination of the base layer of the visible image 302 and the detail layer of the NIR image 304 as follows:
I_en=Vis_base+*NIR_detailEquation 11
(76) where controls the amount of contribution from the NIR detail layer, which plays a role in obtaining the desired output (e.g., final enhanced image) without any artifacts.
(77) The transfer of the detail layer of a portion of the NIR image 304 data to the base layer of the visible image 302 includes transferring the level of detail of the portion of the NIR image 304 data (e.g., transferring the NIR detail of one or more portions of data of the NIR image 304) to a color correction device 316 and a weighted mask device 318 obtained from the saturation and luminance channels of the visible image 302.
(78) Color Correction Device 316:
(79) The color correction device 316 may utilize a color correction matrix for color correction operations such as adjusting white balance, color cast, brightness, or contrast in the visible image 302 (e.g., an RGB image). The standard color correction matrix computation methods can be used for improving color reproduction of fused RGB-IR images (e.g., composite image).
(80) Weighted Mask Device 318:
(81) A weight mask is created using cues obtainable from the visible image 302 and the NIR image 304. The cues include, for example, intensity of visible image 302, saturation of visible image 302, detail level in the visible image 302, and detail level in the NIR image 304. Further, the weighted mask device 318 may selectively fuse portions of the visible image 302 and the NIR image 304 to generate an enhanced visible image (e.g., to enhance the visible image 302). In another embodiment, the weight mask is created based on an analysis of contrast and texture (edges) of the visible image 302 and the NIR image 304, and the weighted mask device 318 may fuse the visible image 302 and the NIR image 304 based on this weight mask (e.g., based on an analysis and comparison of information content of the NIR image 304 and information content of the visible image 302).
(82)
(83) Referring to
(84) In such an example, the weighted mask device 318 may apply a weight mask having a high weight value in the center. The detail level of the NIR image 304 is the high frequency component (edges or texture) present in the NIR image 304, as shown in
(85) Texture Transfer Device 320:
(86) The texture transfer device 320 may selectively transfer details from the NIR image 304 to the visible image 302 according to the weighted mask. For example, the texture transfer device 320 may fuse the detail layer of NIR image 304 with the base layer of the visible image 302 based on the weighted mask.
(87) Contrast Transfer Device 322:
(88) Intensity mismatch between images may lead to the occurrence of unnatural artifacts during fusion. Hence, to reduce intensity mismatch and the occurrence of such unnatural artifacts, the contrast transfer device 322 may apply histogram matching techniques to match the intensity level of the NIR image 304 to that of the visible image 302. For example, the contrast transfer device 322 may apply histogram matching techniques to match the normalized visible image (e.g., the denoised visible image following normalization by the IR normalizer 312) and the normalized NIR image (e.g., the denoised NIR image following normalization by the IR normalizer 312).
(89) Display 324:
(90) The display 324 (similar to the display 108) may display the enhanced image (i.e., composite image) to the user.
(91) Unlike the mechanism of the related art, the processor 104 can be configured to intelligently perform fusion of visible image data with one or more portions of the NIR image data to obtain an enhanced output image based on the presence of an IR light source, which will be disclosed described in greater detail with respect to
(92) In conjunction with
(93)
(94) Referring to
(95)
(96) The processor 104 may detect for contextual events in the visible image 502 and the NIR image 504. For example, the processor 104 may detect for a weather event such as hazy weather conditions. Further, upon detecting a contextual event (e.g., weather event), the processor 104 may identify a portion(s) of the NIR image 504 having a level of detail greater than or equal to a threshold and generate a composite image 506 (e.g., visible image 502 data+NIR image 504 data) having a wavelength outside of visible wavelength, by fusing the visible image 502 data with the identified portion(s) of the NIR image 504 data.
(97)
(98) The processor 104 may identify a portion(s) of an NIR image 604 having a higher level of detail (e.g., contrast) than a visible image 602. The processor 104 can be configured to generate a composite image 606 (e.g., visible image 602 data+NIR image 604 data) having improved range, by fusing the visible image 602 data with the identified portion(s) of the NIR image 604 data.
(99)
(100) Referring to
(101) The processor 104 may transfer the details of the visible image 302 and the NIR image 304 to the noise suppressor 308. The operations of the noise suppressor 308 are described above in reference to
(102) Luminance Channel Extractor 326:
(103) The luminance channel extractor 326 can be configured to extract intensity (e.g., calculate brightness) from the NIR image 304 (e.g., from a portion of the NIR image 304) and adjust the intensity to that of the luminance channel of the visible image 302 such that detail layers of the visible image 302 and the NIR image 304 are comparable with respect to intensity. For example, the processor 104 may shift the mean of the luminance of the NIR image 304 towards the mean of the luminance of the visible image 302 and then renormalize the shifted mean of the luminance to the range of visible image 302.
(104) The operations of the IR Normalizer 312 and the display 324 are described above in reference to
(105) Detail Layer Extractor 328:
(106) The detail layer extractor 328 includes the texture transfer device 320 and the contrast transfer device 322. The operations of each of the texture transfer device 320 and the contrast transfer device 322 are described above in reference to
(107) Base Layer Extractor 330:
(108) the base layer extractor 330 can be configured to decompose a portion(s) of the NIR image 304 and the visible image 302 to a base layer (e.g., low frequency). The base layer extractor 330 may include the decomposition device 314 (as described above in reference to
(109) Fusion Device 332:
(110) The fusion device 332 can be configured to fuse portions of the visible image 302 and the NIR image 304.
(111) Chrominance and Color Converter 334:
(112) The chrominance and color converter 334 can be configured to adjust white balance, color cast, brightness, or contrast in the visible image 302 (e.g., an RGB image). The standard color correction matrix computation methods can be used for improving color reproduction of fused RGB-IR images (e.g., composite image).
(113)
(114) Referring to
(115) For example, in operation 802, the processor 104 of the electronic device 100 illustrated in
(116) In operation 804, the IR application of the electronic device 100 detects whether the IR light source is in ON state. For example, in operation 804, the processor 104 of the electronic device 100 illustrated in
(117) In operation 806, if the electronic device 100 detects the presence of the IR light source (i.e., detects that the IR light source is ON), then in operation 808, the electronic device 100 captures a first frame type (e.g., an IR image frame).
(118) In operation 810, if the electronic device 100 detects that the IR light source is in an OFF state, then in operation 812, the electronic device 100 captures a second frame type (e.g., a visible image frame or RGB image frame).
(119) In operation 814, the electronic device 100 detects whether a portion(s) of the IR image (e.g., portion(s) of data in the IR image) has a level of detail greater than or equal to a threshold. The value of the threshold may be predetermined. When the electronic device 100 detects that a portion(s) of the IR image (e.g., a portion(s) of data of the IR image) has a level of detail greater than or equal to a threshold, the electronic device 100 performs the fusion of the IR image and the visible image. When the electronic device 100 detects that no portion of the IR image (e.g., no portion of data in the IR image) has a level of detail greater than or equal to the threshold, the electronic device 100 may output the visible image frame, without fusing the IR image frame and the visible image frame.
(120) In operation 816, the electronic device 100 terminates the fusion process.
(121) The various actions, acts, blocks, operations, etc., as illustrated in
(122)
(123) Under the presence of an IR light source, it is useful to combine visible (e.g., RGB) and NIR images. Combining improves quality of images captured in portrait mode (e.g., skin smoothing of portrait photos) and a larger amount of detail is captured (e.g., higher visibility) under low light conditions.
(124) The two main areas of interest for imaging are the epidermis and the dermis. The epidermis, also called melanin layer, is responsible for most of the pigmented coloration of the skin. It is the melanin concentration that gives the skin a color that can vary from pale to dark brown. The epidermis layer is otherwise fairly thin and no significant scattering occurs. Just underneath the epidermis lies the dermis layer. In the dermis layer, significant scattering occurs, as well as hemoglobin absorption (giving the skin its reddish shade). The building block for skin smoothing is the same as for low light enhancement. Referring back to
(125)
(126)
(127) Referring to
(128) The overall computing environment 1000 can be composed of multiple homogeneous and/or heterogeneous cores, multiple CPUs of different kinds, special media and other accelerators. The processor 1006 is responsible for processing the instructions of the technique. Further, the processor 1006 may be implemented as multiple processors located on a single chip or multiple chips.
(129) The technique comprising of instructions and codes required for the implementation are stored in either the memory 1008, the storage 1010, or both. At the time of execution, the instructions may be fetched from the memory 1008 or storage 1010, and executed by the processor 1006.
(130) In the case of any hardware implementations, various networking devices 1014 or external I/O devices 1012 may be connected to the computing environment to support the implementation.
(131) The various embodiments disclosed herein can be implemented through at least one software program running on at least one hardware device and performing network management functions to control the elements. The elements shown in
(132) While the present disclosure has been shown and described with reference to the various embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the present disclosure as defined by the appended claims and their equivalents.