Color signal and near-infrared signal generated by using pattern information defining intensity-corresponding pattern
10440292 ยท 2019-10-08
Assignee
Inventors
Cpc classification
H04N23/55
ELECTRICITY
H04N23/16
ELECTRICITY
H04N23/10
ELECTRICITY
H04N23/11
ELECTRICITY
H04N23/45
ELECTRICITY
G02B5/208
PHYSICS
International classification
Abstract
The invention enables image processing of visible light and near-infrared light using an imaging device. An acquisition unit (110) acquires an image signal representing an image including near-infrared light that has an intensity according to a pattern having a prescribed geometric shape. A signal processing unit (120) uses pattern information which defines the pattern to output, a color signal representing visible light components corresponding to the image signal and a near-infrared signal representing near-infrared light components corresponding to the image signal.
Claims
1. An image processing device comprising: an acquirer implemented at least by hardware configured to acquire a video signal representing a video including near-infrared light having an intensity corresponding to a pattern having a prescribed geometric shape; and a signal processer implemented at least by the hardware configured to output a color signal and a near-infrared signal by using pattern information for defining the pattern, the color signal representing a visible light component corresponding to the acquired video signal, the near-infrared signal representing a near-infrared light component corresponding to the video signal, wherein when a difference between the acquired video signal and an estimated value of the video signal calculated by using an estimated value of the color signal, an estimated value of an intensity ratio between the color signal and the near-infrared signal, and the pattern information satisfies a first condition, and a spatial variation in an estimated value of the color signal and an estimated value of the intensity ratio satisfies a second condition, the signal processer outputs, as an output signal, an estimated value of the color signal and a value of the near-infrared signal obtained from an estimated value of the color signal and an estimated value of the intensity ratio.
2. The image processing device according to claim 1, wherein the pattern information indicates a position and a shape of the pattern.
3. The image processing device according to claim 1, further comprising a light receiver implemented at least by the hardware and including an optical filter including a transmission configured to transmit near-infrared light with the pattern.
4. The image processing device according to claim 3, wherein the light receiver includes a plurality of sensors each corresponding to one of a plurality of color components, and the pattern has a correspondence relation with an array of the plurality of sensors.
5. An image processing method comprising: acquiring a video signal representing a video including near-infrared light having an intensity corresponding to a pattern having a prescribed geometric shape; outputting, by using pattern information for defining the pattern, a color signal representing a visible light component corresponding to the acquired video signal and a near-infrared signal representing a near-infrared light component corresponding to the video signal, and outputting, as an output signal, an estimated value of the color signal and a value of the near-infrared signal obtained from an estimated value of the color signal and an estimated value of the intensity ratio, when a difference between the acquired video signal and an estimated value of the video signal calculated by using an estimated value of the color signal, an estimated value of an intensity ratio between the color signal and the near-infrared signal, and the pattern information satisfies a first condition, and a spatial variation in an estimated value of the color signal and an estimated value of the intensity ratio satisfies a second condition.
6. A non-transitory computer-readable medium program recording medium recording a program causing a computer to execute: processing of acquiring a video signal representing a video including a near-infrared light having an intensity corresponding to a pattern having a prescribed geometric shape; and processing of outputting, by using pattern information for defining the pattern, a color signal representing a visible light component corresponding to the acquired video signal and a near-infrared signal representing a near-infrared light component corresponding to the video signal, and processing of outputting, as an output signal, an estimated value of the color signal and a value of the near-infrared signal obtained from an estimated value of the color signal and an estimated value of the intensity ratio, when a difference between the acquired video signal and an estimated value of the video signal calculated by using an estimated value of the color signal, an estimated value of an intensity ratio between the color signal and the near-infrared signal, and the pattern information satisfies a first condition, and a spatial variation in an estimated value of the color signal and an estimated value of the intensity ratio satisfies a second condition.
Description
BRIEF DESCRIPTION OF DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
DESCRIPTION OF EMBODIMENTS
First Example Embodiment
(24)
(25) The term video described herein refers to an image to be captured through an optical system, such as a lens, and may be either a still image or a moving image. The color signal is a signal representing visible light components in the video signal. On the other hand, the near-infrared signal is a signal representing near-infrared light components in the video signal. The color signal and the near-infrared signal represent, for example, the brightness of a pixel, but are not limited only to the brightness. In the following description, it is assumed that the color signal and the near-infrared signal represent brightness of each pixel of a still image or an image of a video at a specific time of the video.
(26) In this example embodiment, the term visible range refers to a wavelength range of 400 to 700 nm. In the wavelength range, a blue (B) wavelength range is from 400 to 500 nm, a green (G) wavelength range is from 500 to 600 nm, and a red (R) wavelength range is from 600 to 700 nm. The term near-infrared range refers to a wavelength range of 700 nm to 2.5 m. However, classification of the wavelength ranges indicated herein is merely an example.
(27) The video processing device 100 includes an acquisition unit 110 and a signal processing unit 120. The video processing device 100 is connected to an external device that supplies a video signal. The external device is, for example, an imaging device including an image sensor. The acquisition unit 110 acquires a video signal from the external device. The signal processing unit 120 outputs a color signal and a near-infrared signal based on the video signal acquired by the acquisition unit 110.
(28) The acquisition unit 110 acquires a video signal representing a video including near-infrared light having an intensity corresponding to a pattern having a prescribed geometric shape. The term pattern described herein refers to a pattern in which figures, such as a circle or a rectangle, are repeatedly regularly arranged. Such a video signal is obtained in such a manner that, for example, the image sensor is provided with an optical filter (hereinafter referred to as an NIR cut tilter) that cuts near-infrared light.
(29)
(30) The pattern in which near-infrared light components included in the video signal are formed on the video does not necessarily match the pattern of the NIR cut filter. This is because the near-infrared light causes diffraction after the near-infrared light is transmitted through the NIR cut filter. Each figure in the pattern of the near-infrared light components appearing as a video in the video signal is larger than the pattern (corresponding to the infrared transmissive portion 12 illustrated in
(31) The signal processing unit 120 acquires such a video signal from the acquisition unit 110, and outputs a color signal and a near-infrared signal. The color signal is typically output as signals of three components of R, G, and B, but is not necessarily limited to this form. The signal processing unit 120 executes prescribed arithmetic processing by using pattern information for defining a pattern of the near-infrared light, thereby enabling the color signal and the near-infrared signal to be output.
(32) The pattern information is information for specifying a pattern of the near-infrared light on the video. For example, the pattern information is data representing a position and shape of the pattern in an NIR cut filter. In the case of the NIR cut filter illustrated in
(33) As described above, the video processing device 100 can separate the color signal and the near-infrared signal from the video signal by using a combination of the video signal representing the video including near-infrared light with a prescribed pattern and the pattern information. Therefore, according to the video processing device 100, it is possible to respectively execute visible light video processing and near-infrared light video processing based on the video signal including the color signal and the near-infrared signal.
(34) Further, when the video signal is obtained by an imaging device, it is sufficient to use a typical imaging device provided with the NIR cut filter as illustrated in
Second Example Embodiment
(35)
(36) The video data acquisition unit 210 acquires video data. The video data acquisition unit 210 can acquire video data from an external device similar to that of the first example embodiment. The video data includes at least a plurality of color signals. In this case, the plurality of color signals are color signals represented by separate color components of three colors of R, G, and B, and each pixel is represented by a value of a prescribed bit number. The term color signals described herein refers to a video in a state where near-infrared light components are superimposed on visible light components. Such color signals are hereinafter also referred to as a first color signal. The first color signal is a signal obtained by adding a second color signal and a near-infrared signal which are described below.
(37) The first color signal acquisition unit 220 acquires the first color signal from the video data acquisition unit 210. The first color signal acquisition unit 220 acquires the first color signal for each color.
(38) The pattern storage unit 230 stores pattern information. The pattern storage unit 230 is composed of, for example, a storage medium such as a hard disk or a flash memory. As the pattern information of this example embodiment, data similar to that of the first example embodiment can be used. As the pattern information, data common to each color can be used.
(39) The second color signal estimation unit 240 estimates the second color signal which is a color signal obtained by removing near-infrared light components from the first color signal. Further, the second color signal estimation unit 240 estimates not only the second color signal, but also the intensity ratio between the second color signal and the near-infrared signal. The second color signal estimation unit 240 estimates the second color signal for each color and the intensity ratio thereof based on the first color signal acquired by the first color signal acquisition unit 220 and the pattern information stored in the pattern storage unit 230.
(40) The near-infrared signal calculation unit 250 calculates a near-infrared signal for each color. The near-infrared signal calculation unit 250 can calculate a near-infrared signal by using the second color signal estimated by the second color signal estimation unit 240 and the intensity ratio between the second color signal and the near-infrared signal.
(41) The output unit 260 outputs a second color signal and a near-infrared signal. The output unit 260 executes a prescribed calculation (e.g., addition) on the near-infrared signal for each color calculated by the near-infrared signal calculation unit 250, and outputs calculation results.
(42) The first color signal acquisition unit 220, the second color signal estimation unit 240, and the near-infrared signal calculation unit 250 may sequentially execute processing for each color, or may simultaneously execute the processing in parallel.
(43)
(44) The initial value estimation unit 241 calculates initial values of estimated values of the second color signal and the intensity ratio between the second color signal and the near-infrared signal. The initial value estimation unit 241 calculates initial values of the estimated value of the second color signal and the estimated value of the intensity ratio based on the first color signal.
(45) The estimated value selection unit 242 selects the estimated values of the second color signal and the intensity ratio. The estimated value selection unit 242 repeatedly performs processing of selecting these estimated values. In the initial selection processing, the estimated value selection unit 242 selects the initial value calculated by the initial value estimation unit 241, while in the second and subsequent selection processing, the estimated value selection unit 242 selects an estimated value updated by the estimated value update unit 246.
(46) The smoothness evaluation unit 243 evaluates smoothness of the estimated values of the second color signal and the intensity ratio. In this case, the smoothness indicates a degree of spatial variation in values. For example, the phrase the estimated value of the second color signal is smooth indicates that the difference between a maximum value and a minimum value of an estimated value of each pixel in a certain range constituting a video is equal to or less than a prescribed threshold. The smoothness evaluation unit 243 calculates an evaluated value of the smoothness according to a prescribed algorithm.
(47) The first color signal estimation unit 244 estimates the first color signal. The first color signal estimation unit 244 calculates an estimated value of the first color signal based on the estimated value selected by the estimated value selection unit 242 and the pattern information stored in the pattern storage unit 230.
(48) The error calculation unit 245 compares the estimated value of the first color signal with the actual first color signal, and calculates an error therebetween. Specifically, the error calculation unit 245 compares the first color signal estimated by the first color signal estimation unit 244 with the first color signal acquired by the first color signal acquisition unit 220.
(49) The estimated value update unit 246 updates the estimated values of the second color signal and the intensity ratio. The estimated value update unit 246 updates the estimated values based on the estimated value calculated by the smoothness evaluation unit 243 and the error calculated by the error calculation unit 245.
(50) Further, the estimated value update unit 246 compares the estimated values before and after the update. When the amount of update of each estimated value is sufficiently small, the update is finished. Specifically, the estimated value update unit 246 compares the amount of update of each estimated value with a prescribed threshold, and when the amount of update is equal to or less than the threshold, the estimated value update unit 246 finishes the update. The estimated value update unit 246 sets the estimated value obtained at the time when the update is finished as an output value of the second color signal estimation unit 240.
(51) On the other hand, when the amount of update exceeds the threshold, the estimated value update unit 246 supplies the estimated value to the estimated value selection unit 242. In this case, the estimated value selection unit 242, the smoothness evaluation unit 243, the first color signal estimation unit 244, the error calculation unit 245, and the estimated value update unit 246 executes the above-described processing by using the updated estimated value, and repeatedly perform this processing until the update of the estimated value is finished.
(52) The video processing device 200 has a configuration as described above. In this configuration, when the video processing device 200 acquires video data, the video processing device 200 outputs the color signal and the near-infrared signal. Specific operations of the video processing device 200 will be described below. It is assumed herein that color signals for R, G, and B colors are set to all pixels of video data.
(53)
(54) First, the video data acquisition unit 210 acquires video data (step S1). Next, the first color signal acquisition unit 220 selects any one of a plurality of first color signals included in the video data acquired by the video data acquisition unit 210 (step S2). At this time, the first color signal acquisition unit 220 selects a first color signal on which the processing of steps S3 and S4, which are described below, has not been executed yet.
(55) When any one of the first color signals is selected by the first color signal acquisition unit 220, the second color signal estimation unit 240 estimates a second color signal and an intensity ratio between the second color signal and a near-infrared signal based on the selected first color signal (step S3). In other words, the second color signal estimation unit 240 calculates an estimated value of the second color signal and an estimated value of the intensity ratio. Next, the near-infrared signal calculation unit 250 calculates a near-infrared signal based on these estimated values (step S4).
(56) After obtaining the necessary second color signal and near-infrared signal, the output unit 260 outputs the second color signal and near-infrared signal. Specifically, the output unit 260 determines whether or not the processing of steps S2 to S4 has been executed for all colors (step S5). When the processing for all colors has been finished (step S5: YES), the second color signal and the near-infrared signal are output (step S6).
(57) On the other hand, when there is any color for which the processing of steps S2 to S4 has not been executed (step S5: NO), the first color signal acquisition unit 220 selects a first color signal that has not been processed (step S2). The second color signal estimation unit 240 and the near-infrared signal calculation unit 250 executes the processing of steps S3 and S4 again according to the selection in step S2.
(58) The processing of steps S3 and S4 will be described in more detail below. The following description is made by using a color of G for convenience of explanation, but processing for other colors is performed in a similar manner.
(59)
(60) The near-infrared signal has a value significant to a range wider than the actual area of the infrared transmissive portion due to an effect of diffraction of near-infrared light, and the value gradually decreases toward the outside from the center of the infrared transmissive portion. When a distance between adjacent infrared transmissive portions is short, the near-infrared signal may contain a mixture of components derived from a certain infrared transmissive portion and components derived from another infrared transmissive portion.
(61)
(62) In this case, an intensity I(w) on the image sensor when the near-infrared light having a wavelength and an incident intensity I.sub.0 is incident into one infrared transmissive portion located on the NIR cut filter is represented by the following Formula (1).
(63)
where J.sub.1(w) represents the Bessel function of the first kind of order 1, and C represents a prescribed correction coefficient. The correction coefficient C is a coefficient for adjusting the intensity I(w) to match the pattern formed in actual video. w is represented by the following Formula (2).
(64)
where a represents a radius of the infrared transmissive portion. q and R respectively correspond to a distance between a point p and a point where a perpendicular to the image sensor from the center of the infrared transmissive portion intersects with the image sensor, and a distance between the center of the infrared transmissive portion and the point p, when any point on the image sensor is set as the point p.
(65)
(66) Accordingly, assuming that an intensity of the near-infrared signal at a position X on the image corresponding to the unit that transmits near-infrared light (i.e., the infrared transmissive portion) is represented by I.sub.NIR_G(X), an intensity I.sub.NIR_G(X,x) that is observed in a pixel located at a position x by the near-infrared light transmitted through the same infrared transmissive portion is represented by the following Formula (3).
I.sub.NIR_G(X,x)=k.sub.X.fwdarw.xI.sub.NIR_G(X)(3)
where k.sub.X.fwdarw.x represents a coefficient calculated by using Formulas (1) and (2) from a distance between the position X and the position x on the image sensor. However, the method for calculating the coefficient k.sub.X.fwdarw.x is not limited to this method. As the method for calculating the coefficient k.sub.X.fwdarw.x, for example, when spectral distribution of the near-infrared signal I.sub.NIR_G(X) is known, there is a method for combining coefficients calculated by using Formulas (1) and (2) at each wavelength based on the spectral distribution. The coefficient k.sub.X.fwdarw.x can also be obtained by calculation based on the standard spectral distribution of near-infrared light in a capturing scene, or computation in advance using another means.
(67) Further, light that reaches the pixel located at the position x is mixed light of visible light and near-infrared light. Specifically, the first color signal in the pixel located at the position x is obtained by adding the second color signal, which is the color signal composed only of visible light components, and a signal representing near-infrared light that reaches the pixel from a plurality of infrared transmissive portions constituting the pattern on the NIR cut filter. Accordingly, the first color signal can be represented by I.sub.G_NIR(x) illustrated in Formula (4). In the formula, I.sub.G(x) corresponds to the second color signal representing a G component in the visible light components.
(68)
where the relationship between the second color signal I.sub.G(X) and the near-infrared signal I.sub.NIR_G(X) is represented by Formula (5) when using an intensity ratio m.sub.G. Formula (4) can be transformed into Formula (6) by using Formula (5).
(69)
(70) The use of the model formula represented by Formula (6) enables estimation of the second color signal and the intensity ratio from the first color signal. Assuming that a vector including the first color signal (I.sub.G_NIR(x)) for each pixel as an element is represented by I.sub.G_NIR, I.sub.G_NIR can be theoretically represented by Formula (7).
I.sub.G_NIR=I.sub.G+KD(M.sub.G)SI.sub.G(7)
where I.sub.G represents a vector including the second color signal (I.sub.G(x)) for each pixel as an element; S represents a sampling matrix for extracting the second color signal in a portion that transmits the near-infrared light; D(M.sub.G) represents a diagonal matrix including, as a diagonal element, each element of a vector M.sub.G including the value of an intensity ratio (m.sub.G) for each portion that transmits the near-infrared light as an element; and K represents a matrix including the value of the coefficient k.sub.X.fwdarw.x as an element.
(71) The second color signal I.sub.G and the intensity ratio M.sub.G are obtained by calculating a value for minimizing an energy function E represented by the following Formula (8).
(72)
(73) The first term on the right side of Formula (8) represents a value larger than 0 unless the second color signal I.sub.G and the intensity ratio M.sub.G satisfy the relationship of Formula (7). The second and third terms on the right side of Formula (8) are regularization terms for preventing the energy minimization in the first term on the right side derived from Formula (7) from being ill-posed. These terms are terms obtained by multiplying coefficients .sub.1 and .sub.2 that are preliminarily set to cost functions C.sub.1(I.sub.G) and C.sub.2(M.sub.G) for evaluating spatial smoothness of the second color signal and the intensity ratio.
(74) The cost functions for evaluating the smoothness are represented by, for example, C(p) in Formulas (9) and (10). In this case, .sub.p represents a set of pixels constituting the image sensor, and N(x) represents a function indicating a position of a spatially adjacent pixel group. Further, p(x) represents data (the second color signal I.sub.G or the intensity ratio M.sub.G) corresponding to the pixel located at the position x on the image sensor, and p(y) represents data (the second color signal I.sub.G or the intensity ratio M.sub.G) corresponding to any arbitrary pixel in a pixel group spatially adjacent to the pixel located at the position x.
(75)
(76) The second color signal I.sub.G and the intensity ratio M.sub.G are calculated in such a manner that, specifically, the second color signal estimation unit 240 updates the value by using a repeat operation. This repeat operation is a repeat operation in which I.sup.0.sub.G is set as I.sub.G NIR and values of all elements of M.sup.0.sub.G are set as a matrix of 1.0, and the update formulas as represented by Formulas (11) and (12) are repeated until the amount of update becomes sufficiently small.
(77)
where V.sup.t.sub.G and M.sup.t.sub.G respectively represent I.sub.G and M.sub.G at a repetition number t.Math.k.sup.t represents a coefficient for adjusting the amount of updateat the repetition number t, and satisfies 0<k.sup.t<1; E represents a unit matrix; and superscript T represents the transposition of a matrix.
(78) Thus, when the second color signal I.sub.G and the intensity ratio M.sub.G are calculated, the near-infrared signal I.sub.NIR_G can be calculated. Specifically, when the near-infrared signal calculation unit 250 substitutes the second color signal I.sub.G and the intensity ratio M.sub.G into Formula (5), thereby calculating a near-infrared signal I.sub.NIR_G.
(79) The second color signal and the near-infrared signal for R and B components can also be calculated in a manner similar to those for the G component. Specifically, by a calculation similar to that for the G component, the second color signal I.sub.R and the near-infrared signal I.sub.NIR_R for the R component, and the second color signal I.sub.B and the near-infrared signal I.sub.NIR_B for the B component can be calculated.
(80) A near-infrared signal output from the video processing device 200 is obtained by adding the near-infrared signals corresponding to the R, G, and B components, respectively. Specifically, assuming that the output near-infrared signal is represented by I.sub.NIR, I.sub.NIR is represented by the following Formula (13).
I.sub.NIR=I.sub.NIR_R+I.sub.NIR_G+I.sub.NIR_B(13)
(81) The video processing device 200 executes such arithmetic processing, thereby outputs video data including the near-infrared signal I.sub.NIR and second color signals I.sub.R, I.sub.G, and I.sub.B. The video processing device 200 can obtain the second color signal and the near-infrared signal from the first color signal only by preparing an NIR cut filter and pattern information corresponding to the NIR cut filter. In this case, the imaging device does not require any special configuration other than the NIR cut filter.
Third Example Embodiment
(82)
(83) The NIR cut filter 311 is an optical filter having a configuration similar to that of the NIR cut filter according to the first and second example embodiments. The NIR cut filter 311 is provided on the front side in a travelling direction of the incident light with respect to the color filter 312 and the photo sensor 313. The NIR cut filter 311 is provided at a prescribed distance from the color filter 312 and the photo sensor 313 so that the near-infrared light diffused by diffraction is received by the photo sensor 313. The NIR cut filter 311 may be detachably or movably configured.
(84)
(85) The color filter 312 is a three-color optical filter having a typical configuration. The color filter 312 has, for example, spectral characteristics as illustrated in
(86)
(87) The pattern of the NIR cut filter 311 may have a correspondence relation with an array of pixels of the photo sensor 313. The term correspondence relation described herein indicates that, for example, an interval between infrared transmissive portions of the NIR cut filter 311 is equal to an interval between the pixels of the photo sensor 313, or has an integral multiple relationship. Specifically, the infrared transmissive portions of the NIR cut filter 311 may be provided so as to overlap the pixels corresponding to a specific color in the photo sensor 313. However, the pattern of the NIR cut filter 311 need not necessarily have the correspondence relation with the array of pixels of the photo sensor 313.
(88) The capturing device 300 has a configuration as described above. The capturing device 300 can generate video data represented by four colors (four components) of R, G, B, and NIR based on video data represented by three colors of R, G, and B in this configuration. Major points of operations of the capturing device 300 are the same as those of the video processing device 200 of the second example embodiment. However, the capturing device 300 executes demosaicing processing prior to the operations described in the second example embodiment.
(89)
(90) In the following description, pieces of color information (color signal values) each representing RGB colors at coordinates (i,j) are respectively represented by R(i,j), G(i,j), and B(i,j). For example, R(1,1) represents color information about the R component of the pixel at the coordinates (1,1). The color information obtained at the time when demosaicing processing is executed actually includes NIR components. However, it is assumed herein that, for convenience of explanation, the NIR components of the color information are not taken into consideration.
(91) The pixel at the coordinates (1,1) corresponds to the R component. Accordingly, the color information about the R component at the coordinates (1,1) is represented by the following Formula (14).
R(1,1)=R(1,1)(14)
(92) On the other hand, the pixel at the coordinates (1,1) does not receive other color components. Accordingly, the color information about the G and B components of the pixel at the coordinates (1,1) is obtained by interpolating peripheral pixels as expressed by Formulas (15) and (16).
G(1,1)=(G(2,1)+G(1,2))/2(15)
B(1,1)=B(2,2)(16)
(93) Next, the color information about the pixel at the coordinates (1,2) is expressed by Formulas (17) to (19).
G(1,2)=G(1,2)(17)
R(1,2)=R(1,1)(18)
B(1,2)=B(2,2)(19)
(94) Note that the color information about the pixel at the coordinates (2,1) is obtained in a manner similar to that for the color information about the pixel at the coordinates (1,2). Further, the color information about the pixel at the coordinates (2,2) is obtained in a manner similar to that for the color information about the pixel at the coordinates (1,1).
(95) The video processing unit 320 executes such processing on all pixels to acquire color information for each color. Subsequently, the video processing unit 320 calculates the near-infrared signal by the operations described in the second example embodiment. The demosaicing processing is not limited to the method described above, but instead may be executed by using, for example, methods disclosed in NPLs 4 to 6.
(96) The capturing device 300 can provide advantageous effects similar to those of the video processing device 200 according to the second example embodiment. Further, the capturing device 300 can disperse infrared light by diffraction in the NIR cut filter 311. With this configuration, even when infrared light having an intensity at which the near-infrared signal is saturated is incident on the photo sensor 313, the capturing device 300 can decrease the intensity of the near-infrared signal per pixel and can increase an apparent dynamic range.
(97) The video processing unit 320 may neglect the near-infrared light components included in the color signals of R and B components obtained after the demosaicing processing. Specifically, as for the R and B components, the video processing unit 320 may regard the second term on the right side of Formula (7) as 0 (i.e., I.sub.R_NIR=I.sub.R, I.sub.B_NIR=I.sub.B). In this case, since I.sub.NIR_R=I.sub.NIR_B=0 holds, I.sub.NIR=I.sub.NIR_G is established by Formula (13).
(98)
MODIFIED EXAMPLES
(99) Example embodiments of the present invention are not limited to the first to third example embodiments described above. For example, the present invention can also be implemented by aspects of modified examples described below. The present invention may also be implemented by aspects in which the first to third example embodiments and the modified examples are combined as appropriate.
(1) Modified Example 1
(100) In the example embodiments of the present invention, the specific shape of the pattern of the NIR cut filter is not limited, as long as the pattern can be described as pattern information. For example, in the pattern formed in the NIR cut filter, the shape of each infrared transmissive portion is not limited to a circular shape, and all the infrared transmissive portions need not necessarily have the same shape.
(2) Modified Example 2
(101) In the example embodiments of the present invention, the visible light components are not limited to three components of R, G, and B. As the visible light components, for example, three components of cyan (C), magenta (M), and yellow (Y) may be used. Further, the visible light components are not necessarily composed of three components, but instead may be composed of components less or more than three components.
(3) Modified Example 3
(102)
(103) The capturing device 400 includes a prism 410, photo sensors 420, 430, and 440, an NIR cut filter 450, and a video processing unit 460. The prism 410 decomposes incident light and outputs the decomposed light in directions corresponding to the R, G, and B components, respectively. The photo sensors 420(R), 430(G), and 440(B) each generate a signal corresponding to the intensity of incident light of each color.
(104) The NIR cut filter 450 is an optical filter similar to the NIR cut filter 311 of the third example embodiment. There is no need to provide the NIR cut filter 450 in all the photo sensors 420, 430, and 440. The NIR cut filter 450 may be provided in any one of the photosensors (photosensor 420 in
(105) The video processing unit 460 may have a configuration similar to that of the video processing unit 320 described in the third example embodiment. However, in the example illustrated in
(106) The capturing device 500 includes an NIR cut filter 510, a stacked sensor 520, and a video processing unit 530. The NIR cut filter 510 and the video processing unit 530 may have configurations similar to those of the NIR cut filter 450 and the video processing unit 460, respectively, illustrated in
(107) The stacked sensor 520 is a sensor in which sensors 521, 522, and 523 are stacked. The sensor 521 has sensitivity in the wavelength range of the B component. The sensor 522 has sensitivity in the wavelength range of the G component. The sensor 523 has sensitivity in the wavelength ranges of the R component and the near-infrared light components.
(4) Modified Example 4
(108) The whole or a part of the configuration according to the present invention can be implemented by a computer. For example, the video processing devices 100 and 200 and the video processing unit 320 can be implemented by a processing device (processor), such as a central processing unit (CPU), and a memory. The present invention may also be implemented by a general-purpose processor or a processor dedicated to video processing.
(109) The present invention may also be provided in the form of a program that can be executed by a computer. This program may be provided in the form in which the program is downloaded from another device (a server or the like) via a network, or may be provided in the form of a computer-readable storage medium. Furthermore, the present invention can be provided not only as a video processing device, a capturing device, a program, and a storage medium, but also as a video processing method.
(110) The present invention has been described above by citing the example embodiments described above as exemplary embodiments. However, the present invention is not limited to the example embodiments described above. In other words, the present invention can be applied in various forms that can be understood by those skilled in the art within the scope of the present invention.
(111) This application is based upon and claims the benefit of priority from Japanese patent application No. 2015-184885, filed on Sep. 18, 2015, the disclosure of which is incorporated herein in its entirety by reference.
REFERENCE SIGNS LIST
(112) 10, 311 NIR cut filter 12, 311a Infrared transmissive portion 100, 200 Video processing device 110 Acquisition unit 120 Signal processing unit 210 Video data acquisition unit 220 First color signal acquisition unit 230 Pattern storage unit 240 Second color signal estimation unit 241 Initial value estimation unit 242 Estimated value selection unit 243 Smoothness evaluation unit 244 First color signal estimation unit 245 Error calculation unit 246 Estimated value update unit 250 Near-infrared signal calculation unit 260 Output unit 300, 400, 500 Capturing device 310 Light receiving unit 312 Color filter 313 Photo sensor 320 Video processing unit