Video format determination device, video format determination method, and video display device
09967536 ยท 2018-05-08
Assignee
Inventors
Cpc classification
International classification
Abstract
There is provided a video format determination device including a video input unit that receives video having a feature amount for each pixel, a region representative value calculation unit that divides a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions, and then computes representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region, a correction value calculation unit that calculates a correction value to correct the representative values, a data correction unit that corrects the representative values, an inter-region correlation calculation unit that calculates the correlation between the left- and right-eye video regions, and an evaluation determination unit that evaluates the correlation to determine whether input video is in the three-dimensional video format.
Claims
1. An apparatus included in a display device, the apparatus comprising: a transmitter that transmits signals to shutter glasses; and a video format determination device including processing circuitry configured to: receive at least one video having a feature amount for each pixel; divide a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into a plurality of blocks having M rows and N columns, respectively, and compute representative values of feature amounts of the respective plurality of blocks for each of the left-eye video region and the right-eye video region, both M and N being 2 or more; calculate a correction value to correct the representative values based on an average value of the feature amounts of the left-eye video region and an average value of the feature amounts of the right-eye video region; correct the representative values of the respective plurality of blocks computed for one of the left-eye video region and the right-eye video region using the correction value; calculate a correlation between the left-eye video region and the right-eye video region by cumulatively adding differences of the representative values of the corresponding plurality of blocks in the left-eye video region and the right-eye video region; calculate an intra-region correlation that is a correlation value within one of the left-eye video region and the right-eye video region or is an average of the correlation value in the left-eye video region and the correlation value in the right-eye video region; and determine whether the input video is in the three-dimensional video format or in a two-dimensional video format based on a ratio between the correlation and the intra-region correlation.
2. The apparatus according to claim 1, wherein the processing circuitry is configured to use, for each respective pixel, a luminance value of the pixel as the feature amount for the pixel.
3. The apparatus according to claim 1, wherein the processing circuitry is configured to calculate the correlation between the left-eye video region and the right-eye video region based on the differences of the representative values of the corresponding plurality of blocks in the left-eye video region and the right-eye video region.
4. The apparatus according to claim 1, wherein the processing circuitry is configured to calculate the correlation between the left-eye video region and the right-eye video region by performing cumulative weighted addition of the differences of the representative values of the corresponding plurality of blocks between the left-eye video region and the right-eye video region.
5. The apparatus according to claim 4, wherein the processing circuitry is configured to perform the cumulative weighted addition using a weighting function that suppresses addition of a difference whose value is equal to or lower than a predetermined value.
6. The apparatus according to claim 1, wherein the processing circuitry is further configured to calculate an evaluation value based on the ratio between the correlation and the intra-region correlation, and determine whether the input video is in the three-dimensional video format or in the two-dimensional video format based on the evaluation value.
7. The apparatus according to claim 6, wherein the processing circuitry is configured to defer determination of a video format of the input video when the correlation within at least one region of the left-eye video region and the right-eye video region has a value equal to or lower than a predetermined threshold value.
8. The apparatus according to claim 1, wherein the processing circuitry is further configured to dispose the left-eye video region and the right-eye video region on the left and right of the input video, respectively, and compute the representative values of the respective plurality of blocks in a determination mode to determine whether the input video is the three-dimensional video in a side-by-side format.
9. The apparatus according to claim 1, wherein the processing circuitry is further configured to dispose the left-eye video region and the right-eye video region on the top and bottom of the input video, respectively, and compute the representative values of the respective plurality of blocks in a determination mode to determine whether the input video is the three-dimensional video in a top-and-bottom format.
10. The apparatus according to claim 1, wherein a plurality of determination modes in which a video format is determined with regard to each of a plurality of three-dimensional video formats is provided, and the processing circuitry is configured to, when the input video is determined to not be a three-dimensional video in a certain determination mode, determine whether the input video is the three-dimensional video by switching to another determination mode.
11. The apparatus according to claim 1, wherein the processing circuitry is configured to, during determination of a video format of a moving image, decide whether a determination state of the video format should be transitioned based on a determination result of the video format for a current frame, a determination state of the video format for the previous frame, and a time elapsed from a change in the determination result.
12. The apparatus according to claim 1, wherein the processing circuitry is configured to set a threshold value to be used in determining an evaluation value according to a determination state of a video format.
13. A video format determination method, comprising: receiving video having a feature amount for each pixel; dividing a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into a plurality of blocks having M rows and N columns, respectively, and computing representative values of feature amounts of the respective plurality of blocks for each of the left-eye video region and the right-eye video region, both M and N being 2 or more; calculating, by processing circuitry, a correction value to correct the representative values based on an average value of the feature amounts of the left-eye video region and an average value of the feature amounts of the right-eye video region; correcting, by the processing circuitry, the representative values of the respective plurality of blocks computed for one of the left-eye video region and the right-eye video region using the correction value; calculating, by the processing circuitry, a correlation between the left-eye video region and the right-eye video region by cumulatively adding differences of the representative values of the corresponding plurality of blocks in the left-eye video region and the right-eye video region; calculating, by the processing circuitry, an intra-region correlation that is a correlation value within one of the left-eye video region and the right-eye video region or is an average of the correlation value in the left-eye video region and the correlation value in the right-eye video region; and determining, by the processing circuitry, whether the input video is in the three-dimensional video format or in a two-dimensional video format based on a ratio between the correlation and the intra-region correlation.
14. A display device, comprising: a display panel; and video signal processing circuitry configured to: receive at least one video signal; divide a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in the input video signal into plurality of blocks having M rows and N columns, respectively, and compute representative values of feature amounts of the respective plurality of blocks for each of the left-eye video region and the right-eye video region, both M and N being 2 or more; calculate a correction value to correct the representative values based on an average value of the feature amounts of the left-eye video region and an average value of the feature amounts of the right-eye video region; correct the representative values of the respective plurality of blocks computed for one of the left-eye video region and the right-eye video region using the correction value; calculate a correlation between the left-eye video region and the right-eye video region by cumulatively adding differences of the representative values of the corresponding plurality of blocks in the left-eye video region and the right-eye video region; calculate an intra-region correlation that is a correlation value within one of the left-eye video region and the right-eye video region or is an average of the correlation value in the left-eye video region and the correlation value in the right-eye video region; generate an evaluation result, based on a ratio between the correlation and the intra-region correlation, that indicates whether the input video signal is in the three-dimensional video format or in a two-dimensional video format; and process the input signal by switching display formats according to a format indicated in the evaluation result.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
(27)
(28)
DETAILED DESCRIPTION OF THE EMBODIMENT(S)
(29) Hereinafter, preferred embodiments of the present disclosure will be described in detail with reference to the appended drawings. Note that, in this specification and the appended drawings, structural elements that have substantially the same function and structure are denoted with the same reference numerals, and repeated explanation of these structural elements is omitted.
(30)
(31) The display device 2000 alternately displays left-eye videos L and right-eye videos R in a time division manner. On the other hand, the shutter glasses 2100 are synchronized with switching timings of the left-eye videos L and the right-eye videos R on the display device 2000 side so as to perform switching to opening and closing of left and right shutter lenses 2101 and 2102. In communication between the display device 2000 and the shutter glasses 2100, a wireless network using radio wave communication such as Wi-Fi, IEEE802.15.4, or the like is used so that packets in which information necessary for controlling opening and closing timings of the left and right shutter lenses 2101 and 2102 is described are transmitted from the display device 2000 to the shutter glasses 2100. Of course, infrared communication or communication methods other than the wireless network can be applied thereto.
(32) The display device 2000 includes a video signal processing unit 2001, a timing control unit 2002, a gate driver 2003, a data driver 2004, a liquid crystal display panel 2005, and a communication unit 2007.
(33) The liquid crystal display panel 2005 includes a liquid crystal layer and transparent electrodes which face each other with the liquid crystal layer interposed therebetween, color filters, and the like (none of which is shown). In addition, on the back side of the liquid crystal display panel 2005, a backlight (surface light source) 2006 is disposed. The backlight 2006 includes LEDs (Light Emitting Diodes), and the like having satisfactory persistence characteristics.
(34) In the video signal processing unit 2001, an image quality correction process for enhancing sharpness of video images and improving contrast is performed.
(35) In addition, in the video signal processing unit 2001, a video format determination device to be described later is installed so as to switch display formats of input video signals based on a determination result of a video format. In other words, it is determined whether an input video signal is a two-dimensional video signal or a three-dimensional video signal, and when the signal is a three-dimensional video signal, the format thereof is determined. Then, when the input signal is determined to be a three-dimensional video signal, the signal is separated into left-eye video images L and right-eye video images R, and correctly arranged along the time axis, and left and right video signals are output in order to alternately display the left-eye video images L and the right-eye video images R on the liquid crystal display panel 2005 in a time division manner.
(36) Left-eye video signals and right-eye video signals which are converted in the video signal processing unit 2001 are input to the timing control unit 2002. The timing control unit 2002 converts the input left-eye video signals and right-eye video signals into signals to be input to the liquid crystal display panel 2005, and generates pulse signals used to cause a panel driving circuit that includes the gate driver 2003 and the data driver 2004 to operate.
(37) The gate driver 2003 is a drive circuit that generates signals for sequential driving, and outputs a driving voltage to gate bus lines connected to each pixel in a display panel 134 according to signals transmitted from the timing control unit 2002. In addition, the data driver 2004 is a drive circuit that outputs driving voltages based on video signals, and generates and outputs signals applied to data lines based on signals transmitted from the timing control unit 2002.
(38) In addition, in order to compensate for response speed of the liquid crystal display panel 2005, over-driving is appropriately performed. Over-driving is a process for improving response characteristics of the panel by applying a driving voltage equal to or higher than a target voltage of the panel drive circuit to liquid crystal elements so as to quickly reach the target voltage.
(39) The communication unit 2007 operates as an access point in the wireless network such as Wi-Fi, or IEEE802.15.4, and includes one or more shutter glasses 2100 operating as terminals in its own basic service set (BSS). The communication unit 2007 transmits packets in which information necessary for controlling opening and closing timings of the left and right shutter lenses 2101 and 2102 on the shutter glasses 2100 side is described.
(40)
(41) In addition,
(42) The display device 2000 alternately displays the left-eye video image L and the right-eye video image R on the liquid crystal display panel 2005 by each field in a time division manner. On the shutter glasses 2100 side, the left and right shutter lenses 2101 and 2102 are synchronized with switching of video images of each field of the display device 2000 and alternately perform opening and closing operations. The brain of the user who observes displayed images through the shutter glasses 2100 fuses the left-eye video images L and the right-eye video images R and thereby three-dimensionally recognizes the video images displayed on the display device 2000.
(43) Note that the technology disclosed in the present specification can be applied not only to the three-dimensional video display format using active glasses such as the shutter glasses 2100 as described above but also to a three-dimensional video display format for the naked eye and to a three-dimensional video display format using passive glasses.
(44)
(45) Herein, the video signal input to the video format determination device 100 is assumed to be a video signal having resolution of, for example, an HD (High Definition) signal of 60 Hz having a pixel size of 19201080, an SD (Standard Definition) signal having a pixel size of 720480, or a still image signal recorded by a digital camera.
(46) The video format determination device 100 includes a feature amount calculation unit 110, a region representative value calculation unit 120, a correction value calculation unit 130, a data correction unit 140, a cumulative difference value calculation unit 150, an intra-region correlation calculation unit 160, an evaluation value calculation unit 170 and a determination unit 180.
(47) The feature amount calculation unit 110 calculates a feature amount when a video format is determined based on an input video signal. To calculate the feature amount, a method of excluding high-frequency information using a low-frequency pass filter, and a method of acquiring edge information using a band-pass filter or slope calculation based on the differences between adjacent pixels are considered, but any method may be used. In addition, particularly, a luminance value of input video may be used as it is without calculating a feature amount. Hereinbelow, a luminance value of input video is set to be used as a feature amount as it is.
(48) The region representative value calculation unit 120 divides a left-eye video region and a right-eye video region in a video frame respectively into a plurality of small regions according to determination modes (in other words, according to a three-dimensional video format to be determined), and then computes a representative value of feature amounts calculated by the feature amount calculation unit 110 for each small region. Herein, as a representative value, an average value of feature amounts (luminance values) of each region is calculated. Which three-dimensional video format should be determined is instructed automatically or manually by a user from, for example, the outside of the video format determination device 100.
(49) When the determination mode in which the side-by-side three-dimensional video format is determined is set, for example, the region representative value calculation unit 120 divides the input video into two halves on the left and right sides, and then divides respective left and right regions into MN small regions, in other words, into 2MN small regions in total as shown in
(50) Next, the region representative value calculation unit 120 computes an average value from feature amounts of each of the small regions as, for example, a representative value of the small regions. The average value of small regions L(p, q) and R(p, q) on the respective left and right sides shown in
(51)
(52) Wherein, each of the variables in the above formulas (1) and (2) is defined as follows.
(53) L(p, q): A representative value of the small regions for a left-eye video image
(54) R(p, q): A representative value of the small regions for a right-eye video image
(55) F(x, y): A feature amount (luminance value) in a unit of a pixel of input video
(56) x: A horizontal pixel position of input video, y: A vertical pixel position of input video
(57) X: The number of horizontal pixels of F(x, y), Y: The number of vertical lines of F(x, y)
(58) M: The number of divided regions of respective left and right video regions in the horizontal direction
(59) N: The number of divided regions of respective left and right video regions in the vertical direction
(60) P: The number of horizontal pixels in each of the divided regions (P=X/M)
(61) Q: The number of vertical pixels in each of the divided regions (Q=Y/N)
(62) p: A horizontal pixel position in a unit of a small region, q: A vertical pixel position in a unit of a small region
(63) In addition, when a determination mode in which the top-and-bottom three-dimensional video format is determined is set, the region representative value calculation unit 120 divides input video into two halves on the upper and lower sides, and then divides the respective upper and lower regions into MN small regions, in other words, into M2N small regions in total as shown in
(64) Note that, as methods for dividing input video, a method in which contracted regions of blocks obtained by dividing a left-eye video region and a right-eye video region so as to respectively have M rows and N columns are used (refer to
(65) In addition, the number of divisions within the range not affected by parallax is also a characteristic of the technology disclosed in the present specification. In other words, in the examples shown in
(66) The correction value calculation unit 130 calculates, for a feature amount of input resolution supplied from the feature amount calculation unit 110, average values APL_L and APL_R of feature amounts (luminance values) of each of the left and right video regions respectively as representative values of the regions as shown in the following formulas (3) and (4), separate from the calculation of the representative value for the small regions in the region representative value calculation unit 120.
(67)
(68) Then, the difference between the calculated values APL_L and APL_R is computed as a correction value a shown in the following formula (5), and supplied to the data correction unit 140 in the latter stage.
=APL_LAPL_R(5)
(69) When the representative value L(p, q) of each small region in the left-eye video region is acquired from the region representative value calculation unit 120, the data correction unit 140 corrects the value according to the following formula (6) using the correction value a supplied from the correction value calculation unit 130, and then acquires a representative value L(p, q) after correction.
L(p,q)=L(p,q)(6)
(70) With an operation of the representative values shown in the above formula (6), the influence of luminance, contrast, and (differences between the left and right video regions attributable to the differences of characteristics of left and right cameras) can be eliminated for three-dimensional video acquired using a twin lens camera.
(71) The cumulative difference value calculation unit 150 obtains a cumulative sum S obtained by weighting the difference of the representative values of the corresponding small regions between the left- and right-eye video regions. However, as the representative value of each small region on the left-eye video region side, the correction value L(p, q) obtained in the data correction unit 140 is used. The cumulative sum S can be calculated according to the following formula (7), and the sum serves as a correlation value S between the left-eye and right-eye video regions in the video.
(72)
(73) Note that, as a weight W (in the above formula (7)) given to the difference of representative values of corresponding small regions of the left- and right-eye video regions, a weighting function that suppresses addition of a difference that is equal to or lower than a predetermined value is used. Specifically, a weighting function as shown in
(74)
(75) The intra-region correlation calculation unit 160 calculates the correlativity of each video signal within the left- and right-eye video regions. For an intra-region correlation C for the whole video, either of an intra-region correlation value CL obtained in the left-eye video region and an intra-region correlation value CR obtained in the right-eye video region may be used, or the intra-region correlation C may be obtained from the average of CL and CR as shown in the following formula (9).
C=(CL+CR)/2(9)
(76) Herein, the intra-region correlation CL in the left-eye video region and the intra-region correlation value CR in the right-eye video region can be respectively obtained according to the following formulas (10) and (11).
(77)
(78) The evaluation value calculation unit 170 calculates an evaluation value Z by obtaining the ratio of the correlation value S between the left- and right-eye video regions obtained by the cumulative difference value calculation unit 150 to the intra-region correlation C in the left- and the right-eye video regions obtained by the intra-region correlation calculation unit 160 according to the following formula (12).
Z=S/C(12)
(79) The determination unit 180 determines whether the input video is of a three-dimensional video signal in the side-by-side format, or the like, or a two-dimensional video signal based on the computed evaluation value Z. When the input video is of a three-dimensional video format, the correlativity of the left-eye video region and the right-eye video region in the input video increases. In this case, since the correlation value S between the regions shown in the above formula (7) becomes small, the evaluation value shown in the above formula (12) decreases. Thus, the determination unit 180 can determine that the input video is of a three-dimensional video format if the evaluation value Z output from the evaluation value calculation unit 170 is low.
(80)
(81) The determination unit 180 first checks whether the intra-region correlation C computed by the intra-region correlation calculation unit 160 is equal to or higher than a predetermined threshold value th_c (Step S701).
(82) At this moment, when the intra-region correlation C is less than the predetermined threshold value th_c (Yes in Step S701), the determination result D is output as 1, in other words, determination is deferred (Step S702). When the intra-region correlation C has a value close to 0, the reason is that there is a high possibility that the entire screen is flat, and thus there is not sufficient information for determining a video format.
(83) On the other hand, when the intra-region correlation C is equal to or higher than the threshold value th_c (No in Step S701), sufficient information for determining a video format can be obtained from the input video, and thus determination of a video formation is attempted based on the evaluation value Z computed by the evaluation value calculation unit 170 (Step S703).
(84) In the present embodiment, two threshold values of th_a and th_b are used in the determination process of a video format based on the evaluation value Z (wherein, th_a<th_b).
(85) If the evaluation value Z is low, the input video can be determined to be in a three-dimensional video format (as described above). Thus, when the evaluation value Z is smaller than the threshold value th_a, the determination unit 180 determines that the input video is three-dimensional video, and outputs 0 as the determination result D. On the other hand, when the evaluation value Z is equal to or greater than the threshold value th_b, the determination unit 180 determines that the input video is two-dimensional video, and outputs 2 as the determination result D. In addition, when the evaluation value Z is equal to or greater than th_a and smaller than th_b, the determination unit 180 defers determination, and outputs 1 as the determination result D.
(86) Note that the threshold values th_a and th_b are determined to be values, for example, 0.2 and 0.4, or the like. In addition, determination of the evaluation value Z may be performed using positions of formulas having three or more stages (for example, 10 stages) rather than threshold values of two stages as shown in
(87) When a still image in the side-by-side format is desired to be determined in the video format determination device 100, a side-by-side determination mode is set as a determination mode. When the determination result D obtained from the video format determination device 100 is 0, which is for three-dimensional video, the image is determined to be in the side-by-side format, and when the result is 2, which is for two-dimensional video, the image is determined to be two-dimensional video. In addition, when the determination result D is 1, which is for deferment of determination, the image is determined to be two-dimensional video because there is also a possibility of erroneous determination.
(88) In addition, when a still image in the top-and-bottom format is desired to be determined in the video format determination device 100, a top-and-bottom determination mode is set as a determination mode. When the determination result D obtained from the video format determination device 100 is 0, which is for three-dimensional video, the image is determined to be in the top-and-bottom format, and when the result is 2, which is for two-dimensional video, the image is determined to be two-dimensional video. In addition, when the determination result D is 1, which is for deferment of determination, the image is determined to be two-dimensional video because there is also a possibility of erroneous determination.
(89) In addition, when there is no knowledge of whether the input video is of a three-dimensional video signal in either of the side-by-side format or the top-and-bottom format, or of a two-dimensional video signal, it is necessary for the video format determination device 100 to determine such a plurality of video formats.
(90)
(91) First, by setting the determination mode to be the side-by-side (SBS) determination mode, a determination process is performed, and the determination result D is obtained (Step S901).
(92) Herein, when the obtained determination result D is 0 (Yes in Step S902), a final determination result E=SBS which indicates that the format of the input video is the side-by-side format is output (Step S903), and then this process routine ends.
(93) In addition, when the obtained determination result D is not 0 (No in Step S902), the determination mode is subsequently switched to the top-and-bottom (TAB) determination mode, the determination process is performed, and then the determination result D is obtained again (Step S904).
(94) Herein, when the obtained determination result D is 0 (Yes in Step S905), a final determination result E=TAB which indicates that the format of the input video is the top-and-bottom format is output (Step S906), and then this process routine ends.
(95) In addition, when the obtained determination result D is not 0 (No in Step S905), it is checked whether or not the determination result D is 1 (Step S907). Then, when D is 1 (Yes in Step S907), a final determination result E=HOLD which indicates that the determination is deferred is output (Step S908), and when D is not 1 (No in Step S907), a final determination result E=2D which indicates that the input video is in the two-dimensional video format is output (Step S909), and the process routine ends.
(96) As described above, the display device 2000 mounted with the video format determination device 100 switches display formats of input video based on determination results of video formats. Herein, in the determination methods as shown in
(97)
(98) In the 2D state, if the determination result of the current frame is maintained to be 2D, the state returns to the 2D state. On the other hand, if the determination result of the current frame is change to SBS and a predetermined period of time elapses (in other words, if the determination result is SBS over several consecutive frames), the state is transitioned from the 2D state to SBS. In addition, if the determination result of the current frame is changed to TAB and a predetermined period of time elapses, the state is transitioned from the 2D state to TAB.
(99) In addition, in the SBS state, if the determination result of the current frame is maintained to be SBS, the state returns to the SBS state. On the other hand, if the determination result of the current frame is changed to TAB and a predetermined period of time elapses, the state is transitioned from the SBS state to the TAB state. In addition, if the determination result of the current frame is changed to 2D and a predetermined period of time elapses, the state is transitioned from the SBS state to the 2D state.
(100) In addition, in the TAB state, if the determination result of the current frame is maintained to be TAB, the state returns to the TAB state. On the other hand, if the determination result of the current frame is changed to 2D and a predetermined period of time elapses, the state is transitioned from the TAB state to the 2D state. In addition, if the determination result of the current frame is changed to SBS and a predetermined period of time elapses, the state is transitioned from the TAB state to the SBS state.
(101)
(102) First, the determination unit 180 sets different determination threshold values in each state according to a state Tin a previous frame (Step S1101).
(103) In the 2D state, for example, determination threshold values that facilitate detection of two-dimensional video such as th_a=0.15 and th_b=0.3 are set so that transition from detection of the two-dimensional video does not frequently occur. In addition, in the case of a three-dimensional video state such as the SBS state or the TAB state, determination threshold values that facilitate detection of three-dimensional video such as th_a=0.25 and th_b=0.45 are set so that transition from detection of the three-dimensional video does not frequently occur.
(104) When a determination result is ambiguous, the determination unit 180 is made to respond only when video in a clearly different format is input without allowing the determination state T to be transitioned, by suppressing the frequency of transition from each state by the determination unit 180 in this manner, and accordingly, the intervals of determination results can be stabilized.
(105) Herein the determination unit 180 retains E for a variable F as the previous determination result (previous frame) before performing determination of a plurality of video formats (Step S1102). Note that, at this moment, when E means HOLD (determination deferred), the value of the variable F is not updated. With this operation, E is not substituted for a previous determination result F. The initial value of E is set to be the same value as T.
(106) Then, the determination unit 180 executes a determination process for the plurality of video formats according to the process shown in
(107) Next, the determination unit 180 checks whether or not the determination result E of the current frame is HOLD (determination deferred) (Step S1104).
(108) Herein, when the determination result E of the current frame is HOLD (determination deferred) (Yes in Step S1104), the process routine ends without updating either of the state T and a counter CNT that measures the time elapsed from detection of a change in the state.
(109) On the other hand, when the determination result E of the current frame is not HOLD (determination deferred) (No in Step S1104), it is further checked whether the determination result E of the current frame is different from the current state T and the same as the previous determination result F (Step S1105).
(110) When the determination result E of the current frame is different from the current state T and the same as the previous determination result F (Step S1105), the counter CNT that measures the elapsed time counts up values (Step S1106).
(111) Then, when a value of the counter CNT that measures the elapsed time exceeds a predetermined threshold value (change time) th_d (Yes in Step S1107), the current state T is updated to the determination result E of the current frame, and the counter CNT is reset so as to have the value of 0 (Step S1108), and the process routine ends.
(112) In addition, when the determination result E of the current frame is the same as the current state T, or different from the previous determination result F (No in Step S1105), the changed determination result is not stabilized, and thus the counter CNT that measures the elapsed time is reset so as to have the value of 0 (Step S1109), and then the process routine ends.
(113) With the process shown in
(114) The characteristics of the video format determination device 100 shown in
(115) (1) The region representative value calculation unit 120 converts an input video signal of high resolution into representative values of feature amounts for every small region. Accordingly, format determination that is hardly affected by parallax, bands of video signals, and noise can be performed. In addition, by performing conversion into representative values of feature amounts of every small region, format determination can be performed with a small calculation amount.
(116) (2) The correction value calculation unit 130 computes a correction value based on an average value of feature amounts computed for each region, and the data correction unit 140 corrects the representative values of each region using this correction value, and thus, when the correlation between the left- and right-eye video regions is calculated in the latter stage, influence of contrast, , and color can be suppressed.
(117) (3) Since the intra-region correlation calculation unit 160 calculates intra-region correlation between the left- and right-eye video regions and the evaluation value calculation unit 170 computes a determination value from the ratio of the inter-region correlation value to the intra-region correlation value at the same time as calculation of the correlation between the regions, format determination that is hardly affected by contrast, , and color can be performed. In addition, since there is little influence of two-dimensional video (refer to
(118) (4) During the calculation of the correlation between the left- and right-eye video regions, the cumulative difference value calculation unit 150 uses a weighting function. Accordingly, by suppressing the influence of parallax, bands of video signals, and noise, determination accuracy of three-dimensional video improves.
(119) (5) The determination unit 180 determines a format of input video using a determination value computed based on an evaluation value of the correlation between the regions, but by providing deferment of determination in determination of a three-dimensional video format and a two-dimensional video format, the risk of erroneous determination can be lowered.
(120) (6) By changing a method of dividing and processing the left- and right-eye video regions, the video format determination device 100 can be applied to determination of a plurality of different three-dimensional video formats, beginning from the side-by-side format and the top-and-bottom format.
(121) (7) In addition, by switching a plurality of determination modes and consecutively executing determination processes, the video format determination device 100 can simultaneously determine a plurality of three-dimensional video formats.
(122) (8) When the video formation of a moving image is determined, the video format determination device 100 sets transition of determination results not to frequently occur, or easily determines a desired video format by setting determination threshold values according to a previous determination value.
(123) (9) When the video formation of a moving image is determined, the video format determination device 100 can stabilize the intervals of determination results with reference to the history of the determination results.
(124)
(125) The video format determination device 1200 includes a feature amount calculation unit 1210, a region representative value calculation unit 1220, a luminance ranking calculation unit 1230, a luminance direction calculation unit 1240, an evaluation value calculation unit 1250, and a determination unit 1260. A video signal input to the video format determination device 1200 is, for example, an HD signal of 60 Hz having a pixel size of 19201080, a video signal having resolution of an SD signal having a pixel size of 720480, or a still image signal recorded by a digital camera. In the case of a video signal, a video format thereof can be determined by processing one frame out of a moving image of which frames are consecutively input.
(126) The feature amount calculation unit 1210 calculates feature amounts when the video format of input video is determined. Hereinafter, luminance values of input video are set to be used as feature amounts.
(127) The region representative value calculation unit 1220 divides a left-eye video region and a right-eye video region in a video frame respectively into a plurality of small regions according to a determination mode instructed from outside of the video format determination device 1200 automatically or manually by a user, and computes a representative value of feature amounts calculated by the feature amount calculation unit 1210 for each small region. Herein, an average value of the feature amounts (luminance values) of each region is calculated as a representative value.
(128) When, for example, a determination mode in which three-dimensional video in the side-by-side format is determined is set, the region representative value calculation unit 1220 divides input video into two halves on the left and right sides, further divides the respective left and right regions into MN small regions (refer to
(129) Note that, as methods for dividing input video, a method in which contracted regions of blocks obtained by dividing a left-eye video region and a right-eye video region so as to respectively have M rows and N columns are used (refer to
(130) In addition, the number of divisions within the range not affected by parallax is also a characteristic of the technology disclosed in the present specification. In other words, in the examples shown in
(131) The luminance ranking calculation unit 1230 ranks the small regions in each row respectively for the left-eye display region on the left half of a video frame and the right-eye display region on the right half thereof based on the representative value of each small region, that is average luminance computed by the region representative value calculation unit 1220. For example, it is assumed that input video is three-dimensional video in the side-by-side format shown in
(132) The luminance direction calculation unit 1240 calculates directions in which luminance changes in the small regions in each of the left-eye display region and the right-eye display region in a video frame based on the representative values of each small region, that is, the average luminance computed in the region representative value calculation unit 1220.
(133) For example, the direction of luminance changes can be obtained for small regions at the center of 33 by multiplying a 33 weighting filter coefficient for each small 33 region. As a weighting filter of this kind, template matching of Prewitt (for example, refer to Japanese Unexamined Patent Application Publication No. 2009-217606) is known. In this matching method, 8 kinds of masks respectively having values and directions of masks as shown in
(134) The evaluation value calculation unit 1250 compares the results obtained from ranking by the luminance ranking calculation unit 1230 for corresponding small regions of the left-eye video region and the right-eye video region. Then, small regions of which the rankings coincide on the left and right sides are set to have 1, those that do not coincide are set to have 0, and the number of regions R of which the rankings coincide on the left and right sides is computed as an evaluation value for ranking. As the number of small regions of which the rankings coincide increases, in other words, as the value of R becomes large, input video can be determined to be three-dimensional video.
(135) In addition, the evaluation value calculation unit 1250 compares the results obtained by calculating the directions of luminance changes using the luminance direction calculation unit 1240 for corresponding small regions of the left-eye video region and the right-eye video region. Then, small regions of which the directions of the luminance changes coincide on the left and right sides are set to have 1, those that do not coincide are set to have 0, and the number of regions P of which the directions of the luminance changes coincide on the left and right sides is computed as an evaluation value for a direction of a luminance change. If the 8 masks shown in
(136) For example, the determination unit 1260 provides three-dimensional video determination threshold values th_R_3D and th_P_3D, and two-dimensional video determination threshold values th_R_2D and th_P_2D for each of the evaluation values R and P. Then, since the degree of similarity of the left- and right-eye video regions increases as the evaluation values become higher, there is a possibility of three-dimensional video. Conversely, since the degree of similarity of the left- and right-eye video regions decreases as the evaluation values become lower, there is a possibility of two-dimensional video. In addition, when a video is not determined to be three-dimensional video or two-dimensional video due to an intermediate degree of similarity, the determination unit 1260 defers determination. Specifically, based on the rankings of luminance and directions of luminance changes, determination can be made using the following respective formulas (12) and (13).
(137)
(138) The determination unit 1260 can determine the format of input video with reference to either or both of the evaluation values R and P calculated by the evaluation value calculation unit based on calculation results by the luminance ranking calculation unit 1230 and the luminance direction calculation unit 1240.
(139) In addition, since the above determination is executed in units of one input frame, there is a possibility of the determination result changing in units of one frame due to changes and noise of an image in the case of a moving image. If the determination result changes in a short period of time, display formats output by the display device 2000 on the screen thereof change too often, and thus there is concern of such changes causing viewers displeasure. Thus, it is necessary to stabilize the intervals of the determination results output from the video format determination device 1200. Therefore, using the following methods, the determination unit 1260 may suppress determination changes in time directions.
(140) (1) When the same determination result is obtained N consecutive times (for N frames), the determination is confirmed.
(141) (2) M or more determination results out of the N consecutive determinations (for N frames) are adopted.
(142) In addition, when there is no determination result that satisfies the condition (1) or (2), an input video cannot be determined to be three-dimensional video or two-dimensional video, and thus the determination unit 1260 may output a result of determination deferred.
(143) The characteristics of the video format determination device 100 shown in
(144) (1) Since determination is made without considering absolute values of luminance in the left- and right-eye video regions, a video format can be accurately determined even for a three-dimensional video image of which luminance significantly deviates in the left- and right-eye video regions.
(145) (2) The left- and right-eye video regions are respectively divided into small regions, and evaluation is performed using a representative value of average luminance values, or the like for each region. Accordingly, frame determination is hardly affected by parallax, bands of video signals, and noise.
(146) (3) The left- and right-eye video regions are respectively divided into small regions, and evaluation is performed using a representative value of average luminance values, or the like for each region. Thus, a calculation amount may be small in comparison to a method in which evaluation is performed in a unit of pixels.
(147) (4) The video format determination device 1200 can be applied to determination of a plurality of different three-dimensional video formats beginning from the side-by-side format and the top-and-bottom format by changing a method of dividing and processing left- and right-eye video regions.
(148) Additionally, the present technology may also be configured as below.
(149) (1) A video format determination device including:
(150) a video input unit that receives video having a feature amount for each pixel;
(151) a region representative value calculation unit that divides a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computes representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region;
(152) a correction value calculation unit that calculates a correction value to correct the representative values based on an average value of feature amounts of the left-eye video region and an average value of feature amounts of the right-eye video region;
(153) a data correction unit that corrects the representative values of the respective small regions computed for one of the left-eye video region and the right-eye video regions using the correction value;
(154) an inter-region correlation calculation unit that calculates the correlation between the left- and right-eye video regions by cumulatively adding differences of the representative values of corresponding small regions in the left-eye video region and the right-eye video region; and
(155) an evaluation determination unit that evaluates the correlation between the left- and right-eye video regions to determine whether input video is in the three-dimensional video format.
(156) (2) A video format determination device including:
(157) a video input unit that receives video having a feature amount for each pixel;
(158) a region representative value calculation unit that divides a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computes representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region;
(159) an inter-region correlation calculation unit that calculates the correlation between the left- and right-eye video regions by cumulatively adding differences of the representative values of corresponding small regions in the left-eye video region and the right-eye video region;
(160) an intra-region correlation calculation unit that calculates the correlation within at least one region of the left-eye video region and the right-eye video region; and
(161) an evaluation determination unit that calculates an evaluation value based on a ratio of the correlation between the left- and right-eye video regions to the correlation within the region to determine whether input video is in the three-dimensional video format based on the evaluation value.
(162) (3) The video format determination device according to (1), wherein the region representative value calculation unit uses a luminance signal of each pixel as a feature amount.
(163) (4) The video format determination device according to (1) or (2), wherein the inter-region correlation calculation unit calculates the correlation between the left- and right-eye video regions based on the differences of the representative values of corresponding small regions in the left-eye video region and the right-eye video region.
(164) (5) The video format determination device according to (1) or (2), wherein the inter-region correlation calculation unit calculates the correlation between the left- and right-eye video regions by performing cumulative weighted addition of the differences of the representative values of corresponding small regions between the left-eye video region and the right-eye video region.
(165) (6) The video format determination device according to (5), wherein the inter-region correlation calculation unit performs the cumulative weighted addition using a weighting function that suppresses addition of a difference whose value is equal to or lower than a predetermined value.
(166) (7) The video format determination device according to (1), further including:
(167) an intra-region correlation calculation unit that calculates the correlation within at least one region of the left-eye video region and the right-eye video region,
(168) wherein the evaluation determination unit calculates an evaluation value based on a ratio of the correlation between the left- and right-eye video regions to the correlation within the region to determine whether input video is in the three-dimensional video format based on the evaluation value.
(169) (8) The video format determination device according to (2) or (7), wherein the evaluation determination unit defers determination of a video format of input video when the correlation within the region has a value equal to or lower than a predetermined threshold value.
(170) (9) The video format determination device according to (1) or (2), wherein the region representative value calculation unit disposes the left-eye video region and the right-eye video region on left and right of the input video, respectively, and computes representative values of the respective small regions in a determination mode to determine whether or not the input video is three-dimensional video in a side-by-side format.
(171) (10) The video format determination device according to (1) or (2), wherein the region representative value calculation unit disposes the left-eye video region and the right-eye video region on top and bottom of the input video, respectively, and computes representative values of the respective small regions in a determination mode to determine whether or not the input video is three-dimensional video in a top-and-bottom format.
(172) (11) The video format determination device according to (1) or (2),
(173) wherein a plurality of determination modes in which a video format is determined with regard to each of a plurality of three-dimensional video formats is provided, and
(174) wherein, when input video is determined not to be three-dimensional video in a certain determination mode, the evaluation determination unit determines whether or not the input video is three-dimensional video by switching to another determination mode.
(175) (12) The video format determination device according to (1) or (2), wherein, during determination of a video format of a moving image, the evaluation determination unit decides whether or not a determination state of a video format should be transitioned based on a determination result of a video format for a current frame, a determination state of a video format for the previous frame, and a time elapsed from a change in the determination result.
(176) (13) The video format determination device according to (12), wherein the evaluation determination unit sets a threshold value to be used in determining an evaluation value according to a determination state of a video format.
(177) (14) A video format determination device including:
(178) a video input unit that receives video having a feature amount for each pixel;
(179) a region representative value calculation unit that divides a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computes representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region;
(180) a representative value ranking calculation unit that calculates rankings of the small regions for each row and each column based on the computed representative values in each of the left-eye video region and the right-eye video region; and
(181) an evaluation determination unit that evaluates a degree of similarity of rankings of corresponding small regions between the left-eye video region and the right-eye video region to determine whether or not the input video is in the three-dimensional video format.
(182) (15) A video format determination device including:
(183) a video input unit that receives video having a feature amount for each pixel;
(184) a region representative value calculation unit that divides a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computes representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region;
(185) a representative value direction calculation unit that calculates directions in which the representative values change in each of the small regions in each of the left-eye video region and the right-eye video region; and
(186) an evaluation determination unit that further evaluates a degree of similarity of directions in which the representative values change in corresponding small regions in the left-eye video region and the right-eye video region to determine whether or not the input video is in the three-dimensional video format.
(187) (16) The video format determination device according to (15), wherein, by multiplying a weighting filter coefficient of n m by each of n m small regions in each of the left-eye video region and the right-eye video region, the representative value direction calculation unit calculates the directions in which the representative values change in each of the small regions.
(188) (17) A video format determination method including:
(189) receiving video having a feature amount for each pixel;
(190) dividing a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computing representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region;
(191) calculating a correction value to correct the representative values based on an average value of feature amounts of the left-eye video region and an average value of feature amounts of the right-eye video region;
(192) correcting the representative values of the respective small regions computed for one of the left-eye video region and the right-eye video regions using the correction value;
(193) calculating the correlation between the left- and right-eye video regions by cumulatively adding differences of the representative values of corresponding small regions in the left-eye video region and the right-eye video region; and
(194) evaluating the correlation between the left- and right-eye video regions to determine whether input video is in the three-dimensional video format.
(195) (18) A video format determination method including:
(196) receiving video having a feature amount for each pixel;
(197) dividing a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computing representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region;
(198) calculating the correlation between the left- and right-eye video regions by cumulatively adding differences of the representative values of corresponding small regions in the left-eye video region and the right-eye video region;
(199) a. calculating the correlation within at least one region of the left-eye video region and the right-eye video region; and
(200) calculating an evaluation value based on a ratio of the correlation between the left- and right-eye video regions to the correlation within the region to determine whether input video is in the three-dimensional video format based on the evaluation value.
(201) (19) A video format determination method including:
(202) receiving video having a feature amount for each pixel;
(203) dividing a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computing representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region;
(204) calculating rankings of the small regions for each row and each column based on the computed representative values in each of the left-eye video region and the right-eye video region; and
(205) evaluating a degree of similarity of rankings of corresponding small regions between the left-eye video region and the right-eye video region to determine whether or not the input video is in the three-dimensional video format.
(206) (20) A video format determination method including:
(207) receiving video having a feature amount for each pixel;
(208) dividing a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computing representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region;
(209) calculating directions in which the representative values change in each of the small regions in each of the left-eye video region and the right-eye video region; and
(210) evaluating a degree of similarity of directions in which the representative values change in corresponding small regions in the left-eye video region and the right-eye video region to determine whether or not the input video is in the three-dimensional video format.
(211) (21) A display device including:
(212) an input unit that receives video signals;
(213) a video signal processing unit that determines the format of input video signals, and performs processing of a three-dimensional video signal or a two-dimensional video signal by switching display formats according to the determination result; and
(214) a display unit that displays video signals that have been processed in the video signal processing unit on a screen,
(215) wherein the video signal processing unit divides a left-eye video region and a right-eye video region in a three-dimensional video format to be determined in input video into small regions having M rows and N columns, respectively, and then computes representative values of feature amounts of the respective small regions for each of the left-eye video region and the right-eye video region, calculates a correction value to correct the representative values based on an average value of feature amounts of the left-eye video region and an average value of feature amounts of the right-eye video region, calculates a correction value to correct the representative values based on an average value of feature amounts of the left-eye video region and an average value of feature amounts of the right-eye video region using the correction value, calculates the correlation between the left- and right-eye video regions by cumulatively adding differences of the representative values of corresponding small regions in the left-eye video region and the right-eye video region, and evaluates the correlation between the left- and right-eye video regions to determine whether input video is in the three-dimensional video format.
(216) Hereinabove, the technology disclosed in the present specification has been described in detail with reference to a specific embodiment. However, it should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof
(217) In the present specification, description has been provided focusing on an embodiment in which a three-dimensional video format such as the side-by-side format and the top-and-bottom format in which left and right videos are transmitted as one frame is determined, but the application range of the technology disclosed in the present specification is not limited to any specific three-dimensional video format. For example, in a case of the frame sequential scheme in which left-eye video and right-eye video are alternately transmitted for each frame, the technology disclosed in the present specification can be realized in the same manner by performing calculation of the correlation between regions and calculation of intra-region correlation.
(218) In addition, the technology disclosed in the present specification can be applied not only to a three-dimensional video display format using active glasses such as the shutter glasses as described above but also to a three-dimensional video display format using passive glasses and a three-dimensional video display format for the naked eye.
(219) It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.