Stereoscopic image display system, disparity conversion device, disparity conversion method, and program
09832445 · 2017-11-28
Assignee
Inventors
Cpc classification
H04N13/122
ELECTRICITY
H04N2013/0081
ELECTRICITY
International classification
Abstract
To reduce burden of a viewer at the time of performing stereoscopic display by a stereoscopic display device and to show a stereoscopic image in a natural and comfortable manner. A disparity estimating unit 210 estimates disparity from the left and right images of an input image to generate a disparity map. A disparity analyzing unit 230 analyzes the disparity map to generate a disparity control parameter for performing suitable disparity control. A disparity control unit 240 controls the content of processing at an image conversion unit 250 in accordance with the disparity control parameter. The image conversion unit 250 performs image conversion on the input image based on the control by the disparity control unit 240 to output an output image. The image conversion unit 250 performs image conversion of two steps. The image conversion unit 250 performs shift processing for shifting the relative positions of the left and right images of the input image in the horizontal direction as first step image conversion, for example. The image conversion unit 250 performs scaling processing for performing enlargement/reduction of the entire screen with each center of the left and right images as a reference, as second step image conversion, for example.
Claims
1. An image processing apparatus comprising: circuitry configured to perform a first image conversion process on a left image and a right image serving as an input stereoscopic image; perform a second image conversion process on the left image and the right image converted by the first image conversion process, the second image conversion process including an enlargement/reduction process with center positions of the left image and right image as references; estimate regions of interest in the left image and the right image of the input stereoscopic image to generate a region-of-interest map illustrating positions and degrees of interest in the input stereoscopic image; and determine a disparity control parameter such that a total number of disparities of the region-of-interest are within a predetermined disparity range.
2. The image processing apparatus according to claim 1, wherein, as part of the first image conversion process, the circuitry is further configured to shift relative positions of the left image and the right image in a horizontal direction; perform the enlargement/reduction process with inner positions of the left image and right image as references; and perform the enlargement/reduction process with outer positions of the left image and right image as references.
3. The image processing apparatus of claim 1, wherein the circuitry is further configured to: estimate disparity from the left image and right image of the input stereoscopic image to generate a disparity map that holds disparity for each pixel or each pixel group; analyze the disparity map to generate a disparity control parameter so that the distribution of the disparity in the input stereoscopic image is included in a predetermined range; and control image conversion for the first and second image conversion processes based on the disparity control parameter.
4. The image processing apparatus of claim 1, wherein the circuitry is further configured to: calculate a disparity histogram based on the disparity map and the region-of-interest map, wherein the disparity histogram includes a left side representing a deeper than a display surface of a stereoscopic display and includes a right side representing a nearer than the display surface of the stereoscopic display.
5. The image processing apparatus of claim 4, wherein the disparity histogram includes a weight factor associated with an influence of an image frame or subject within the region-of-interest.
6. The image processing apparatus of claim 5, wherein the region-of-interest map illustrates that the degree of interest increases with a degree of whiteness in the region displayed.
7. The image processing apparatus of claim 1, wherein the circuitry is further configured to determine a disparity control parameter to minimize a total amount of predetermined stress values of the input stereoscopic image.
8. The image processing apparatus of claim 1, wherein the circuitry is further configured to define a range of stereopsis including a first range on a nearer side of a 3D view and a second range on a deeper side of the 3D view, the nearer side being a side beyond a display surface displaying the stereoscopic image, and position content of the input stereoscopic image between the first range and the second range based on performing the first image conversion process and the second image conversion process.
9. An image processing method comprising: performing, with circuitry, a first image conversion process on a left image and a right image serving as an input stereoscopic image; performing, with the circuitry, a second image conversion process, on the left image and the right image converted by the first image conversion process, the second image conversion process including enlargement/reduction process with center positions of the left image and right image as references; estimating, with the circuitry, regions of interest in the left image and the right image of the input stereoscopic image to generate a region-of-interest map illustrating positions and degrees of interest in the input stereoscopic image; and determining, with the circuitry, a disparity control parameter such that a total number of disparities of the region-of-interest are within a predetermined disparity range.
10. An image processing apparatus comprising: circuitry configured to display a stereoscopic image in a 3-dimensional (3D) view, the stereoscopic image including a left side image and a right side image; perform a shifting process of the left side image and the right side image to change a disparity amount of the input stereoscopic image to a nearer side of the 3D view or a deeper side of the 3D view of a display surface of a stereoscopic display depending on a depth disparity amount between the left side image and right side image; and perform a scaling process to change a range in a depth direction of the input stereoscopic image, the scaling process includes shifting positions of the left side image and the right side image in a horizontal direction using inner positions of the left side image and the right side image as reference points.
11. The image processing apparatus of claim 10, wherein the shifting process includes shifting the left side image and the right side image toward one another or shifting the left side image and the right side image away from one another.
12. The image processing apparatus of claim 10, wherein a pixel located on an inner position of the left side image or the right side image is shifted by a smaller scale than a pixel located on an outer position of the left side image or the right side image.
Description
BRIEF DESCRIPTION OF DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
(27)
(28)
(29)
(30)
(31)
(32)
(33)
(34)
(35)
DESCRIPTION OF EMBODIMENTS
(36) Hereafter, description will be made regarding modes for implementing the present invention (hereafter, referred to as embodiments). Description will be made in accordance with the following sequence. 1. First Embodiment (disparity control according to parameter specification) 2. Second Embodiment (disparity control based on disparity analysis) 3. Third Embodiment (disparity control in consideration of region of interest) 4. Fourth Embodiment (disparity control in consideration of around image frame) 5. Modifications
(37) <1. First Embodiment>
(38) Configuration Example of Stereoscopic Image Display System
(39)
(40) The image storage device 100 is a device configured to store image data for stereoscopic display. Here, the image data is a stereoscopic image with a left image to be perceived by the left eye of a person, and a right image to be perceived by the right eye of the person as a pair, may be a still image made up of a set of the left and right images, or may be a moving image where the left and right images (frames) are arrayed in the time sequence. Note that the image storage device 100 is an example of an input image supply unit described in the Claims.
(41) The disparity conversion device 200 is a device configured to convert the disparity of a stereoscopic image in the image data stored in the image storage device 100. That is to say, the input and output of this disparity conversion device 200 are both stereoscopic images, and conversion is made so that disparities recognized by both eyes differ.
(42) The display control device 300 is a device configured to effect control to display the image data output from the disparity conversion device 200 on the image display device 400. The image display device 400 is a stereoscopic display configured to display the image data as a stereoscopic image. As for a stereoscopic display method, an arbitrary method may be applied such as a method for alternately disposing the left and right images for each scanning line, a method for displaying the left and right images in chronological order, or the like. The display control device 300 performs display control so as to handle the display method of the image display device 400.
(43) Configuration Example of Disparity Conversion Device
(44)
(45) The operation acceptance unit 290 is a user interface configured to accept operation input from a user. As for operation input, a later-described disparity control parameter may directly be input, or a parameter abstracted in accordance with a fixed rule may be input.
(46) The disparity control unit 240 is a unit configured to control the processing content at the image conversion unit 250 based on the operation input accepted by the operation acceptance unit 290, thereby controlling the disparity of the input image 10.
(47) The image conversion unit 250 subjects the input image 10 to image conversion based on the control by the disparity control unit 240 to output an output image 30. As for the image conversion at the image conversion unit 250, processing of two steps is performed as to each of the left and right images, which will be described later.
(48) Configuration Example of Image Conversion Unit
(49)
(50) The left image pre-conversion unit 251 is a unit configured to subject the left image 11 to first step image conversion. The right image pre-conversion unit 252 is a unit configured to subject the right image 12 to first step image conversion. The left image post-conversion unit 253 is a unit configured to subject the left image 11 to second step image conversion. The right image post-conversion unit 254 is a unit configured to subject the right image 12 to second step image conversion. These left image pre-conversion unit 251, right image pre-conversion unit 252, left image post-conversion unit 253, and right image post-conversion unit 254 are each configured so as to be independently controlled by the disparity control unit 240. Note that the left image pre-conversion unit 251 and right image pre-conversion unit 252 are examples of the first image conversion unit described in the Claims. Also, the left image post-conversion unit 253 and right image post-conversion unit 254 are examples of the second image conversion unit described in the Claims.
(51) As for the first step image conversion at the image conversion unit 250, shift processing for moving the left and right images is assumed to be used, which will be described later. However, instead of this shift processing, scaling processing for enlarging/reducing the entire screen in the horizontal direction with the edges of the left and right images as references may be employed. Also, as for the second step image conversion at the image conversion unit 250, scaling processing for enlarging/reducing the entire screen is assumed to be used with the centers of the left and right images as references, which will be described later. More natural stereoscopic effect can be reproduced by performing such image conversion of two steps regarding each of the left and right images.
(52) The image conversion at this image conversion unit 250 is generalized as follows. Now, let us say that the pixel of coordinates (x, y) is p(x, y), the pixels of interest on the left and right images are p(xL, yL) and p(xR, yR), and the pixels on the left and right images after disparity control are p(x′L, y′L) and p(x′R, y′R).
(53)
Here, (xLi, yLi) and (xRi, yRi) are offset terms of the left and right images, respectively (where i is 0 or 1).
(54) The shift processing for shifting the relative positions of the left and right images in the horizontal direction is defined by the following expression from the above general expression.
(55)
At this time, the disparity control parameter is shift amount s.
(56) Also, the scaling processing for enlarging/reducing the entire screen in the horizontal direction with the centers of the left and right images as references is defined by the following expression.
(57)
At this time, the disparity control parameters are a scaling ratio r, the horizontal coordinates xLC of the center of the left image, and the horizontal coordinates xRC of the center of the right image.
(58) Also, the scaling processing for enlarging/reducing the entire screen in the horizontal direction with the frame inner side positions as references is defined by the following expression.
(59)
At this time, the disparity control parameters are a scaling ratio r, the horizontal coordinates xLR of the right edge of the left image, and the horizontal coordinates xRL of the left edge of the right image.
(60) Also, the scaling processing for enlarging/reducing the entire screen in the horizontal direction with the frame outer side positions as references is defined by the following expression.
(61)
At this time, the disparity control parameters are a scaling ratio r, the horizontal coordinates xLL of the left edge of the left image, and the horizontal coordinates xRR of the right edge of the right image.
(62) Relationship between Disparity and Stereoscopic Image
(63)
d=XR−XL
That is to say, in the event that the stereoscopic image exists deeper than the display surface of the stereoscopic display, the disparity d has a positive value, and in the event that the stereoscopic image exists nearer than the display surface of the stereoscopic display, the disparity d has a negative value.
(64) Also, in this drawing,
Ld:Ls=de:(de−d)
holds. The binocular interval de is 6.5 [cm] or so, and in the usual case, the Ls is thought as 1.5 through 1.7 [m] or so, whereby these can be realized as fixed values. Thus, it can be understood that relationship between the disparity d and the stereoscopic image distance Ld is represented by the following expression, and stereoscopic effect (stereoscopic image distance Ld) changes depending on the disparity d.
Ld=Ls.Math.de/(de−d)
(65) Image Conversion by Shift Processing
(66)
(67) (b) in
(68) (c) in
(69) If we say that the pixel of the coordinates (x, y) is taken as p(x, y), and the pixels of interest on the left and right images are taken as p(xL, yL) and p(xR, yR), pixels p(x′L, y′L) and p(x′R, y′R) on the left and right images after disparity control according to the shift amount s become as follows, respectively.
p(x′L,y′L)=p(xL−s/2,yL)
p(x′R,y′R)=p(xR+s/2,yR)
(70)
(71) In the event that the shift processing has been performed with the shift amount s, the disparity is changed to d+s, and accordingly, the stereoscopic image distance L′d after shift processing is as the following expression.
L′d=Ls.Math.de/(de−(d+s))
(72) That is to say, in the event that the shift processing has been performed in a direction where the left and right images recede from each other (s>0), the stereoscopic image is perceived to have moved deeper. At this time, the stereoscopic image is perceived such that the more the shift amount increases, the deeper the stereoscopic image moves, and the deeper the stereoscopic image exists, the deeper the stereoscopic image moves. On the other hand, in the event that the shift processing has been performed in a direction where the left and right images approach to each other (s<0), the stereoscopic image is perceived to have moved nearer. At this time, the stereoscopic image is perceived such that the more the absolute value of the shift amount increases, the nearer the stereoscopic image moves, and the deeper the stereoscopic image exists, the nearer the stereoscopic image moves.
(73)
(74) Image Conversion by Scaling Processing with Frame Center Positions as References
(75)
(76) (b) in
(77) (c) in
(78) If we say that the pixel of the coordinates (x, y) is taken as p(x, y), and the pixels of interest on the left and right images are taken as p(xL, yL) and p(xR, yR), pixels p(x′L, y′L) and p(x′R, y′R) on the left and right images after disparity control by the scaling ratio r are as follows, respectively.
p(x′L,y′L)=p(r(xL−xLC)+xLC,yL)
p(x′R,y′R)=p(r(xR−xRC)+xRC,yR)
(79)
(80) In the event that the scaling processing has been performed with the scaling ratio r, the disparity changes to r×d, and accordingly, stereoscopic image distance L′d after shift processing is as in the following expression.
L′d=Ls.Math.de/(de−r.Math.d)
(81) That is to say, in the event that the reduction processing has been performed (r<1), the stereoscopic image is perceived such that the dynamic range of depth is reduced, the stereoscopic image deeper than the position of the display surface of the stereoscopic display moves nearer, and the stereoscopic image nearer than the position of the display surface of the stereoscopic display moves deeper. On the other hand, in the event that the enlargement processing has been performed (r>1), the stereoscopic image is perceived such that the dynamic range of depth is enlarged, the stereoscopic image deeper than the position of the display surface of the stereoscopic display moves deeper, and the stereoscopic image nearer than the position of the display surface of the stereoscopic display moves nearer.
(82)
(83) Comfortable Range of Stereopsis
(84)
|α−β|≦1°(=π/180)
α=β±(π/180)
hold.
(85) Here, relationship between the visual distance Ls and β is
(de/2)/Ls=tan(β/2)
β=2 tan−1(de/(2Ls)).
Also, relationship between the stereoscopic image distance Ld and α is
Ld=de/2 tan(α/2),
and accordingly, the range in the depth direction whereby viewing can comfortably be performed is as follows.
de/2 tan(αmax/2) through de/2 tan(αmin/2)
where
αmax=β+π/180
αmin=β−π/180
hold.
(86) Disparity Control by Two Step Image Conversion
(87)
(88) With this example, let us assume that the stereoscopic image exists in a rage from −80 pixel to +32 pixel as disparity. First, upon the shift processing of 25 pixels in the depth direction as the first step image conversion, the stereoscopic image moves in a range of −55 pixel to +57 pixel as disparity. In this case, the stereoscopic image is in a state somewhat protruding from the comfortable range at the depth portion. Therefore, the scaling processing of 0.95 [times] in the horizontal direction is performed as the second step image conversion. This scaling processing is performed, as described above, with the frame center positions as references. Thus, the stereoscopic image is in a range of −52 pixel to +54 pixel as disparity, whereby it can be understood that the disparity is included in the above comfortable range.
(89)
(90) With this example as well, let us say that the stereoscopic image exists in a range from −80 pixel to +32 pixel as disparity. First, in the same way as with the example in
(91) In this way, the shift processing is performed as the first step image conversion, and the scaling processing with the frame center positions as references is performed as the second step image conversion, whereby the stereoscopic image can be included in a range where comfortable viewing can be performed while suppressing side effects by the image conversions.
(92) In this case, if we say that the sift amount of the first step image conversion is s, and the scaling ratio of the second step image conversion is r1, pixels p(x″L, y″L) and p(x″R, y″R) on the left and right images in the event of applying both are as follows, respectively.
p(x″L,y″L)=p(r1(xL—s/2−xLC)+xLC,yL)
p(x″R,y″R)=p(r1(xR+s/2−xRC)+xRC,yR)
The coordinate conversions of scaling of multiple steps may be executed together by the above expressions, and also may be executed by dividing these into two steps.
(93) Processing Procedure of Disparity Conversion
(94)
(95) With the above embodiment, though description has been made assuming the shift processing as the first step image conversion, scaling processing with the frame inner side or outer side positions as references may be employed instead of the shift processing. Hereafter, the scaling processing with the frame inner side or outer side positions as references will be described.
(96) Image Conversion by Scaling Processing with Frame Inner Side Positions as References
(97)
(98) (b) in
(99) (c) in
(100) If we say that the pixel of the coordinates (x, y) is taken as p(x, y), and the pixels of interest on the left and right images are taken as p(xL, yL) and p(xR, yR), pixels p(x′L, y′L) and p(x′R, y′R) on the left and right images after disparity control according to the scaling ratio r become as follows, respectively.
p(x′L,y′L)=p(r(xL−xLR)+xLR,yL)
p(x′R,y′R)=p(r(xR−xRL)+xRL,yR)
(101)
(102) In the event that the reduction processing has been performed (r<1), the stereoscopic image is perceived such that the stereoscopic image moves nearer. At this time, the stereoscopic image is perceived such that the smaller the scaling ratio becomes, the nearer the stereoscopic image moves, and the deeper the stereoscopic image exists, the nearer the stereoscopic image moves. On the other hand, in the event that the enlargement processing has been performed (r>1), the stereoscopic image is perceived such that the stereoscopic image has moved deeper. At this time, the stereoscopic image is perceived such that the more the scaling ratio increases, the deeper the stereoscopic image moves, and the deeper the stereoscopic image exists, the deeper the stereoscopic image moves.
(103) As can be understood by comparing this drawing and
(104) Now, in order to consider a case where scaling processing with the frame inner side positions as references, and scaling processing with the frame centers as references are executed with the framework of the shift processing and the scaling processing with the frame centers as references, expression modifications are performed as follows.
x″L=r1r0(xL−(r0−1)(xLR−xLC)/r0−xLC)+xLC
x″R=r1r0(xR−(r0−1)(xRL−xRC)/r0−xRC)+xRC
That is to say, shift processing with shift amount (2(r0−1)(xLR−xLC)/r0) is performed as to the left image, which is equivalent to perform scaling processing with the frame centers of the scaling ratio r1r0 as references. Also, shift processing with shift amount (−2(r0−1)(xRL−xLC)/r0) is performed as to the right image, which is equivalent to perform scaling processing with the frame centers of the scaling ratio r1r0 as references.
(105) Image Conversion by Scaling Processing with Frame Outer Side Positions as References
(106)
(107) (b) in
(108) (c) in
(109) If we say that the pixel of the coordinates (x, y) is taken as p(x, y), and the pixels of interest on the left and right images are taken as p(xL, yL) and p(xR, yR), pixels p(x′L, y′L) and p(x′R, y′R) on the left and right images after disparity control according to the scaling ratio r become as follows, respectively.
p(x′L,y′L)=p(r(xL−xLL)+xLL,yL)
p(x′R,y′R)=p(r(xR−xRR)+xRR,yR)
(110)
(111) In the event that the reduction processing has been performed (r<1), the stereoscopic image is perceived such that the stereoscopic image moves nearer. At this time, the stereoscopic image is perceived such that the smaller the scaling ratio becomes, the nearer the stereoscopic image moves, and the deeper the stereoscopic image exists, the nearer the stereoscopic image moves. On the other hand, in the event that the enlargement processing has been performed (r>1), the stereoscopic image is perceived such that the stereoscopic image has moved deeper. At this time, the stereoscopic image is perceived such that the more the scaling ratio increases, the deeper the stereoscopic image moves, and the deeper the stereoscopic image exists, the deeper the stereoscopic image moves.
(112) As can be understood by comparing this drawing and
(113) Now, in order to consider a case where scaling processing with the frame outer side positions as references, and scaling processing with the frame centers as references are executed with the framework of the shift processing and the scaling processing with the frame centers as references, expression modifications are performed as follows.
x″L=r1r0(xL−(r0−1)(xLL−xLC)/r0−xLC)+xLC
x″R=r1r0(xR−(r0−1)(xRR−xRC)/r0−xRC)+xRC
That is to say, shift processing with shift amount (2(r0−1)(xLL−xLC)/r0) is performed as to the left image, which is equivalent to perform scaling processing with the frame centers of the scaling ratio r1r0 as references. Also, shift processing with shift amount (−2(r0−1)(xRR−xLC)/r0) is performed as to the right image, which is equivalent to perform scaling processing with the frame centers of the scaling ratio r1r0 as references.
(114)
(115) Also, scaling processing with the frame center positions as references is performed as the second step image conversion. Thus, the entire screen is reduced so that the coordinate of the pixel of interest in the horizontal direction of the left image is changed from the x′L to x″L, and the entire screen is reduced so that the coordinate of the pixel of interest in the horizontal direction of the right image is changed from the x′R to x″R. At this time, the entire screen is in a state in which there is no original image on the image frame portion, and accordingly, processing for inserting an image, such as black or white or the like, is performed. The image necessary for this insertion processing can further be reduced as compared to the case of performing shift processing as the first step image conversion.
(116) In this way, the scaling processing with the frame inner positions as references is performed as the first step image conversion, and the scaling processing with the frame center positions as references is performed as the second step image conversion, whereby image insertion such as black around the image frame can be suppressed to the minimum.
(117) If we say that the scaling ratio of the first step image conversion is r0, and the scaling ratio of the second step image conversion is r1, pixels p(x″L, y″L) and p(x″R, y″R) on the left and right images in the event of applying both are as follows, respectively.
p(x″L,y″L)=p(r1(r0(xL−xLR)+xLR−xLC)+xLC,yL)
p(x″R,y″R)=p(r1(r0(xR−xRL)+xRL−xRC)+xRC,yR)
The coordinate conversions of scaling of multiple steps may be executed together by the above expressions, and also may be executed by dividing these into two steps.
(118)
(119) Also, scaling processing with the frame center positions as references is performed as the second step image conversion. Thus, the entire screen is reduced so that the coordinate of the pixel of interest in the horizontal direction of the left image is changed from the x′L to x″L, and the entire screen is reduced so that the coordinate of the pixel of interest in the horizontal direction of the right image is changed from the x′R to x″ R. At this time, the entire screen is in a state in which there is no original image on the image frame portion, and accordingly, processing for inserting an image, such as black or white or the like, is performed. The image necessary for this insertion processing can further be reduced as compared to the case of performing shift processing as the first step image conversion.
(120) In this way, the scaling processing with the frame outer positions as references is performed as the first step image conversion, and the scaling processing with the frame center positions as references is performed as the second step image conversion, whereby image insertion such as black around the image frame can be suppressed to the minimum.
(121) If we say that the scaling ratio of the first step image conversion is r0, and the scaling ratio of the second step image conversion is r1, pixels p(x″L, y″L) and p(x″R, y″R) on the left and right images in the event of applying both are as follows, respectively.
p(x″L,y″L)=p(r1(r0(xL−xLL)+xLL−xLC)+xLC,yL)
p(x″R,y″R)=p(r1(r0(xR−xRR)+xRR−xRC)+xRC,yR)
The image conversions of scaling of multiple steps may be executed together by the above expressions, and also may be executed by dividing these into two steps.
(122) Cubic Interpolation
(123)
(124) It is assumed to obtain a pixel by taking advantage of 4×4 taps around a generated pixel P. If we say that the integer portion of the coordinates of the generated pixel P are taken as (x, y), the tap of a filter is taken as p(x+i, y+j), and a filter coefficient is taken as wi, j, the generated pixel P is represented by the following expression.
(125)
(126) Now, the filter coefficient W is obtained as follows according to distance between the generated pixel P and the tap.
(127)
(128) For example, weight W−1, −1 of a coefficient corresponding to the pixel of (x−1, y−1) are represented with
horizontal Wx=4−8(1+Δx)+5(1+Δx)2−(1+Δx)3
horizontal Wy=4−8(1+Δy)+5(1+Δy)2−(1+Δy)3
being multiplied.
W−1,−1=WxWy
Here, (Δx, Δy) becomes the decimal-point portion of the generate pixel P.
(129) Recommended Setting Range of Disparity Control Parameter
(130)
(131) First, in the event of performing shift processing as image conversion, it is recommended to set a value from −128 pixel to +128 pixel as the shift amount s. Note that in the event that the shift amount s>0, shift processing is performed in a direction where the left and right images recede to each other, and in the vent that the shift amount s<0, shift processing is performed in a direction where the left and right images approach to each other. This shift processing is assumed to be used as the first step image conversion.
(132) Also, in the event of performing scaling processing as image conversion, it is recommended to set a 0.8 to 1.20-fold value as the scaling ratio r. Note that in the event that the scaling ratio r>1, scaling processing for enlargement is performed, and in the event that the scaling ratio r<1, scaling processing for reduction is performed. Also, as for the reference position at the time of scaling processing, it is recommended to set ones of the frame center positions (
(133) Abstraction Disparity Control Parameter
(134)
(135) Now, abstraction disparity control parameters of four steps of “strong”, “middle”, “weak”, and “off” are assumed as an example of strength of disparity control. Disparity control is performed most strongly in the event of “strong”, the strength of disparity control is weakened as the strength is changed from “strong” to “weak”, and no disparity control is performed in the event of “off”. In the event of “strong”, shift processing with the shift amount s=60 pixels is performed as the first step image conversion, and scaling processing with the frame centers as references of the scaling ratio r=0.85 times is performed as the second step image conversion. In the event of “middle”, shift processing with the shift amount s=40 pixels is performed as the first step image conversion, and scaling processing with the frame centers as references of the scaling ratio r=0.90 times is performed as the second step image conversion. In the event of “weak”, it is handled that shift processing with the shift amount s=0 pixel has been performed as the first step image conversion, i.e., actual image conversion has not been performed. In this case, scaling processing with the frame inner sides of the scaling ratio r=1.03 times as references is performed as the second step image conversion. In the event of “off”, it is handled that shift processing with the shift amount s =0 pixel has been performed as the first step image conversion, i.e., actual image conversion has not been performed. In this case, it is handled that scaling processing with the scaling ratio r=1.00 times has been performed as the second step image conversion, i.e., actual image conversion has not been performed.
(136) In this way, by allowing the user to select out of the abstraction disparity control parameters of the four steps that have been preset, the disparity control parameter can be readily specified. In this case, the disparity control unit 240 generates a specific disparity control parameter based on the abstraction disparity control parameter to control the image conversion unit 250.
(137) <2. Second Embodiment>
(138) Configuration of Disparity Conversion Device
(139)
(140) The disparity estimating unit 210 is a unit configured to estimate disparity from the left image 11 and right image 12 of the input image 10 to generate a disparity map. This disparity map holds disparity for each pixel or for each pixel group of the input image 10. In this case, either the left image 11 or the right image 12 may be taken as a reference as the input image 10. An estimation method for disparity is a known technique, and a technique has been known wherein the disparity of each of the left and right images is estimated to generate a disparity map by performing matching regarding the foreground images obtained by removing the background images from the left and right images, for example (e.g., see Japanese Unexamined Patent Application Publication No. 2006-114023).
(141) The disparity analyzing unit 230 is a unit configured to analyze the disparity map estimated at the disparity estimating unit 210 to generate a disparity control parameter for performing suitable disparity control. Specifically, the disparity analyzing unit 230 generates a disparity histogram form the disparity map to determine a disparity control parameter so that the distribution of this histogram is included in a suitable range.
(142) Shift Processing and Disparity Histogram
(143)
(144) As for a calculation expression of the shift amount s in the shift processing, the following expression is available.
s[pixels]=α×intra-image disparity average [pixels]−β×intra-image disparity deflection [pixels]−γ
Here, α, β, and γ are parameter values that can arbitrarily be set. For example, with a 42-V television set, it is conceived to set these such that α=1.0 or so, β=1.5 or so, and γ=20.0 or so. An arrangement may be made wherein these parameter values are set to a default value beforehand, and are subsequently changed according to the user's preference. Note that in the event of setting that α=1.0, β=0.0, and γ=20.0, the average after the shift processing in (b) in
(145) Scaling Processing and Disparity Histogram
(146)
(147) As for a calculation expression of the scaling ratio r in the scaling processing, the following expression is available.
r[times]=ε×intra-image disparity deflection [pixels]
(148) Also, the following expression is available as another expression.
r[times]=comfortable disparity range [pixels]/intra-image disparity range [pixels]
Thus, disparity can be reduced by the scaling processing so that the disparity range within an image is included in the comfortable disparity range. Here, as for the comfortable disparity range, with the above “3DC Safety Guidelines”, around 100 pixels are defined in the standard visual distance of the 42-V television set.
(149) In this way, according to the second embodiment of the present invention, a disparity control parameter is generated based on the disparity histogram generated from the left and right images, whereby image conversion can be performed using this disparity control parameter.
(150) Disparity Control Parameter Based on Stress Value
(151)
(152) A stress value en of a pixel n is defined by the following expression with the disparity of the pixel n as xn.
en=f(xn)
Here, f(x) is a function representing relationship between the disparity and stress, and indicates relationship such as
f(x)=λ×2,
and with x conf<x,
f(x)=2λx conf(x−x conf)+λx 2conf
and with x<−x conf,
f(x)=x 4/2 x 2conf+x 2conf/2.
(153) This function indicates that the higher the stress value is as the position moves upwards. Specifically, with the nearer side than the display surface of the stereoscopic display, the nearer the position is, the higher the stress value is. On the other hand, with the deeper side than the display surface of the stereoscopic display, the stress value is low within a suitable range, but the stress value suddenly increases when the position is deeper than a fixed position. This is because the greater the stress value is when exceeding a comfortable disparity range, and the stress extremely increases when disparity is deeper than both-eye distance. Also, the stress is conceived to be smaller on the nearer side than the depth side in the event of exceeding the comfortable disparity range. Here, the comfortable disparity range is around “±50 pixel” at the standard visual distance of the 46-V type. Also, the depth position where disparity is greater than both-eye distance is equivalent to around “−100 pixel” at the standard visual distance of the 46-V type.
(154) If we say that the shift amount in shift processing is taken as x shift, and the scaling amount in scaling processing is taken as a scale, a stress value e′n after these processes is represented by the following expression.
e′n=f(a scale.Math.xn−x shift)
Therefore, according to this modification, a disparity control parameter can be determined by obtaining the shift amount x shift and scaling amount a scale whereby total amount E′ of the stress values of the entire screen becomes the minimum.
(155) Processing Procedure of Disparity Conversion
(156)
(157) <3. Third Embodiment>
(158) Configuration of Disparity Conversion Device
(159)
(160) Region-of-interest Map and Disparity Histogram
(161)
(162) The region-of-interest map indicates the position and degree of a region of interest in the input image, and indicates that the whiter the region is, the higher the influence ratio, the degree of interest is. Specifically, in (c) in
(163)
(164) On the other hand, with this third embodiment, the region-of-interest map generated by the region-of-interest estimating unit 220 is added, and accordingly, a disparity histogram with weight such as (b) in
(165) Processing Procedure of Disparity Conversion
(166)
(167) In this way, according to the third embodiment of the present invention, a disparity control parameter is generated in consideration of influence in the region of interest, whereby image conversion can be performed with this disparity control parameter.
(168) <4. Fourth Embodiment>
(169) Configuration Example of Disparity Conversion Device
(170)
(171) Existence of an object protruding around the image frame tends to give impression to the effect that the object is hard to see, to the viewer. On the other hand, an object receding around the image frame is not frequently hard to see. Accordingly, an uncomfortable region displayed on only one eye is a region where the image is input to only one eye of the viewer, and is also a region protruding nearer than the display surface of the display. Such an uncomfortable region is distinguishable using a depth map. With this fourth embodiment, let us say that an uncomfortable region is determined at the disparity analyzing unit 230 with a depth map being assumed to be generated at the disparity estimating unit 210. Based on the determination result at the disparity analyzing unit 230, the image is converted at the image frame conversion unit 280 so as to fill an uncomfortable region displayed on only one eye of the viewer.
(172) Image Frame Circumference and Disparity Control
(173)
(174) (b) in
(175) Processing Procedure of Disparity Conversion
(176)
(177) In this way, according to the fourth embodiment of the present invention, the image is converted so as to fill an uncomfortable region displayed on only one eye around the image frame, whereby displeasure at the time of stereopsis can be reduced.
(178) <5. Modifications>
(179) With the above embodiments, description has been made assuming scaling processing for performing enlargement/reduction in the horizontal direction alone, but the present invention is not restricted to this, and enlargement/reduction may also be performed in the vertical direction at the same time. In this case, control can be performed so as to maintain the aspect ratio, but circumstances in the horizontal direction such as overflow of an image at the time of enlargement, image insertion processing of a image frame portion at the time of reduction, and so forth may be caused in the same way as with the vertical direction.
(180) Also, with the above embodiments, description has been made regarding convergence to a comfortable range of a stereoscopic image by the image conversion of two steps serving as the first embodiment, and generation of a disparity control parameter based on a disparity histogram serving as the second embodiment. Further, description has been made regarding correspondence to a region of interest serving as the third embodiment, and correspondence to around an image frame serving as the fourth embodiment. These may be carried out by being combined as appropriate. For example, after performing correspondence to a region of interest according to the third embodiment, correspondence to around an image frame according to the fourth embodiment may be performed.
(181) Note that the embodiments of the present invention illustrate an example for realizing the present invention, and as apparently illustrated in the embodiments of the present invention, the matters in the embodiments of the present invention, and the invention specification matters in the Claims have a corresponding relation, respectively. Similarly, the invention specification matters in the Claims, and matters in the embodiments of the present invention denoted with the same name have a corresponding relation, respectively. However, the present invention is not restricted to the embodiments, and can be realized by subjecting the embodiments to various modifications without departing from the essence of the present invention.
(182) Also, the processing procedures described in the embodiments of the present invention may be taken as a method having these series of procedures, or may be taken as a program causing a computer to execute these series of procedures through a recording medium for recording the program thereof. As for this recording medium, for example, CD (Compact Disc), MD (MiniDisc), DVD (Digital Versatile Disk), memory cards, Blu-ray Disc (registered trademark), or the like may be employed.
Reference Signs List
(183) 100 image storage device
(184) 200 disparity conversion device
(185) 210 disparity estimating unit
(186) 220 region-of-interest estimating unit
(187) 230 disparity analyzing unit
(188) 240 disparity control unit
(189) 250 image conversion unit
(190) 251 left image pre-conversion unit
(191) 252 right image pre-conversion unit
(192) 253 left image post-conversion unit
(193) 254 right image post-conversion unit
(194) 280 image frame conversion unit
(195) 290 operation acceptance unit
(196) 300 display control device
(197) 400 image display device