Image processing device, image display device, and program
11240407 · 2022-02-01
Assignee
Inventors
Cpc classification
H04N23/81
ELECTRICITY
H04N5/66
ELECTRICITY
H04N23/683
ELECTRICITY
International classification
Abstract
Provided is an image processing device, image display device, and program that allow an image captured by an imaging device not having a vibration suppression function or an image whose vibration has been suppressed incompletely to be displayed on a display device with the vibration suppressed. The image processing device includes a motion estimator configured to estimate the amount of motion of an object between a first image and a second image later than the first image and a motion compensator configured to perform a conversion process on the second image so that vibration of the object between the first image and the second image is suppressed, on the basis of the amount of motion of the object. The motion estimator has a first estimation mode in which the amount of motion of the object is estimated in a predetermined search area and a second estimation mode in which the amount of motion of the object is estimated in a larger area than the search area in the first estimation mode.
Claims
1. An image processing device comprising: a motion estimating circuit configured to estimate the amount of motion of an object between a first image and a second image later than the first image; a motion compensating circuit configured to perform a conversion process on the second image so that vibration of the object between the first image and the second image is suppressed, on the basis of the amount of motion of the object; and a mode switching circuit configured to switch between a first estimation mode and a second estimation mode, wherein the motion estimating circuit has the first estimation mode in which the amount of motion of the object is estimated in a predetermined search area and the second estimation mode in which the amount of motion of the object is estimated in a larger area than the predetermined search area in the first estimation mode, when a predetermined criterion is satisfied, the motion compensating circuit performs the conversion process on the basis of the amount of motion estimated in the second estimation mode, and the predetermined search area of the first estimation mode is in the first image and is smaller than the first image, the larger area of the second estimation mode is in the first image and includes the predetermined search area of the first estimation mode, and when the predetermined criterion is not satisfied, the motion compensating circuit performs the conversion process on the basis of the amount of motion estimated in the first estimation mode, in the first estimation mode, the motion estimating circuit defines, as an estimation error, a value calculated on the basis of a difference in luminance between the object included in the first image and a predetermined area included in the second image, and the predetermined criterion is that the estimation error exceeds a predetermined value, or that the estimation error exceeds a predetermined multiple of an average value, or both.
2. The image processing device of claim 1, wherein the second estimation mode is a global motion estimation mode in which the amount of motion of the object is estimated from the entire first and second images.
3. The image processing device of claim 1, wherein in the second estimation mode, the amount of motion of the object is estimated in a new search area obtained by expanding a size of the search area in the first estimation mode.
4. The image processing device of claim 1, wherein the motion estimator estimates the amount of motion of the object using pattern matching or a feature vector.
5. The image processing device of claim 1, further comprising: an area setting circuit configured to set an area corresponding to the object in the first estimation mode.
6. The image processing device of claim 5, wherein the motion compensating circuit performs the conversion process on the second image so that the area set by the area setting circuit is located in an approximate center of a display surface on which the first and second images are displayed.
7. The image processing device of claim 1, wherein when the predetermined criterion is not satisfied, the mode switching circuit switches the second estimation mode to the first estimation mode.
8. An image display device comprising the image processing device of claim 1.
9. An image processing program for causing a computer to function as: a motion estimating circuit configured to estimate the amount of motion of an object between a first image and a second image later than the first image; a motion compensating circuit configured to perform a conversion process on the second image so that vibration of the object between the first image and the second image is suppressed, on the basis of the amount of motion of the object; and a mode switching circuit configured to switch between a first estimation mode and a second estimation mode, wherein the image processing program is stored in a computer-readable non-transitory storage medium, the motion estimating circuit has the first estimation mode in which the amount of motion of the object is estimated in a predetermined search area and the second estimation mode in which the amount of motion of the object is estimated in a larger area than the predetermined search area in the first estimation mode, when a predetermined criterion is satisfied, the motion compensating circuit performs the conversion process on the basis of the amount of motion estimated in the second estimation mode, and the predetermined search area of the first estimation mode is in the first image and is smaller than the first image, the larger area of the second estimation mode is in the first image and includes the predetermined search area of the first estimation mode, and when the predetermined criterion is not satisfied, the motion compensating circuit performs the conversion process on the basis of the amount of motion estimated in the first estimation mode, in the first estimation mode, the motion estimating circuit defines, as an estimation error, a value calculated on the basis of a difference in luminance between the object included in the first image and a predetermined area included in the second image, and the predetermined criterion is that the estimation error exceeds a predetermined value, or that the estimation error exceeds a predetermined multiple of an average value, or both.
10. An image processing device comprising: a motion estimating circuit configured to estimate the amount of motion of an object between a first image and a second image later than the first image; and a motion compensating circuit configured to perform a conversion process on the second image so that vibration of the object between the first image and the second image is suppressed, on the basis of the amount of motion of the object, wherein the motion estimating circuit has a first estimation mode in which the amount of motion of the object is estimated in a predetermined search area and a second estimation mode in which the amount of motion of the object is estimated in a larger area than the predetermined search area in the first estimation mode, the predetermined search area of the first estimation mode is in the first image and is smaller than the first image, the second estimation mode is a global motion estimation mode, and the larger area of the global motion estimation mode match with the entire first image, in the global motion estimation mode, the larger area than the predetermined search area is divided into a plurality of divided areas, and movement direction is estimated for each divided area, and the amount of motion of the object is estimated from the movement directions of the divided areas, when a predetermined criterion is satisfied, the motion compensating circuit performs the conversion process on the basis of the amount of motion estimated in the second estimation mode, and when the predetermined criterion is not satisfied, the motion compensating circuit performs the conversion process on the basis of the amount of motion estimated in the first estimation mode.
Description
BRIEF DESCRIPTION OF DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
DESCRIPTION OF EMBODIMENTS
(14) Now, embodiments of the present invention will be described with reference to the drawings. Various features described in the embodiments below can be combined with each other.
First Embodiment
(15) Referring now to
Vibration of Object in Image
(16) As shown in
(17) For example, an image shown in
(18) To suppress such vibration, the image processing device 1 according to the first embodiment is used. Thus, the image is displayed on an image display device with the vibration suppressed.
Image Processing Device 1
(19) Referring now to
(20) The first embodiment will be described using consecutive two images, an image 1 and an image 2, as examples of the first image and the second image later than the first image.
(21) The image processing device 1 includes the motion estimator 2, an area setter 3, a tracking failure detector 5, a mode switch 7, and the motion compensator 8. The motion estimator 2 includes a local area tracking unit 4 and a global motion estimator 6. The local area tracking unit 4 includes a first motion estimator 41 and an image cutter 42. The global motion estimator 6 includes a second motion estimator 61 and a motion corrector 62. These functions may be implemented by software or hardware. When implementing the functions by software, a CPU may be caused to execute a program. The program may be stored in built-in memory or a computer-readable non-transitory storage medium. The program may also be stored in external memory and then read so that the functions are implemented by so-called cloud computing. When implementing the functions by hardware, ASIC, FPGA, DRP, or the like may be used.
(22) The area setter 3 sets an area corresponding to an object in a local motion estimation mode, which is an example of a first estimation mode. In the first embodiment, the area corresponding to the object is referred to as a local area LA (to be discussed later with reference to
(23) The motion estimator 2 estimates the amount of motion of the object between the first image and the later second image. The first and second images may be consecutive images included in multiple images or may be two non-consecutive images.
(24) The local area tracking unit 4 tracks the local area LA in the local motion estimation mode. The local motion estimation mode is a mode in which the amount of motion of the object is estimated in the predetermined search area SA. This mode can save the calculation resources compared to when estimating the amount of motion of the object using the entire first and second images. Also, the local motion estimation mode is a mode in which the object can be tracked in the search area SA, and therefore allows for accurate estimation of the motion of the object compared to when estimating the amount of motion of the object using the entire first and second images.
(25) The image cutter 42 cuts out parts of the first and second images and outputs the cut-out image parts to the first motion estimator 41. The image cutter 42 determines the range of the part of the image 1 to be cut out, on the basis of coordinates received from the area setter 3.
(26) The first motion estimator 41 estimates the motion of the object in the search areas SA of the first and second images using the image parts received from the image cutter 42. In the first embodiment, the amount of motion of the object estimated by the first motion estimator 41 is referred to as an amount of tracked motion. The first motion estimator 41 then outputs the amount of tracked motion to the motion compensator 8. The first motion estimator 41 also outputs, to the tracking failure detector 5, an estimation error (to be discussed later with reference to
(27) The global motion estimator 6 estimates the motion of the object in a global motion estimation mode, which is an example of a second estimation mode. The global motion estimation mode is a mode in which the amount of motion of the object is estimated from the entire first and second images. For this reason, the global motion estimation mode requires a larger amount of calculation than the local motion estimation mode, while it allows for estimation of the motion of the object that cannot be coped with in the local motion estimation mode. In the global motion estimation mode, the term “object” refers to the entire image.
(28) The second motion estimator 61 estimates the amount of motion of the object using the entire first and later second images. In the first embodiment, the amount of motion of the object estimated by the second motion estimator 61 is referred to as an amount of global motion. The second motion estimator 61 then outputs the amount of global motion to the motion corrector 62.
(29) The motion corrector 62 calculates an amount of correction used when performing a conversion process on the second image, on the basis of the amount of global motion received from the second motion estimator 61. Such correction will be described later with reference to
(30) The tracking failure detector 5 detects that the first motion estimator 41 has failed to track the local area LA.
(31) The mode switch 7 switches between the local motion estimation mode and the global motion estimation mode, which is an example of the second estimation mode. When switching between the local motion estimation mode and global motion estimation mode, the mode switch 7 outputs a mode switch signal to the local area tracking unit 4 and global motion estimator 6. Thus, the mode of the image processing device 1 is switched.
(32) The motion compensator 8 performs a conversion process on the second image on the basis of the amount of motion of the object to suppress the vibration of the object between the first and second images. When the predetermined criterion is satisfied, the motion compensator 8 performs the conversion process on the basis of the amount of motion estimated in the second estimation mode; when the predetermined criterion is not satisfied, it performs the conversion process on the basis of the amount of motion estimated in the first estimation mode. Thus, the input image 2 is converted into an image 2′, which is then outputted as an output image. As a result, the vibration of the object between the first and second images is suppressed.
Flowchart
(33) Next, referring to
(34) S1: Set Local Area
(35) First, in S1, the area setter 3 sets a local area LA. The image processing device 1 is previously placed in the local motion estimation mode.
(36) The setting of the local area LA by the area setter 3 will be described with reference to
(37) The area setter 3 then outputs the coordinates indicating the local area LA to the image cutter 42 shown in
(38) S2: Track Local Area
(39) Then, in S2, the local area tracking unit 4 tracks the local area LA in the search area SA.
(40) A technique by which the local area tracking unit 4 tracks the local area LA will be described with reference to
(41) The local area tracking unit 4 estimates the local area LA from the image 2 by moving a fitting area FA in the search area SA as shown in
(42) Specifically, the image cutter 42 shown in
(43) As shown in
(44) The first motion estimator 41 sequentially performs this calculation while shifting the fitting area FA in the search area SA on a pixel-by-pixel basis. Specifically, in the first embodiment, the first motion estimator 41 first sets the fitting area FA on the upper-left edge of the search area SA and then calculates Σ|Y.sub.0-Y.sub.1| while shifting the fitting area FA on a pixel-by-pixel basis in the right direction of the figure. When the fitting area FA reaches the upper-right edge of the search area SA, the first motion estimator 41 returns the fitting area FA to the initial position and then shifts it by one pixel in the downward direction of the figure, that is, sets it in a lower position than the initial position by one line. The first motion estimator 41 then repeatedly performs this calculation until the fitting area FA reaches the lower-right edge of the search area SA.
(45) The first motion estimator 41 then makes a comparison among Σ|Y.sub.0-Y.sub.1| of all positions and identifies the coordinates of a fitting area FA having the smallest value. The first motion estimator 41 then estimates an area represented by the coordinates, as an area corresponding to the local area LA included in the input image 1. This is because the area having the smallest luminance difference is most likely to correspond to the local area LA included in the input image 1. Note that when the local area LA does not change between the input images 1 and 2 at all, the estimation error is 0.
(46) In this way, the local area tracking unit 4 tracks the movement of the local area LA.
(47) S3: Detect Failure Flag
(48) Then, in S3, the tracking failure detector 5 determines whether a failure flag has been detected. A failure flag is detected on the basis of the predetermined criterion.
(49) For example, when the local area LA makes a large movement between the two images or when the luminance of the local area LA changes significantly therebetween, the first motion estimator 41 may have difficulty in tracking the local area LA. For this reason, in the first embodiment, the predetermined criterion is previously set as a criterion indicating that tracking is difficult, and when this criterion is satisfied (that is, when the first motion estimator 41 has difficulty in tracking the local area LA), the tracking failure detector 5 detects a failure flag and generates a failure flag signal indicating that tracking is difficult.
(50) The predetermined criterion will be described below with reference to
(51) In the first embodiment, the predetermined criterion is that “the estimation error exceeds a predetermined value” or that “the estimation error exceeds a predetermined multiple of the average value.” More specifically, the predetermined criterion is that “the estimation error exceeds the predetermined value and exceeds the predetermined multiple of the average value.” As used herein, the term “predetermined value” refers to a threshold (absolute value) and tracking of the local area LA will become difficult when the estimation error exceeds this value. The predetermined value may be any value. In the first embodiment, the threshold is, for example, 25,000. The term “average value” refers to the average of the amounts of motion of multiple images arranged in chronological order. In
(52) Specifically, in the first embodiment, when the estimation error exceeds 25,000 and the estimation error exceeds 15,000, it is determined that tracking of the local area LA is difficult. In the example of
(53) When the determination result in S3 is “No”, that is, when no failure flag has been detected, the process proceeds to S4.
(54) S4: Output Amount of Motion
(55) Then, in S4, as shown in
(56) S5: Motion Compensation
(57) Then, in S5, the motion compensator 8 performs a compensation process.
(58) The processes in S4 and S5, that is, the local motion estimation mode will be described below with reference to
(59) Local Motion Estimation Mode
(60)
(61) First, the area setter 3 generates the coordinates of the local area LA by performing the process of S1 in
(62) Since the input image 1 is only one input image, tracking of the local area LA is difficult. Accordingly, the image 1 is outputted as an output image as it is.
(63) Then, when the image 2, which is the input image of the subsequent frame, is inputted, data indicating the image 2 is outputted to the local area tracking unit 4 along with data indicating the image 1. Then, the local area tracking unit 4 estimates the amount of motion between the image 1 and image 2 (the amount of motion 1) and outputs the amount of motion 1 and the data indicating the image 2 to the motion compensator 8. The motion compensator 8 generates an image 2′ by performing a conversion process on the image 2 on the basis of the amount of motion 1 and outputs the image 2′ as an output signal subsequent to the image 1.
(64) The conversion process based on the amount of motion will be described below with reference to
(65) In the first embodiment, the motion compensator 8 compensates for the position of the image so that the local area LA is displayed in the approximate center of the display surface on which the image is displayed. Specifically, the motion compensator 8 moves the image 1 on the basis of the coordinates set by the area setter 3 and the coordinates of the approximate center of the display surface. In the first embodiment, to simplify the description, a case in which the coordinates of the approximate center of the display surface are previously inputted to the area setter 3 will be described. That is, in
(66) As shown in
(67) Next, as shown in
(68) To suppress the vibration of the local area LA, the motion compensator 8 moves the image 2 along with D2 by the amount of motion 1 (Δx1,Δy1). In the first embodiment, the right direction of
(69) If it is not desired to generate an area where no image is displayed as seen in
(70) Referring back to
(71) The local area tracking unit 4 calculates the amount of motion 2 between the image 2 and image 3. In the first embodiment, as shown in
(72) As shown in
(73) Then, when an image 4 is inputted, a process similar to those with respect to the images 2 and 3 is repeated. This process is performed each time one image is inputted until the input image of the last frame is inputted.
(74) In the local motion estimation mode, the subsequent image is converted in accordance with P0 (x0,y0). For this reason, instead of calculating the amount of motion 2 between the image 2 and image 3, the amount of motion between the images 1 and image 3 may be directly calculated by comparing the image 1 and image 3. However, the local area LA is tracked with higher accuracy by comparing the image 2 and image 3 than by comparing the image 1 and image 3. The reason is that the luminance in the local area LA included in actual images 1, 2, 3, and the like changes from moment to moment and a smaller estimation error (Σ|Y.sub.0-Y.sub.1|) is obtained by comparing with an image as recent as possible.
(75) As seen above, in the local motion estimation mode, the local area LA can be tracked in the search area SA using pattern matching.
(76) Referring back to
(77) When the determination result in S3 is “Yes”, that is, a failure flag is detected, the tracking failure detector 5 outputs a failure flag signal to the mode switch 7 and motion compensator 8, and the process proceeds to S6.
(78) In S6, the mode switch 7 outputs a mode switch signal to the local area tracking unit 4 and global motion estimator 6 and switches the image processing device 1 from the local motion estimation mode to the global motion estimation mode. Then, the amount of motion calculated in the global motion estimation mode is outputted to the motion compensator 8, and the process proceeds to S4 and S5 as in the local motion estimation mode.
(79) Referring now to
(80) S6: Global Motion Estimation Mode
(81) The global motion estimation mode is a mode in which the amount of motion of the object is estimated from the entire first and second images. In the global motion estimation mode, the coordinates of the local area LA are not set unlike in the local motion estimation mode.
(82) As shown in
(83) The second motion estimator 61 estimates the movement direction of each divided area using pattern matching or a feature vector. As used herein, the term “movement direction of each divided area” refers to the direction in which an object included in each divided area moves. While any estimation method may be used, a process similar to that with respect to the local area LA, for example, may be performed on each divided area. Specifically, the movement direction of each divided area can be estimated by setting the center of each divided area as a local area LA, setting the outer circumference of the divided area as a search area SA, and performing a process similar to tracking of the local area LA on each of the nine divided areas. The second motion estimator 61 shown in
(84) The significance of the motion corrector 62 will be described below with reference to
(85)
(86) As shown in
(87) As shown in
(88) Next, as shown in
(89) In the global motion estimation mode, the coordinates of the local area LA are not set. For this reason, even if the motion compensator 8 moves the image 3 by the amount of global motion between the image 2 and image 3 in a conversion process, the image 3 is returned to the state of the image 2 shown in
(90) The motion compensator 8 moves the image 3 along with D3 by the corrected amount of motion 2′ (Δx1′,Δy1′). In
(91) Next, as shown in
(92) The motion compensator 8 moves the image 4 along with D4 by the corrected amount of motion 3′ (Δx3′,Δy3′). This process is performed by adding the corrected amount of motion 3′ to the coordinates of the entire image 4 and rendering the entire image 4. Thus, the motion compensator 8 generates an image 4′ and outputs the image 4′ as an output image subsequent to the image 3′. Note that (Δx3′,Δy3′) corresponds to the corrected amount of global motion in
(93) This process is performed each time one image is inputted until the input image of the last frame is inputted.
(94) As seen above, in the global motion estimation mode, the amount of motion of the object is estimated from the entire image using pattern matching.
(95) Referring back to
(96) S7: Determine Whether Current Image is Image of Last Frame
(97) After the motion compensator 8 completes the compensation process, it is determined in S7 whether the current image is the image of the last frame. When the determination result in S7 is “No”, that is, the current image is not the image of the last frame, the processes of S2 to S5 are repeatedly performed. Even if the mode is switched to the global motion estimation mode in S6, the process returns to S2. As described above, this is because it is necessary to determine whether tracking of the local area LA has been allowed again. In this case, it can be said that the process of the local motion estimation mode is being performed in the background, although the local motion estimation mode has been switched to the global motion estimation mode. For this reason, when the predetermined criterion is not satisfied, the second estimation mode can be switched to the first estimation mode.
(98) On the other hand, when the determination result in S7 is “Yes”, that is, the current image is the image of the last frame, the image processing device 1 ends the vibration suppression process.
(99) As described above, the image processing device 1 according to the first embodiment determines in which of the local motion estimation mode and global motion estimation mode the motion compensator 8 performs the conversion process, on the basis of the predetermined criterion. Thus, when tracking of the local area LA is allowed, the image processing device 1 is able to perform image processing in the local motion estimation mode, in which calculation is small in amount. On the other hand, when tracking of the local area LA is difficult, the local motion estimation mode is switched to the global motion estimation mode, in which the entire image is used. Thus, the image processing device 1 is able to efficiently suppress the vibration of the image while saving the calculation resources.
Second Embodiment
(100) Next, referring to
(101) The image processing device 1 according to the second embodiment includes a first motion estimator 41 that estimates the amount of motion of a local area LA in a predetermined search area SA and a motion compensator 8 that performs a conversion process on a second image so that the vibration of the local area LA between a first image and the second image is suppressed, on the basis of the amount of motion of the local area LA. The first motion estimator 41 has a first estimation mode in which the amount of motion of an image is estimated by tracking the local area LA in the search area SA and a second estimation mode in which the local area LA is tracked on the basis of a larger area than the search area SA in the first estimation mode.
(102) As shown in
(103) A comparison is made between an input image shown in
(104) For this reason, the local area tracking unit 4 newly sets a search area SA1 by expanding the search area SA0 using an area expander (not shown). This state is shown in
(105) Then, the local area LA is tracked by again performing the process shown in
(106) In the second embodiment, the search area SA0 is expanded, and the first estimation mode is switched to the second estimation mode. Thus, the local area LA which has been difficult to be tracked before being expanded can be tracked. If necessary, the search area SA may be expanded predetermined times repeatedly. If the size of the expanded search area SA becomes equal to the size of the entire image, the local area LA can be tracked accurately as long as the local area LA is located in the image, although calculation is large in amount.
(107) Next, referring to
(108) Then, in S3, the tracking failure detector 5 determines whether a failure flag has been detected. When the determination result in S3 is “No”, that is, when no failure flag has been detected, the process proceeds to S5.
(109) Then, in S5, the local area tracking unit 4 outputs the amount of motion to the motion compensator 8. Then, in S6, the motion compensator 8 performs a conversion process. On the other hand, when the determination result in S3 is “Yes”, that is, a failure flag has been detected, the process proceeds to S6 to determine whether the number of times of expansion of the search area SA has reached the upper limit. Note that the upper limit number of times may be any number of times. Hereinafter, an example in which the upper limit number of times is three will be described.
(110) In the initial state, the number of times of expansion is zero and does not exceed the upper limit. Therefore, the determination result in S6 becomes “No”. Then, the process proceeds to S7, and the local area tracking unit 4 sets a search area SA1 by expanding the search area SA0. Then, the process again proceeds to S3 through S2. When tracking of the local area LA is allowed due to the setting of the search area SA1, the process proceeds to S4. On the other hand, when tracking of the local area LA is still difficult although the search area SA1 has been set, that is, when the predetermined criterion shown in
(111) When the number of times of expansion exceeds the upper limit in S6, specifically, when tracking of the local area LA is difficult although a search area SA3 has been set, the determination result in S6 becomes “Yes”, and the process proceeds to S8. That is, when tracking of the local area LA is difficult although the search area SA has been expanded until the number of times of expansion exceeds the upper limit, the image is displayed as it is without performing a compensation process. In the present embodiment, the upper limit number of times of expansion of the search area SA is set to three. However, if the size of the search area SA3 that has been expanded three times is set so as to be equal to the size of the entire image, the process proceeds to S4 after the upper limit number of times is exceeded in S6, and a compensation process can be performed.
(112) Then, the process proceeds to S8, and a process similar to S7 in
(113) When the determination result in S8 is “Yes”, that is, the current image is the image of the last frame, the image processing device 1 ends the vibration suppression process.
(114) As described above, the local area LA can be accurately tracked by expanding the search area SA using the above configuration and process. Thus, it is possible to efficiently suppress the vibration of the image while saving the calculation resources.
(115) When the predetermined criterion is no longer satisfied in the second estimation mode, for example, when the local area LA departs from the search area SA2 and is located in the immediately preceding search area SA1, tracking of the local area LA may be allowed in the search area SA1. In this case, the mode switch 7 may reduce the size of the search area SA by one level so that the local area LA is tracked. Also, when the local area LA is located in the initial search area SA, the second estimation mode may be switched to the first estimation mode.
Third Embodiment
(116) Next, referring to
(117) In the third embodiment, when the upper limit number of times is exceeded in S6, specifically, when tracking of the local area LA is difficult although a search area SA3 has been set, the determination result in S6 becomes “Yes”, and the process proceeds to S8. In S8, the mode is switched to the global motion estimation mode, as in the S6 of
(118) Then, the process proceeds to S5 and S9 through S4, and a process similar to S7 of
(119) On the other hand, when the determination result in S9 is “Yes”, that is, the current image is the image of the last frame, the image processing device 1 ends the vibration suppression process.
(120) As described above, the image processing device 1 according to the third embodiment is able to accurately track the local area LA by expanding the search area SA. Even when the number of times of expansion of the search area SA exceeds the upper limit, the motion compensator 8 can continue the conversion process by switching the mode to the global motion estimation mode.
Others
(121) While the various embodiments have been described, the image processing device 1 according to the present invention is not limited thereto. For example, the criterion for switching between the first estimation mode and the second estimation mode may be a single criterion that “the estimation error exceeds the predetermined value.” Alternately, the criterion may be a single criterion that “the estimation error exceeds the predetermined multiple of the average value.” Also, the threshold shown in
(122) The mode switch 7 may be omitted. In this case, the local motion estimation mode and global motion estimation mode may be performed simultaneously, and the motion compensator 8 may make a selection on whether it should perform the conversion process on the basis of the amount of motion estimated in the local motion estimation mode or on the basis of the amount of motion estimated in the global motion estimation mode.
(123) When estimating the amount of motion in the local motion estimation mode and global motion estimation mode, a feature vector may be used in place of pattern matching.
(124) The image processing apparatus 1 may be provided as a PC, a server or computer, or a set-top box connected to a display device. The image processing device 1 may be incorporated into a display device. The image processing device 1 may also be provided as a computer-readable non-transitory storage medium having the functions of the image processing device 1 implemented thereon. A program for implementing the functions of the image processing apparatus 1 may be distributed through the internet or the like.
(125) The elements of the image processing device 1 may be contained in the same case, or may be disposed in multiple cases in a distributed manner.
DESCRIPTION OF REFERENCE SIGNS
(126) 1: image processing device 2: motion estimator 3: area setter 4: local area tracking unit 41: first motion estimator 42: image cutter 5: tracking failure detector 6: global motion estimator 61: second motion estimator 62: motion corrector 7: mode switch 8: motion compensator 9: image display device LA: local area SA: search area FA: fitting area