Acquiring and displaying images in real-time

09787918 · 2017-10-10

Assignee

Inventors

Cpc classification

International classification

Abstract

An imaging device (100) for acquiring and displaying images in real-time, the imaging device comprising i) an imaging sensor (110) comprising a radiation sensitive array (120) for acquiring an image (142), ii) a readout circuit (140) connected to the radiation sensitive array for reading out the image, iii) a signal processor (160) for processing the image for obtaining a processed image (162), and iv) a display (180) for displaying the processed image, the radiation sensitive array being arranged in rows of sensor pixels and the display being arranged in rows of display pixels, and wherein the readout circuit is a rolling shutter circuit for sequentially reading out the rows of sensor pixels for sequentially providing subsets of pixels, the signal processor is configured for, on availability of one of the subsets of pixels, processing the subset of pixels for providing a processed subset of pixels, and the display is configured for, on availability of the processed subset of pixels, displaying the processed subset of pixels on a thereto corresponding subset of display pixels for displaying the processed image sequentially on the rows of display pixels.

Claims

1. A night vision device arranged for acquiring and displaying images in real-time, the night vision device comprising: i) an imaging sensor comprising a radiation sensitive array for acquiring an image, ii) a readout circuit connected to the radiation sensitive array for reading out the image, iii) a signal processor for processing the image for obtaining a processed image, and iv) a display for displaying the processed image, the radiation sensitive array being arranged in rows of sensor pixels and the display being arranged in rows of display pixels, and wherein: the readout circuit is a rolling shutter circuit for sequentially reading out the rows of sensor pixels by a) initializing an exposure of a row, and b) reading out its contents after said exposure, for sequentially providing subsets of pixels), each subset of pixels corresponding to one row or a subset of the rows of sensor pixels or a subset of pixels from a single row; the signal processor is configured for, on availability of one of the subsets of pixels, processing the subset of pixels for providing a processed subset of pixels; and the display is configured for, on availability of the processed subset of pixels, displaying the processed subset of pixels on a thereto corresponding subset of display pixels for displaying the processed image sequentially on the rows of display pixels, and wherein the direct view system comprises a further imaging sensor and a further readout circuit, the further imaging sensor comprising a further radiation sensitive array for acquiring a further image, the further readout circuit being connected to the further radiation sensitive array for reading out the further image, the further radiation sensitive array being arranged in rows of further sensor pixels, the further readout circuit being a further rolling shutter circuit for sequentially reading out the rows of further sensor pixels for sequentially providing further subsets of pixels, and wherein the direct view system is configured for synchronously displaying the image and the further image on the display by: (1) the rolling shutter circuit and the further rolling shutter circuit being configured for synchronously providing the subset of pixels and one of the further subsets of pixels by substantially synchronously reading out corresponding portions of the image and the further image; and (2) the signal processor being configured for, on availability of the subset of pixels and the further subset of pixels, combining the subset of pixels with the further subset of pixels for obtaining the processed subset of pixels, wherein the radiation sensitive array has a first spatial resolution, the further radiation sensitive array has a second spatial resolution, the second spatial resolution being lower than the first spatial resolution and the further rolling shutter circuit being configured for reading out the further image with a second readout speed that is lower than a first readout speed of the rolling shutter circuit for enabling said synchronously providing the subset of pixels and the further subset of pixels, and wherein the rolling shutter circuit is clocked at a first pixel clock for providing the first readout speed and the further rolling shutter circuit is clocked at a second pixel clock for providing the second readout speed.

2. The night vision device according to claim 1, wherein the imaging sensor is a visible light imaging sensor for sensing visible light and the further imaging sensor is a thermal imaging sensor for sensing infrared radiation for enabling synchronously displaying a visible light image and a thermal image on the display.

3. The night vision device according to claim 1, wherein the signal processor is configured for combining the subset of pixels with the further subset of pixels by fusing the subset of pixels with the further subset of pixels for obtaining as the processed image an image fusion of the image with the further image.

4. The night vision device according to claim 1, wherein the rolling shutter circuit is configured for reading out the image with the first readout speed within an imaging frame time (Ti), and the further rolling shutter circuit is configured for reading out the further image with the second readout speed within the imaging frame time.

5. The night vision device according to claim 1, wherein the night vision device comprises a scaler for spatially scaling the further subset of pixels for providing as the further image a scaled image having the first spatial resolution.

6. The night vision device according to claim 5, wherein the scaler is configured for performing the spatial scaling using at least one technique out of the group of: pixel repetition, first order linear interpolation, higher order linear interpolation and non-linear interpolation techniques.

7. The night vision device according to claim 1, wherein the signal processor comprises an image processing pipeline for obtaining a pipelined processing of the subsets of pixels.

8. The night vision device according to claim 1, wherein the rolling shutter circuit is configured for reading out the image with a first readout speed, and wherein the night vision device is configured for establishing the first readout speed in dependence on an amount of radiation impinging on the radiation sensitive array.

9. The night vision device according to claim 1, wherein the night vision device comprises an image intensifier for providing intensified visible light to the imaging sensor.

10. A helmet, head mount, rifle sight or handheld device comprising the night vision device according to claim 1.

11. A method of acquiring and displaying images in real-time with a night vision device, the night vision device comprising i) an imaging sensor comprising a radiation sensitive array for acquiring an image, ii) a readout circuit connected to the radiation sensitive array for reading out the image, iii) a signal processor for processing the image for obtaining a processed image, and iv) a display for displaying the processed image, the radiation sensitive array being arranged in rows of sensor pixels and the display being arranged in rows of display pixels, and wherein the method comprises: sequentially reading out the rows of sensor pixels with the readout circuit by (a) initializing an exposure of a row, and (b) reading out its contents after said exposure, for sequentially providing subsets of pixels, each subset of pixels corresponding to one row or a subset of the rows of sensor pixels or a subset of pixels from a single row; on availability of one of the subsets of pixels, processing the subset of pixels with the signal processor for providing a processed subset of pixels; and on availability of the processed subset of pixels, displaying the processed subset of pixels with the display on a thereto corresponding subset of display pixels for displaying the processed image sequentially on the rows of display pixels, wherein the night vision device comprises a further imaging sensor and a further readout circuit, the further imaging sensor comprising a further radiation sensitive array for acquiring a further image, the further readout circuit being connected to the further radiation sensitive array for reading out the further image, the further radiation sensitive array being arranged in rows of further sensor pixels, the further readout circuit being a further rolling shutter circuit for sequentially reading out the rows of further sensor pixels for sequentially providing further subsets of pixels, and wherein the method comprises synchronously displaying the image and the further image on the display by: synchronously providing the subset of pixels and one of the further subsets of pixels by substantially synchronously reading out corresponding portions of the image and the further image; and on availability of the subset of pixels and the further subset of pixels, combining the subset of pixels with the further subset of pixels for obtaining the processed subset of pixels, wherein the radiation sensitive array has a first spatial resolution, the further radiation sensitive array has a second spatial resolution, the second spatial resolution being lower than the first spatial resolution and the method further comprising reading out the further image with a second readout speed that is lower than a first readout speed of the rolling shutter circuit for enabling said synchronously providing the subset of pixels and the further subset of pixels, wherein method comprises clocking the rolling shutter circuit at a first pixel clock for providing the first readout speed and clocking the further rolling shutter circuit at a second pixel clock for providing the second readout speed.

12. A computer program stored on a non-transitory computer-readable medium, the computer program comprising instructions for causing a processor system to perform the method according to claim 11.

Description

BRIEF DESCRIPTION OF THE DRAWINGS

(1) These and other aspects of the invention are apparent from and will be elucidated with reference to the embodiments described hereinafter. In the drawings,

(2) FIG. 1 shows a timing diagram of a direct view system;

(3) FIG. 2 shows an imaging device comprising an imaging sensor;

(4) FIG. 3 shows the imaging sensor and a display;

(5) FIG. 4 shows a timing diagram of the imaging device;

(6) FIG. 5 shows an alternate representation of the timing diagram of FIG. 4;

(7) FIG. 6 shows an imaging device comprising a further imaging sensor;

(8) FIG. 7 shows an imaging sensor and the further imaging sensor;

(9) FIG. 8 shows a timing diagram of the imaging device;

(10) FIG. 9 shows an imaging device comprising a scaler;

(11) FIG. 10 shows a schematic functioning of a readout circuit;

(12) FIG. 11 shows a method for acquiring and displaying images in real-time;

(13) FIG. 12 shows a computer program stored on a computer-readable medium.

DETAILED DESCRIPTION OF EMBODIMENTS

(14) FIG. 1 shows a timing diagram of a direct view system comprising a frame buffer memory. Here, the horizontal axis is indicative of time, whereas the vertical axis is used for visually differentiating between timings of a reading of an image, a processing of the image and a displaying of the image. Here, SS.sub.1 indicates a time period for reading an image from an imaging sensor. The reading SS.sub.1 comprises storing the read image in a frame buffer memory. In such a direct view system, processing takes only place when the image is entirely stored in the frame buffer memory. Thus, after the reading SS.sub.1 has completed, the direct view system commences processing PP.sub.1 the image. Typically, portions of the image are read out from the frame buffer memory, processed, and then written back to the frame buffer memory or to a further frame buffer memory. Finally, after the processing PP.sub.1 has completed, the direct view system commences displaying DD.sub.1 the image. Consequently, a time period between a start TT.sub.1 of the reading SS.sub.1 of the image and a start TT.sub.2 of the displaying DD.sub.1 of the image indicates a minimum delay, or latency LL, that a user experiences between an change in a scene and the displayed image reflecting said change.

(15) For increasing a throughput of the direct view system, the reading, processing and displaying may be pipelined. This means that while the processing PP.sub.1 of the image takes place, a reading SS.sub.2 of a following image may take place. Similarly, while the displaying DD.sub.1 of the image takes place, a processing PP2 of the following image may take place, etc. It is noted that such pipelining increases a throughput of the direct view system, i.e., allows a system to read, process and display more images in a given time period. However, said pipelining does not affect the latency LL of the direct view system.

(16) FIG. 2 shows an imaging device 100 for acquiring and displaying images in real-time. The imaging device 100 comprises an imaging sensor 110, and the imaging sensor 110 comprises a radiation sensitive array 120 for acquiring an image 142. The imaging device 100 further comprises a readout circuit 140 connected to the radiation sensitive array 120 for reading out the image 142, and a signal processor 160 for processing the image 142 for obtaining a processed image 162. For that purpose, the readout circuit 140 is shown to be connected to the signal processor 160. The imaging device 100 further comprises a display 180 for finally displaying the processed image 162 on the display 180.

(17) FIG. 3 shows the imaging sensor 110 comprising the radiation sensitive array 120. Also shown is that the radiation sensitive array 120 is arranged in rows of sensor pixels 122. It is noted that, although not explicitly indicated in FIG. 3, the radiation sensitive array 120 is also arranged in columns of sensor pixels as a consequence of being an array. Also shown in FIG. 3 is the display 180 as a display pixel array which is arranged in rows of display pixels 182. It is noted that, although not explicitly indicated in FIG. 3, the display 180 is also arranged in columns of display pixels as a consequence of being an array.

(18) During operation of the imaging device 100, the readout circuit 140 sequentially reads out the rows of sensor pixels 122 for sequentially providing a subset of pixels 124. This reading is indicated in the timing diagrams shown in FIGS. 4 and 5. Here, S.sub.1 indicates a time period for reading the image 142 from the imaging sensor 110. FIG. 4 shows the time period in a similar manner as FIG. 1 for allowing a comparison with the aforementioned direct view system. In FIG. 5, the horizontal axis is indicative of time, whereas the vertical axis is indicative of a row number, with R.sub.n indicating a top row of the radiation sensitive array 120 and R.sub.0 indicating a bottom row. Thus, FIG. 5 shows the readout circuit 140 reading out the row R.sub.n at the beginning of the time period S.sub.1 and the row R.sub.0 at its end. Consequently, during the reading S.sub.1, all rows R.sub.n to R.sub.0 are read out sequentially.

(19) The above described reading S.sub.1 of the radiation sensitive array 120 is achieved by the readout circuit 140 being a rolling shutter circuit. The rolling shutter circuit 140 differs from a circuit configured for reading out the radiation sensitive array 120 using a snapshot shutter. The basic operating principle of a rolling shutter circuit is that the radiation sensitive array 120 is addressed in a row-by-row, i.e., on a line-by-line basis for i) initializing an exposure of a row and ii) reading out its contents after said exposure. This is typically achieved by the use of two pointers, each addressing respective rows in the radiation sensitive array 120. One pointer provides a reset of a currently addressed row for initializing an exposure of the row, whereas the other pointer addresses a row that is to be read out. The difference in location between the two pointers is the effective exposure time, i.e., if the ‘reset’ pointer trails the ‘readout’ pointer by only one row, the exposure time is maximized.

(20) In contrast, a circuit that uses a snapshot shutter typically exposes the entire radiation sensitive array simultaneously before reading out the entire image from the radiation sensitive array into a frame buffer memory. During the time needed for reading out the entire image, the radiation sensitive array is not configured for exposure anymore. Disadvantageously, the exposure time provided by said circuit is less than that of a rolling shutter circuit. A shorter exposure time typically results in an image that has a worse signal-to-noise ratio, i.e., is noisier. A publication “EBAPS: Next Generation, Low Power, Digital Night Vision”, Aebi et al., Intevac Corporation, OPTRO 2005 symposium, Paris, France, describes using a rolling shutter circuit for maximizing the exposure time of a camera sensor.

(21) The pointers of the rolling shutter circuit 140 may be increased by an internal state machine in the imaging sensor 110 itself, i.e., the rolling shutter circuit 140 may be part of the imaging sensor 110. External logic, e.g., a Field Programmable Gate Array (FPGA) located outside of the imaging sensor 110 may be used to clock the state machine and to program the distance between the two pointers for determining the exposure time.

(22) By sequentially reading out the rows of sensor pixels 122, the rolling shutter circuit 140 sequentially provides a subset of pixels 124. The subset of pixels 124 may comprise the pixels of an entire row, or of a subset of rows. The subset of pixels may also comprise a subset of pixels from a single row, e.g., a single pixel or multiple neighbouring pixels. The rolling shutter circuit 140 provides the subset of pixels 124 to the signal processor 160, which, on availability of the subset of pixels 124, processes the subset of pixels to provide a processed subset of pixels. By processing the sequentially provided subset of pixels, the signal processor 160 effectively processes the image 142 and provides a processed image 162. This processing is indicated in the timing diagrams shown in FIGS. 4 and 5. Here, P.sub.1 indicates a time period for processing the image 142. FIG. 5 shows the readout circuit 140 processing the row R.sub.n at the beginning of the time period P.sub.1 and the row R.sub.0 at its end. Consequently, during the processing P.sub.1, all rows R.sub.n to R.sub.0 are processed sequentially.

(23) The time delay between the reading S.sub.1 and the processing P.sub.1, as shown in FIGS. 4 and 5, is dependent on, amongst others, a size of the subset of pixel 124. For example, if the subset of pixels 124 comprises the pixels of a row of sensor pixels 122, the processing P.sub.1 is delayed with respect to the reading S.sub.1 by at least a time period needed for reading and providing said subset of pixels 124 to the signal processor 160. It will be appreciated, however, that said time delay is significantly less than a time delay corresponding to an entire reading out of the image 142 due to the sequential providing of the subset of pixels 124.

(24) On availability of the processed subset of pixels, the display 180 displays the processed subset of pixels on a thereto corresponding subset of display pixels 184. By displaying the sequentially provided processed subset of pixels, the display 180 effectively displays the processed image 162. This displaying is indicated in the timing diagrams shown in FIGS. 4 and 5. Here, D.sub.1 indicates a time period for displaying the processed image 162. FIG. 5 shows the display 180 displaying the row R.sub.n at the beginning of the time period D.sub.1 and the row R.sub.0 at its end. It will be appreciated that for the time delay between the displaying D.sub.1 and the processing P.sub.1, similar considerations holds as for the time delay between the reading S.sub.1 and the processing P.sub.1. The imaging device shown in FIG. 2 thus provides a latency L that corresponds to a time delay between the reading S.sub.1 and the displaying D.sub.1. Moreover, it will be appreciated that a reading S.sub.2 of a following image may commence after the reading S.sub.1 of the image has finished. Similarly, a processing P.sub.2 may commence after the processing P.sub.1 has finished, and a displaying D.sub.2 may commence after the displaying D.sub.1 has finished.

(25) FIG. 6 shows an imaging device 200. The imaging device comprises, next to the imaging sensor 110 and the readout circuit 140, also a further imaging sensor 210 and a further readout circuit 240. The further imaging sensor 210 comprises a further radiation sensitive array 220 for acquiring a further image 242. For reading out the further image 242, the further readout circuit 240 is connected to the further radiation sensitive array 220.

(26) FIG. 7 shows the further imaging sensor 210 comprising the further radiation sensitive array 220 next to the aforementioned imaging sensor 110 and radiation sensitive array 110. Also shown is that the further radiation sensitive array 220 is arranged in rows of further sensor pixels 222. It is noted that, although not explicitly indicated in FIG. 7, the further radiation sensitive array 220 is also arranged in columns of sensor pixels as a consequence of being an array. FIG. 7 also shows the radiation sensitive array 120 having a first spatial resolution, the further radiation sensitive array 220 having a second spatial resolution, with the second spatial resolution being lower than the first spatial resolution. As a consequence, the radiation sensitive array 120 is build up by n+1 rows, i.e., row R.sub.0 to row R.sub.n, whereas the radiation sensitive array 220 is build up by m+1 rows, i.e., row R.sub.0 to row R.sub.n, with m being smaller than n. This configuration is assumed in the remainder of the description of the imaging device of FIG. 6. However, it is noted that the second spatial resolution may also be equal to or larger than the first spatial resolution. Moreover, it is noted that spatial resolution may refer to an image's horizontal or vertical resolution, e.g., having 1280 pixels or 1024 lines, or to a combined resolution, e.g., a 1.3 megapixel image.

(27) Referring to FIG. 6 again, the readout circuit 140 is configured as a further rolling shutter circuit for sequentially reading out the rows of further sensor pixels 222. During operation of the imaging device 200, the rolling shutter circuit 140 and the further rolling shutter circuit 240 synchronously provide the subset of pixels 124 and the further subset of pixels 224 by substantially synchronously reading out corresponding portions of the image 142 and the further image 242. Here, corresponding portions refer to portions of the image that have an associated image contents. For example, when the imaging sensor 110 is a visible light imaging sensor for sensing visible light 112 and the further imaging sensor 210 is a thermal imaging sensor for sensing infrared radiation 212, the image 142 may be a visible light image of a scene and further image 242 may be a thermal image of the same scene. Consequently, corresponding portions may refer to, e.g., the top row R.sub.n of the image 142 corresponding to the top row R.sub.m of the further image 242, the bottom row R.sub.0 of the image 142 corresponding to a same bottom row R.sub.0 of the further image 242, etc. This enables the signal processor 260 to, on availability of the subset of pixels 124 and the further subset of pixels 224, combine both subsets of pixels to obtain the processed subset of pixels for, e.g., providing a processed image 262 in which the thermal image is overlaid on top of the visible light image. It is noted that corresponding portions may also refer to, e.g., when the imaging sensor 110 acquires a left-hand view and the further imaging sensor 210 acquires a right-hand view, portions that have a same vertical position in either image.

(28) In order to compensate for the second spatial resolution being lower than the first spatial resolution, the further rolling shutter circuit 240 is configured for reading out the further image 242 with a second readout speed that is lower than a first readout speed of the rolling shutter circuit 140. This is shown in FIG. 8, where a similar timing diagram is shown as in FIG. 5, with additionally a time period of a reading I.sub.i of the further image 242 being indicated. It will be appreciated that, with the number of rows R.sub.m of the further image 242 being lower than the number of rows R.sub.n of the image 142, the second readout speed needs to be lower to enable the reading I.sub.1 of the further image 242 within a same time interval as the reading S.sub.1 of the image 142. This is reflected in a lower slope of the reading I.sub.1 with respect to a horizontal axis when compared to the reading S.sub.1.

(29) FIG. 8 shows the first readout speed being selected for reading S.sub.1 the image 142 within an imaging frame time T.sub.i and the second readout speed being selected for reading I.sub.1 the further image 242 within the same imaging frame time T.sub.i. The imaging frame time T.sub.i is directly coupled to the imaging frame rate, e.g., is 1/60 s=0.0167 ms with a 60 Hz imaging frame rate. For maximizing an exposure of the radiation sensitive array 120, the first readout speed is selected for reading S.sub.1 the image 142 in substantially said imaging frame time T.sub.i. Furthermore, the second readout speed is selected for substantially reading I.sub.1 the further image 242 within the same imaging frame time T.sub.i. It will be appreciated that the resulting ratio between the first readout speed and the second readout speed inherently follows from the aforementioned configuration of the imaging device 200 for synchronously reading out corresponding portions of the image 142 and the further image 242.

(30) Moreover, in the example depicted in FIG. 8, the first readout speed and the second readout speed are selected for providing a reading S.sub.1 of the image 142 and a reading I.sub.1 of the further image 242 that covers the entire imaging frame time T.sub.i. This may follow out of a preference for the aforementioned maximizing of an exposure time of the radiation sensitive array 120 and/or of the further radiation sensitive array 220. However, the reading S.sub.1 and the reading I.sub.1 may also be faster, e.g., being completed before an end of the imaging frame time T.sub.i. This reduces the time difference between the reading S.sub.1 of the top and bottom part of the image 142, and thus may reduce or avoid so-termed skew artefacts in the image 142. These artefacts are known to occur when said time difference is relatively large. Also, the imaging device 200 may be configured for establishing the first readout speed in dependence on an amount of radiation 112 impinging on the radiation sensitive array 120. As such, the imaging device 200 may dynamically determine a compromise between a needed exposure time and the aforementioned skew effects.

(31) The rolling shutter circuit 140 may be clocked at a first pixel clock for providing the first readout speed and the further rolling shutter circuit 240 may be clocked at a second pixel clock for providing the second readout speed. Since the second readout speed is lower than the first readout speed, the second pixel clock is also lower than the first pixel clock. For example, when the second spatial resolution of the further image 242 is horizontally and vertically one fourth of that of the first spatial resolution of the image 142, e.g., 320 by 256 pixels with respect to 1280 by 1024 pixels, the first readout circuit 140 may be clocked at a system clock of, e.g., 44 MHz, whereas the second readout circuit 240 may be clocked at one sixteenth of that system clock, i.e., a 2.75 MHz pixel clock. Since a lower clock rate typically results in lower power consumption, the power consumption of the imaging device 200 may be reduced. Alternatively, the second readout circuit 240 may be clocked at 44 MHz as well, but may be configured to, on average, only provide one pixel every sixteenth clock cycle.

(32) Since the second spatial resolution is lower than the first spatial resolution, it may be needed to scale the further image 242 to the first spatial resolution or to a spatial resolution of the display 180. It is noted that this may not be needed in all cases, e.g., when the further image 242 is inserted as a so-termed Picture-in-Picture (PiP) into the image 142. FIG. 9 shows an imaging device 300 comprising a scaler 250 for providing as the further image 242 a scaled image 252 having the first spatial resolution. Here, the further rolling shutter circuit 240 is shown to be connected to the scaler 250 for providing the further image 242 to the scaler 250, and the scaler 250 is shown to be connected to the signal processor 260 for providing the scaled image 252 to the signal processor 260. The scaler 250 may comprise line buffer memories for enabling spatial scaling in a vertical direction. The spatial scaling may comprise performing a zero order linear interpolation technique, i.e., a so-termed pixel repetition or nearest neighbour interpolation, as is known from the technical field of image processing. The spatial scaling may also comprise techniques such as first order linear interpolation, e.g., bilinear interpolation, higher order linear interpolation and non-linear interpolation techniques. Such techniques typically introduce fewer interpolation artefacts.

(33) It is noted that the imaging device 300 does not need to comprise an explicit scaler 250. Instead, a buffer may be used that effectively functions as a scaler. For example, the further rolling shutter circuit 240 may comprise a so-termed First-In-First-Out (FIFO) buffer, as is known from the technical field of processor design and architectures. The further rolling shutter circuit 240 may then read a row R.sub.m, as is shown schematically in FIG. 10, from the further radiation sensitive array 220. The reading may be performed using a 2.75 MHz pixel clock. The read out row R.sub.m may then be buffered in the FIFO, and read out with a higher pixel clock, e.g., a 44 MHz pixel clock, for providing the row R.sub.m repeatedly at a same readout speed as the row R.sub.n that is read out by the rolling shutter circuit 140. It is noted that such use of a FIFO buffer effectively performs a nearest neighbour interpolation, although it may conventionally not considered being a scaler. Also, it will be appreciated that such functionality may be also implemented in the signal processor 260 itself, i.e., the signal processor 260 may comprise the FIFO for performing said buffering.

(34) FIG. 11 shows a method 300 of acquiring and displaying images in real-time with an imaging device, the imaging device comprising i) an imaging sensor comprising a radiation sensitive array for acquiring an image, ii) a readout circuit connected to the radiation sensitive array for reading out the image, iii) a signal processor for processing the image for obtaining a processed image, and iv) a display for displaying the processed image, the radiation sensitive array being arranged in rows of sensor pixels and the display being arranged in rows of display pixels, and wherein the method comprises sequentially reading 340 out the rows of sensor pixels with the readout circuit for sequentially providing a subset of pixels, on availability of the subset of pixels, processing 360 the subset of pixels with the signal processor for providing a processed subset of pixels, and on availability of the processed subset of pixels, displaying 380 the processed subset of pixels with the display on a thereto corresponding subset of display pixels for displaying the processed image sequentially on the rows of display pixels.

(35) FIG. 12 shows a computer readable medium 400 comprising a computer program 420, the computer program 420 comprising instructions for causing a processor system to perform the method 300 as shown in FIG. 11. The computer program 420 may be embodied on the computer readable medium 400 as physical marks or by means of magnetization of the computer readable medium 400. However, any other suitable embodiment is conceivable as well. Furthermore, it will be appreciated that, although the computer readable medium 400 is shown in FIG. 12 as an optical disc, the computer readable medium 400 may be any suitable computer readable medium, such as a read-only-memory or random-access memory, e.g., solid state memory, flash memory, etc.

(36) It will be appreciated that the present invention may be used for various kinds of imaging sensors, and thus is not limited to, e.g., visible light imaging sensors or thermal sensors. Moreover, combining the image 142 with the further image 242 may comprise fusing the image 142 with the further image 242 by, e.g., overlaying certain elements of the further image 242 on top of the image 142. However, the combining may also comprise creating a processed image 262 that comprises a side-by-side, picture-in-picture or similar spatial arrangement of the image 142 and the further image 242.

(37) The signal processor 160 may employ various kinds of signal processing next to the aforementioned combining of fusing of the image 142 and the further image 242. For example, the signal processor 160 may perform various kinds of image processing, as are known from the technical field of image processing, such as non-uniformity correction, histogram equalization, noise reduction, sharpening, colour mapping, etc. Moreover, to increase a throughput of the signal processor 160, the signal processing may comprise an image processing pipeline for obtaining a pipelined processing of the subset of pixels 124.

(38) The display 180 may be a micro Organic Light Emitting Diode (OLED) or Liquid Crystal (LC) based display. The imaging sensor 110 may be a CMOS sensor. The signal processor 160 may be embodied in a FPGA. The imaging sensor 160 may be configured for providing synchronization information, e.g., so-termed horizontal and vertical SYNC signals. These may be used by the imaging device 100 to synchronize the reading, the processing and the displaying of the image 142. The synchronization information may also be used for synchronizing reading the further image 242 using the further readout circuit 240.

(39) It will be appreciated that the above description for clarity has described embodiments of the invention with reference to different functional units. However, it will be apparent that any suitable distribution of functionality between different functional units or processors may be used without detracting from the invention. For example, functionality illustrated to be performed by separate processors or controllers may be performed by the same processor or controllers. Hence, references to specific functional units are only to be seen as references to suitable means for providing the described functionality rather than indicative of a strict logical or physical structure or organization.

(40) The invention can be implemented in any suitable form including hardware, software, firmware or any combination of these. The invention may optionally be implemented at least partly as computer software running on one or more data processors and/or digital signal processors. The elements and components of an embodiment of the invention may be physically, functionally and logically implemented in any suitable way. Indeed the functionality may be implemented in a single unit, in a plurality of units or as part of other functional units. As such, the invention may be implemented in a single unit or may be physically and functionally distributed between different units and processors.

(41) Although the present invention has been described in connection with some embodiments, it is not intended to be limited to the specific form set forth herein. Rather, the scope of the present invention is limited only by the accompanying claims. Additionally, although a feature may appear to be described in connection with particular embodiments, one skilled in the art would recognize that various features of the described embodiments may be combined in accordance with the invention. In the claims, the term comprising does not exclude the presence of other elements or steps.

(42) Furthermore, although individually listed, a plurality of means, elements or method steps may be implemented by e.g. a single unit or processor. Additionally, although individual features may be included in different claims, these may possibly be advantageously combined, and the inclusion in different claims does not imply that a combination of features is not feasible and/or advantageous. Also the inclusion of a feature in one category of claims does not imply a limitation to this category but rather indicates that the feature is equally applicable to other claim categories as appropriate. Furthermore, the order of features in the claims do not imply any specific order in which the features must be worked and in particular the order of individual steps in a method claim does not imply that the steps must be performed in this order. Rather, the steps may be performed in any suitable order. In addition, singular references do not exclude a plurality. Thus references to “a”, “an”, “first”, “second” etc do not preclude a plurality. Reference signs in the claims are provided merely as a clarifying example shall not be construed as limiting the scope of the claims in any way.