Method and apparatus for performing object detection based on images captured by a fisheye camera and electronic device
11398009 · 2022-07-26
Assignee
Inventors
Cpc classification
H04N13/111
ELECTRICITY
G06F18/214
PHYSICS
G06V20/52
PHYSICS
B60R2300/102
PERFORMING OPERATIONS; TRANSPORTING
International classification
G06V20/52
PHYSICS
G06V10/24
PHYSICS
G06V10/422
PHYSICS
H04N13/111
ELECTRICITY
Abstract
This disclosure provides an apparatus and method for performing object detection based on images captured by a fisheye camera and an electronic device. The apparatus includes a memory and a processor coupled to the memory. The processor according to an embodiment is configured to: project an original image captured by the fisheye camera onto a cylindrical or spherical projection model, and perform reverse mapping to obtain at least two reversely mapped images, angles of view of the at least two reversely mapped images being towards different directions, detect objects in the reversely mapped images, respectively, and detect an object that is the same among the objects detected in the reversely mapped images. According to this disclosure, information in the wide field of view images obtained by capturing by the fisheye camera may be fully utilized.
Claims
1. An apparatus for performing object detection based on images captured by a fisheye camera, the apparatus comprising: a memory; a processor coupled to the memory and configured to: project original images captured by the fisheye camera onto a cylindrical or spherical projection model, and perform reverse mapping to obtain at least two reversely mapped images corresponding to a same original image, angles of view of the at least two reversely mapped images being towards different directions, detect objects in the at least two reversely mapped images, respectively, and detect an object that is the same among the objects detected in the at least two reversely mapped images, wherein, in performing the reverse mapping, the processor sets an offset related to a width and/or a height of an original image among the original images, the offset being set to be greater than a value and less than the value respectively, so as to obtain the at least two reversely mapped images, wherein, when the offset is equal to the value, the angles of view face a center of the width and/or the height of the original image, wherein the processor is further configured to: map, for objects existing in overlapped regions of the at least two reversely mapped images, contour frames of the objects in the overlapped regions of the at least two reversely mapped images into the original images to obtain mapped contour frames, and detect a degree of overlapping of the mapped contour frames, and determine that the object is the same when the degree of overlapping is greater than a threshold.
2. The apparatus according to claim 1, wherein, the processor performs detection on the at least two reversely mapped images based on a foreground and a background, and/or performs detection of the object based on a classifier obtained based on a training.
3. The apparatus according to claim 1, wherein the processor is further configured to: detect a motion state of the object according to a result of the detection of the object.
4. The apparatus according to claim 3, wherein the processor is further configured to: calculate, when determining that the objects detected in the at least two reversely mapped images are the object that is the same among the objects, motion velocities of the object respectively based on the at least two reversely mapped images, and detect the motion state of the object according to a difference between the motion velocities of the object calculated respectively based on the at least two reversely mapped images.
5. An electronic device, comprising the apparatus as claimed in claim 1.
6. A method for performing object detection based on images captured by a fisheye camera, the method comprising: projecting original images captured by the fisheye camera onto a cylindrical or spherical projection model, and performing reverse mapping to obtain at least two reversely mapped images corresponding to a same original image, angles of view of the at least two reversely mapped images being towards different directions; detecting objects in the at least two reversely mapped images, respectively; and detecting an object that is the same among the objects detected in the at least two reversely mapped images, wherein, in performing the reverse mapping, an offset related to a width and/or a height of an original image among the original images, the offset being set to be greater than a value and less than the value respectively, so as to obtain the at least two reversely mapped images; wherein, when the offset is equal to the value, the angles of view face a center of the width and/or the height of the original image, wherein the detecting of the object that is the same among the objects detected in the at least two reversely mapped images comprises: mapping, for objects existing in overlapped regions of the at least two reversely mapped images, contour frames of the objects in the overlapped regions of the at least two reversely mapped images into the original images to obtain mapped contour frames, and detecting a degree of overlapping of the mapped contour frames, and judging that the object is the same when the degree of overlapping is greater than a threshold.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) Many aspects of the disclosure can be better understood with reference to the following drawings. The components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of this disclosure. To facilitate illustrating and describing some parts of the disclosure, corresponding portions of the drawings may be exaggerated or reduced. Elements and features depicted in one drawing or embodiment of the disclosure may be combined with elements and features depicted in one or more additional drawings or embodiments. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views and may be used to designate like or similar parts in more than one embodiment.
(2) The drawings are included to provide further understanding of this disclosure, which constitute a part of the specification and illustrate the preferred embodiments of this disclosure, and are used for setting forth the principles of this disclosure together with the description. It is obvious that the accompanying drawings in the following description are some embodiments of this disclosure, and for those of ordinary skills in the art, other accompanying drawings may be obtained according to these accompanying drawings without making an inventive effort. In the drawings:
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
DETAILED DESCRIPTION OF THE DISCLOSURE
(14) These and further aspects and features of the present disclosure will be apparent with reference to the following description and attached drawings. In the description and drawings, particular embodiments of the disclosure have been disclosed in detail as being indicative of some of the ways in which the principles of the disclosure may be employed, but it is understood that the disclosure is not limited correspondingly in scope. Rather, the disclosure includes all changes, modifications and equivalents coming within the terms of the appended claims.
(15) In the embodiments of this disclosure, terms “first”, and “second”, etc., are used to differentiate different elements with respect to names, and do not indicate spatial arrangement or temporal orders of these elements, and these elements should not be limited by these terms. Terms “and/or” include any one and all combinations of one or more relevantly listed terms. Terms “contain”, “include” and “have” refer to existence of stated features, elements, components, or assemblies, but do not exclude existence or addition of one or more other features, elements, components, or assemblies.
(16) In the embodiments of this disclosure, single forms “a”, and “the”, etc., include plural forms, and should be understood as “a kind of” or “a type of” in a broad sense, but should not defined as a meaning of “one”; and the term “the” should be understood as including both a single form and a plural form, except specified otherwise. Furthermore, the term “according to” should be understood as “at least partially according to”, the term “based on” should be understood as “at least partially based on”, except specified otherwise.
Embodiment 1
(17) Embodiment 1 provides a method for performing object detection, in which the object detection is performed based on images captured by a fisheye camera.
(18)
(19) block 101: original images captured by the fisheye camera are projected onto a cylindrical or spherical projection model, and reverse mapping is performed to obtain at least two reversely mapped images, angles of view of the at least two reversely mapped images being towards different directions;
(20) block 102: objects in the reversely mapped images are detected respectively; and
(21) block 103: the same object among the objects detected from the reverse mapping images is detected.
(22) According to this embodiment, when an image captured by the fisheye camera is reversely mapped, at least two reversely mapped images with angles of view facing towards different directions are formed, and objects are detected based on the reversely mapped images. Hence, it is not necessary to cut off the images captured by the fisheye camera, and information in the wide field of view images obtained by capturing by the fisheye camera may be fully utilized.
(23) In this embodiment, the original images captured by the fisheye camera may be from images captured by the fisheye camera in a real-time manner, or may be from images captured by the fisheye camera stored in a storage device, which is not limited in this embodiment.
(24) In this embodiment, the subject captured by the fisheye camera may be a road or an indoor environment, and the object may be a person or a vehicle in the image. In the following description of this embodiment, the fisheye camera is used for traffic surveillance, the object is a traffic lane, and the object is a vehicle or the like on a traffic lane. However, this embodiment is not limited thereto, and the fisheye camera may also be used for capturing other subjects, and the object may also be other objects in the image.
(25) In block 101 of this embodiment, the original images captured by the fisheye camera may be mapped onto the cylindrical model or the spherical model, so that the original images may be corrected. Reference may be made to the related art for a particular method for mapping the original images onto the cylindrical surface model or the spherical surface model, such as mapping the original images to the cylindrical surface model or the spherical surface model in performing image stitching.
(26) In one implementation of this embodiment, the original images may be mapped to the cylindrical surface model in block 101, the cylindrical surface model may be, for example, a semi-cylindrical model, thus, images in the original images in a range of field of view of 180 degrees may be mapped onto the semi-cylindrical model.
(27) Furthermore, in this embodiment, before the original images are mapped onto the semi-cylindrical surface model, sizes of the original images may be adjusted so that the adjusted sizes are in conformity with formula (1) below:
(28)
(29) where, width denotes a width of the adjusted image, i.e. a size in the horizontal direction, height denotes a height of the adjusted image, i.e. a size in the vertical direction perpendicular to the horizontal direction, HFOV is a horizontal field of view of an original image, i.e. a field of view in a width direction, UVFOV is a first field of view of the original image in the vertical direction, and DVFOV is a second field of view of the original image in the vertical direction; UVFOV and DVFOV respectively indicate opening angles of a central optical axis of a lens of the fisheye camera with respect to an upper and lower edges of the original image.
(30) Furthermore, the 180 degree in formula (1) refers to a center angle to which a cross-sectional arc of the semi-cylindrical model corresponds; and when the semi-cylindrical model is replaced with a whole cylindrical model, a center angle to which a cross-sectional arc of the cylindrical model corresponds is 360 degrees, hence, 180 degrees in formula (1) may be replaced with 360 degrees.
(31)
(32) In the following description of this embodiment, description shall be given by taking a scheme in which the size adjustment is performed as an example. However, this embodiment may not be limited thereto. For example, the above-described size adjustment may not be performed, and in a case where the size adjustment is not performed, the description for the size-adjusted image in the following description may be applicable to description of the original images.
(33) In block 101 of this embodiment, the image projected onto the cylindrical projection model may be reversely mapped, that is, the image on the cylindrical projection model may be projected onto a plane.
(34)
(35) As shown in
(36)
(37)
(38) where, height.sub.target and width.sub.target respectively denote a height and width of the planar image 302, f denotes a focal lengths of the fisheye camera, and h.sub.1 denotes a height of the above adjusted image.
(39) In a case where the height and width of the planar image 302 are given, there exists a relationship between P′(X,Y) and P(u,v), which is shown by formulae (4) and (5) below:
(40)
(41) The above formulae are limited by conditions: offset∈(0,1),
(42)
where, width denotes the width of the above size-adjusted image.
(43) In the present embodiment, offset appears in formula (4), which indicates an offset related the widths of the original images.
(44) In this embodiment, in performing the reverse mapping based on the above formulae (2)-(4), by adjusting values of the offset, planar images with angles of view towards different directions in the width direction may be obtained. For example, if a value of offset is set to be greater than a first value, a first planar image may be obtained, and if a value of offset is set to be less than the first value, a second planar image may be obtained, angles of view of the first planar image and the second planar image facing towards different directions in the width direction.
(45) In one implementation, when a value of offset is the first value, the angle of view of the planar image in the width direction faces towards a center position in the width direction; when a value of offset is less than the first value, the angle of view of the planar image in the width direction is left to the center position in the width direction; and when a value of offset is great than the first value, the angle of view of the planar image in the width direction is right to the center position in the width direction.
(46) For example, the first value may be 0.5, when a value of offset is less than the first value, the value of offset may be 0.45, and when a value of offset is great than the first value, the value of offset may be 0.55.
(47)
(48) As shown in
(49) Furthermore, in this embodiment, an offset 1 related to the height of the original images may be set in formula (5), and by adjusting offset1, at least two reversely mapped images with angles of vies towards different directions in the height direction may be obtained.
(50) In one implementation, in performing reverse mapping in block 101, the offset may be adjusted to obtain at least two reversely mapped images, and/or offset 1 may be adjusted to obtain at least two reversely mapped images.
(51) Furthermore, in the above description of this embodiment, in the case of projecting the original images onto the cylindrical projection model, reverse mapping is performed and at least two reversely map images are obtained. However, this embodiment is not limited thereto. For example, in the case of projecting the original images onto a spherical projection model, reverse mapping may also be performed and at least two reversely mapped images may be obtained.
(52) In the following description of this embodiment, description shall be given by taking “adjusting the offset to obtain two reversely mapped images” as an example, and processing in other cases may be similar thereto.
(53) In block 102 of this embodiment, objects in the reversely mapped images may be separately detected for the at least two reversely mapped images generated in block 101, and thus, for the reversely mapped images, there are corresponding target detection results. For example, for (b) and (c) in
(54) In one implementation, the objects in the reversely mapped images may be detected based on a method of foreground and background detection, or the objects in the reversely mapped images may be detected based on the classifier, or the two detection methods are combined. For example, the object detected in the method of foreground and background detection and the target detected based on the classifier are fused to obtain a final object detection result.
(55)
(56) block 501: foreground detection and background detection are performed on the reversely mapped images; for example, the foreground detection and background detection may be performed by using a Vibe algorithm;
(57) block 502: the foreground is threshold based on a preset threshold to obtain a binarized image; for example, pixels in the foreground with pixel values greater than or equal to the threshold are set as white pixels, and pixels in the foreground with pixel values less than the threshold value are set as black pixels, so as to obtain the binarized image;
(58) block 503: morphological processing is performed on the binarized image, and regions with a relatively small number of pixels in the binarized image are removed, so as to obtain at least one pixel set (blob);
(59) block 504: connected domain processing are performed on the binarized image after being performed morphological processing, so as to merge adjacent pixel sets into relatively large pixel set;
(60) block 505: blobs are clustered based on a result of the connected domain processing; for example, pixel sets with center distances less than a certain threshold are clustered into a pixel cluster, one pixel cluster being taken as a detected object;
(61) block 506: an object is detected in the reversely mapped images based on the classifier; the classifier may be a classifier obtained based on target training, for example, the classifier may be a classifier based on graphics training of a vehicle; hence, a vehicle may be detected from the reversely mapped images; and the classifier may detect based on, for example, a histogram of oriented gradient (HOG) feature; and
(62) block 507: the object detected in block 505 and the object detected in block 506 are fused to obtain a final object detection result; for example, a contour frame Bbox1 of an object 1 detected in block 505 is compared with a contour frame Bbox2 of an object 2 detected in block 506, and when an intersection-over-union (IoU) of Bbox1 to and Bbox2 exceeds a certain threshold Tmerge, it is determined object 1 is the final object, otherwise, it is not determined object 1 is the final object.
(63) In
(64) In block 103 of this embodiment, the same target may be detected from the objects detected in the reversely mapped images.
(65)
(66) block 601: for objects existing in all overlapped regions of the reversely mapped images, contour frames of the objects in the overlapped regions of the reversely mapped images are mapped into the original images to obtain mapped contour frames; and
(67) block 602: a degree of overlapping of the mapped contour frames is detected, and it is determined that the objects are the same object when the degree of overlapping is greater than a first threshold.
(68) The method for detecting the same object shall be described below by taking the images in
(69) As shown in
(70) Assuming that M objects are detected in (b) in
(71) In this embodiment, a mark may be set for each object to indicate whether the object is located in the region S.sub.public1. For example, a mark F.sub.i.sup.S.sup.
(72) In this embodiment, the number of objects in the region S.sub.public1 may be determined by the mark F.sub.i.sup.S.sup.
(73) Assuming that total N objects are detected in (c) of
(74) In this embodiment, a mark may be set for each object to indicate whether the object is located in the region S.sub.public2. For example, a mark F.sub.j.sup.S.sup.
(75) In this embodiment, the number of objects in the region S.sub.public2 may be determined by the mark F.sub.j.sup.S.sup.
(76) In this embodiment, the bounding box.sub.i.sup.S.sup.
(77) In this embodiment, when a degree of overlapping of the bounding boxes box.sub.i.sup.S.sup.
(78) Furthermore, in this embodiment, when the object in the contour frame box.sub.i.sup.S.sup.
(79) In this embodiment, the determination as to whether the object moves in the positive direction or the inverse direction may be performed, for example, according to a moving trajectory of the object in images in a period of time, and reference may be made to the related art for a particular method.
(80) In this embodiment, as shown in
(81) block 104: a motion state of the object is detected according to a result of detection of block 103.
(82) In block 104 of this embodiment, when it is determined that the objects detected in the reversely mapped images are the same object, motion velocities of the object are calculated respectively based on the reversely mapped images, and the motion state of the object is detected according to a difference between the motion velocities of the object calculated respectively based on the reversely mapped images.
(83) For example, when the object in the contour frame box.sub.i.sup.S.sup.
(84) In the above description of this embodiment, the motion states of the objects are of acceleration or deceleration; however, this embodiment is not limited thereto, and the motion states of the objects may be other.
(85) The method of this embodiment shall be described below by way of an example.
(86)
(87) As shown in
(88) block 701: a fisheye camera shoots to obtain original images;
(89) block 702: sizes of the original images are adjusted, and the adjusted images are projected to the cylindrical projection model;
(90) block 703: an offset less than the first value is set, and reverse mapping is performed to obtain a first reversely mapped image, such as (b) in
(91) block 704: an offset greater than the first value is set, and reverse mapping is performed to obtain a second reversely mapped image, such as (c) in
(92) block 705: an object is detected in the first reversely mapped image;
(93) block 706: an object is detected in the second reversely mapped image;
(94) block 707: it is determined whether the i-th object is located in the region S.sub.public1, that is, whether F.sub.i.sup.S.sup.
(95) block 708: box.sub.i.sup.S.sup.
(96) block 709: a velocity of the i-th object is calculated;
(97) block 710: it is determined whether the j-th object is located in the region S.sub.public2, that is, whether F.sub.j.sup.S.sup.
(98) block 711: box.sub.j.sup.S.sup.
(99) block 712: a velocity of the j-th object is calculated;
(100) block 713: it is determined whether the IoU of box.sub.i.sup.S.sup.
(101) block 714: it is determined whether the moving direction of the object is the positive direction, entering into block 715 if it is determined “Yes”, and entering into block 716 if it is determined “No”;
(102) block 715: F.sub.box.sub.
(103) block 716: F.sub.box.sub.
(104) According to this embodiment, when an image captured by the fisheye camera is reversely mapped, at least two reversely mapped images with angles of view facing towards different directions are formed, and objects are detected based on the reversely mapped images. Hence, it is not necessary to cut off the images captured by the fisheye camera, and information in the wide field of view images obtained by capturing by the fisheye camera may be fully utilized.
Embodiment 2
(105) Embodiment 2 provides an apparatus for performing object detection based on images captured by a fisheye camera. As a principle of the apparatus for solving problems is similar to that of the method in Embodiment 1, reference may be made to the implementation of the method in Embodiment 1 for implementation of the apparatus, with identical contents being not going to be described herein any further.
(106)
(107) a first processing unit 801 configured to project original images captured by the fisheye camera onto a cylindrical or spherical projection model, and perform reverse mapping to obtain at least two reversely mapped images, angles of view of the at least two reversely mapped images being towards different directions;
(108) a second processing unit 802 configured to detect objects in the reversely mapped images respectively; and
(109) a third processing unit 803 configured to detect the same object in the objects detected in the reversely mapped images.
(110) In performing the reverse mapping, the first processing unit 801 sets an offset related to a width and/or a height of an original image to be greater than a first value and less than the first value respectively, so as to obtain two reversely mapped images; wherein, when the offset is equal to the first value, the angles of view face a center of the width and/or height of the original image.
(111) In this embodiment, the second processing unit 802 performs detection based on a foreground and a background, and/or performs detection based on a classifier.
(112)
(113) a fourth processing unit 901 configured to, for objects existing in overlapped regions of the reversely mapped images, map contour frames of the objects in the overlapped regions of the reversely mapped images into the original images to obtain mapped contour frames; and
(114) a fifth processing unit 902 configured to detect a degree of overlapping of the mapped contour frames, and determine that the objects are the same object when the degree of overlapping is greater than a first threshold.
(115) As shown in
(116) a sixth processing unit 804 configured to detect a motion state of the object according to a result of detection of the third processing unit.
(117)
(118) a seventh processing unit 1001 configured to, when it is determined that the objects detected in the reversely mapped images are the same object, calculate motion velocities of the object respectively based on the reversely mapped images; and
(119) an eighth processing unit 1002 configured to detect the motion state of the object according to a difference between the motion velocities of the object calculated respectively based on the reversely mapped images.
(120) Reference may be made to corresponding blocks in Embodiment 1 for detained description of the units in this embodiment, which shall not be described herein any further.
(121) According to this embodiment, when an image captured by the fisheye camera is reversely mapped, at least two reversely mapped images with angles of view facing towards different directions are formed, and objects are detected based on the reversely mapped images. Hence, it is not necessary to cut off the images captured by the fisheye camera, and information in the wide field of view images obtained by capturing by the fisheye camera may be fully utilized.
Embodiment 3
(122) Embodiment 3 provides an electronic device. As a principle of the electronic device for solving problems is similar to that of the apparatus 800 in Embodiment 2, reference may be made to the implementation of the apparatus 800 in Embodiment 2 for implementation of the electronic device, with identical contents being not going to be described herein any further.
(123)
(124) In one implementation, the functions of the apparatus 800 may be integrated into the central processing unit 1101, wherein the central processing unit 1101 may be configured to carry out the method described in Embodiment 1.
(125) The central processing unit 1101 may be configured to perform control, so that the electronic device 1100 carries out the following method:
(126) projecting original images captured by the fisheye camera onto a cylindrical or spherical projection model, and performing reverse mapping to obtain at least two reversely mapped images, angles of view of the at least two reversely mapped images being towards different directions; detecting objects in the reversely mapped images respectively; and detecting the same object in the objects detected in the reversely mapped images.
(127) The central processing unit 1101 may further be configured to: in performing the reverse mapping, set an offset related to a width and/or a height of an original image to be greater than a first value and less than the first value respectively, so as to obtain two reversely mapped images;
(128) when the offset is equal to the first value, the angles of view face a center of the width and/or height of the original image.
(129) The central processing unit 1101 may further be configured to: perform detection based on a foreground and a background, and/or perform detection based on a classifier.
(130) The central processing unit 1101 may further be configured to: for objects existing in overlapped regions of the reversely mapped images, map contour frames of the objects in the overlapped regions of the reversely mapped images into the original images to obtain mapped contour frames; and
(131) detect a degree of overlapping of the mapped contour frames, and determine that the objects are the same object when the degree of overlapping is greater than a first threshold.
(132) The central processing unit 1101 may further be configured to: detect a motion state of the object according to a result of detection of the third processing unit.
(133) The central processing unit 1101 may further be configured to: when it is determined that the objects detected in the reversely mapped images are the same object, calculate motion velocities of the object respectively based on the reversely mapped images; and
(134) detect the motion state of the object according to a difference between the motion velocities of the object calculated respectively based on the reversely mapped images.
(135) Furthermore, as shown in
(136) According to this embodiment, when an image captured by the fisheye camera is reversely mapped, at least two reversely mapped images with angles of view facing towards different directions are formed, and objects are detected based on the reversely mapped images. Hence, it is not necessary to cut off the images captured by the fisheye camera, and information in the wide field of view images obtained by capturing by the fisheye camera may be fully utilized.
(137) An embodiment of the present disclosure provides a computer storage medium, including a computer readable program code, which will cause an apparatus for performing object detection based on images captured by a fisheye camera or an electronic device to carry out the method for performing object detection based on images captured by a fisheye camera as described in Embodiment 1.
(138) An embodiment of the present disclosure provides a computer readable program code, which, when executed in an apparatus for performing object detection based on images captured by a fisheye camera or an electronic device, will cause the apparatus for performing object detection based on images captured by a fisheye camera or the electronic device to carry out the fall detection method as described in Embodiment 1.
(139) The above apparatuses and methods of this disclosure may be implemented by hardware, or by hardware in combination with software. This disclosure relates to such a computer-readable program that when the program is executed by a logic device, the logic device is enabled to carry out the apparatus or components as described above, or to carry out the methods or blocks as described above. The present disclosure also relates to a storage medium for storing the above program, such as a hard disk, a floppy disk, a CD, a DVD, and a flash memory, etc.
(140) The methods/apparatuses described with reference to the embodiments of this disclosure may be directly embodied as hardware, software modules executed by a processor, or a combination thereof. For example, one or more functional block diagrams and/or one or more combinations of the functional block diagrams shown in
(141) The soft modules may be located in an RAM, a flash memory, an ROM, an EPROM, and EEPROM, a register, a hard disc, a floppy disc, a CD-ROM, or any memory medium in other forms known in the art. A memory medium may be coupled to a processor, so that the processor may be able to read information from the memory medium, and write information into the memory medium; or the memory medium may be a component of the processor. The processor and the memory medium may be located in an ASIC. The soft modules may be stored in a memory of a mobile terminal, and may also be stored in a memory card of a pluggable mobile terminal. For example, if equipment (such as a mobile terminal) employs an MEGA-SIM card of a relatively large capacity or a flash memory device of a large capacity, the soft modules may be stored in the MEGA-SIM card or the flash memory device of a large capacity.
(142) One or more functional blocks and/or one or more combinations of the functional blocks in
(143) This disclosure is described above with reference to particular embodiments. However, it should be understood by those skilled in the art that such a description is illustrative only, and not intended to limit the protection scope of the present disclosure. Various variants and modifications may be made by those skilled in the art according to the principle of the present disclosure, and such variants and modifications fall within the scope of the present disclosure.
(144) Following implementations are further provided in this disclosure.
(145) 1. An apparatus for performing object detection based on images captured by a fisheye camera, the apparatus including:
(146) a first processing unit configured to project original images captured by the fisheye camera onto a cylindrical or spherical projection model, and perform reverse mapping to obtain at least two reversely mapped images, angles of view of the at least two reversely mapped images being towards different directions;
(147) a second processing unit configured to detect objects in the reversely mapped images respectively; and
(148) a third processing unit configured to detect the same object in the objects detected in the reversely mapped images.
(149) 2. The apparatus according to supplement 1, wherein,
(150) in performing the reverse mapping, the first processing unit sets an offset related to a width and/or a height of an original image to be greater than a first value and less than the first value respectively, so as to obtain two reversely mapped images;
(151) wherein, when the offset is equal to the first value, the angles of view face a center of the width and/or height of the original image.
(152) 3. The apparatus according to supplement 1, wherein,
(153) the second processing unit performs detection based on a foreground and a background, and/or performs detection based on a classifier.
(154) 4. The apparatus according to supplement 1, wherein the third processing unit includes:
(155) a fourth processing unit configured to, for objects existing in overlapped regions of the reversely mapped images, map contour frames of the objects in the overlapped regions of the reversely mapped images into the original images to obtain mapped contour frames; and
(156) a fifth processing unit configured to detect a degree of overlapping of the mapped contour frames, and determine that the objects are the same object when the degree of overlapping is greater than a first threshold.
(157) 5. The apparatus according to supplement 1, wherein the apparatus further includes:
(158) a sixth processing unit configured to detect a motion state of the object according to a result of detection of the third processing unit.
(159) 6. The apparatus according to supplement 5, wherein the sixth processing unit includes:
(160) a seventh processing unit configured to, when it is determined that the objects detected in the reversely mapped images are the same object, calculate motion velocities of the object respectively based on the reversely mapped images; and
(161) an eighth processing unit configured to detect the motion state of the object according to a difference between the motion velocities of the object calculated respectively based on the reversely mapped images.
(162) 7. An electronic device, including the apparatus as described in any one of supplements 1-6.
(163) 8. A method for performing object detection based on images captured by a fisheye camera, the method including:
(164) projecting original images captured by the fisheye camera onto a cylindrical or spherical projection model, and performing reverse mapping to obtain at least two reversely mapped images, angles of view of the at least two reversely mapped images being towards different directions;
(165) detecting objects in the reversely mapped images respectively; and
(166) detecting the same object in the objects detected in the reversely mapped images.
(167) 9. The method according to supplement 8, wherein,
(168) in performing the reverse mapping, an offset related to a width and/or a height of an original image is set to be greater than a first value and less than the first value respectively, so as to obtain two reversely mapped images;
(169) wherein, when the offset is equal to the first value, the angles of view face a center of the width and/or height of the original image.
(170) 10. The method according to supplement 8, wherein the detecting objects in the reversely mapped images includes:
(171) performing detection based on a foreground and a background, and/or performing detection based on a classifier.
(172) 11. The method according to supplement 8, wherein the detecting the same object in the objects detected in the reversely mapped images includes:
(173) for objects existing in overlapped regions of the reversely mapped images, mapping contour frames of the objects in the overlapped regions of the reversely mapped images into the original images to obtain mapped contour frames; and
(174) detecting a degree of overlapping of the mapped contour frames, and judging that the objects are the same object when the degree of overlapping is greater than a first threshold.
(175) 12. The method according to supplement 8, wherein the method further includes: detecting a motion state of the object according to a result of the detection.
(176) 13. The method according to supplement 12, wherein the detecting a motion state of the object includes:
(177) when it is determined that the objects detected in the reversely mapped images are the same object, calculating motion velocities of the object respectively based on the reversely mapped images; and
(178) detecting the motion state of the object according to a difference between the motion velocities of the object calculated respectively based on the reversely mapped images.