METHOD FOR DETECTING OBJECT IMAGE USING HYPERSPECTRAL IMAGING BY BAND SELECTION
20230281818 · 2023-09-07
Inventors
- HSIANG-CHEN WANG (CHIAYI CITY, TW)
- Yu-Sheng Chi (Chiayi County, TW)
- YU-MING TSAO (CHIAYI COUNTY, TW)
- SIAN-HONG SHIH (CHIAYI COUNTY, TW)
Cpc classification
G06V10/255
PHYSICS
G06V10/7715
PHYSICS
G06V30/18114
PHYSICS
G06V10/77
PHYSICS
International classification
G06V20/69
PHYSICS
G06V10/77
PHYSICS
A61B1/273
HUMAN NECESSITIES
Abstract
The present application related to a method for detecting image using hyperspectral imaging by band selection. Firstly, obtaining a hyperspectral imaging information according to a reference image, hereby, obtaining corresponded hyperspectral image from an input image, and obtaining corresponded feature values by band selection for operating Principal components analysis to simplify feature values. Then, obtaining feature images by Convolution kernel, and then positioning an image of an object under detected by a default box and a boundary box from the feature image. By Comparing with the esophageal cancer sample image, the image of the object under detected is classifying to an esophageal cancer image or a non-esophageal cancer image. Thus, detecting an input image from the image capturing device by the convolutional neural network to judge if the input image is the esophageal cancer image for helping the doctor to interpret the image of the object under detected.
Claims
1. A method for detecting object image using hyperspectral imaging by band selection, comprising steps of: acquiring the hyperspectral image information formed by converting a reference image to a hyperspectral reference image, said reference image including one or more object reference image and a background reference image; an image extraction unit extracting an input image to a host, said input image including one or more object-under-test image and a background image; said host converting said input image according to said hyperspectral image information for giving a hyperspectral image; said host analyzing said hyperspectral image for giving a plurality of first hyperspectral characteristic values; said host performing band selection on said plurality of first hyperspectral characteristic values according to a cell and performing principal component analysis to simplify said hyperspectral image and generate a plurality of second characteristic values correspondingly; said host performing one or more layer of convolution calculation on said plurality of second characteristic values according to a plurality of kernels to filter out said background image and give a convolution result, and giving one or more selected image according to said convolution result and said one or more object-under-test image, said plurality of kernels including a plurality of selected characteristic values and a plurality of peripheral characteristic values, said one or more object-under-test image including a plurality of peripheral images and said one or more selected image, said plurality of peripheral images surrounding said one or more selected image, said one or more selected image corresponding to said plurality of selected characteristics, and said plurality of peripheral images corresponding to said plurality of peripheral characteristic values; said host generating one or more prediction box according to the edge of said one or more selected image; said host extracting a bounding box of said input image for comparing a first central point of said prediction box with a second central point of said bounding box and giving a center displacement between said prediction box and said bounding box, and said bounding box corresponding to an edge of said input image; said host performing a regression calculation according to said center displacement and giving a regression result; said host aligning said obj ect-under-test image according to said regression result and said prediction box so that said selected image can move towards said second central point as said first central point moves towards said second central point; said host matching and comparing said aligned object-under-test image with one or more sample image for generating a comparison result; and said host classifying said input image as a target-object image according to said comparison result.
2. The method for detecting object image using hyperspectral imaging by band selection of claim 1, where in said step of said host performing one or more layer of convolution calculation on said plurality of second characteristic values according to a plurality of kernels, said host sets said plurality of kernels as mxnxp and normalizes a plurality of pixels of said input image to a plurality of pixel normalized values, and by multiplying said plurality of kernels by said plurality of pixel normalized values, said plurality of second characteristic values are extracted in a convolution layer, where m=n and m=1, 3, 5, 10, 19, or 38.
3. The method for detecting object image using hyperspectral imaging by band selection of claim 1, where in said step of giving one or more selected image according to said convolution result and said one or more obj ect-under-test image, said host integrates the region of said plurality of selected characteristic values and gives one or more distribution region on said input image.
4. The method for detecting object image using hyperspectral imaging by band selection of claim 1, where in said step of said host performing one or more layer of convolution calculation on said plurality of second characteristic values according to a plurality of kernels, said host performs convolution on each pixel of said input image according to a single shot multibox detector model for detecting said plurality of second characteristic values.
5. The method for detecting object image using hyperspectral imaging by band selection of claim 1, where in said step of said host performing a regression calculation according to said center displacement, said host performs the regression calculation according to a first location of said prediction box, a second location of said bounding box, and a sizing factor for aligning said object-under-test image.
6. The method for detecting object image using hyperspectral imaging by band selection of claim 1, where in said step of said host matching and comparing said aligned object-under-test image with one or more sample image, said host matches and compares said object-under-test image and said one or more sample image in a fully connected layer.
7. The method for detecting object image using hyperspectral imaging by band selection of claim 1, where in said step of said host classifying said input image as a target-object image according to said comparison result, when said host cannot judge said input image as a target-object image according to said one or more sample image, said host performs similarity comparison on said object-under-test image according to said one or more sample image.
8. The method for detecting object image using hyperspectral imaging by band selection of claim 7, where in said step of said host performing similarity comparison on said object-under-test image according to said one or more sample image, if said host judges that a similarity level of said object-under-test image is greater than a similarity threshold value, said host judges said input image as said target-object image.
9. The method for detecting object image using hyperspectral imaging by band selection of claim 1, wherein said hyperspectral image information corresponds to a plurality of white light images and a plurality of narrow-band images and includes a plurality of color matching functions, a calibration matrix, and a conversion matrix.
10. The method for detecting object image using hyperspectral imaging by band selection of claim 1, where in said step of said host matching and comparing said aligned object-under-test image with one or more sample image, said host reads said one or more sample image from a database for matching and comparing according to said aligned object-under-test image.
11. The method for detecting object image using hyperspectral imaging by band selection of claim 1, where after said step of said host performing band selection on said plurality of first hyperspectral characteristic values according to a cell and performing principal component analysis to simplify said hyperspectral image and generate a plurality of second characteristic values correspondingly, said host converts said plurality of first characteristic values to a simulation image according to said band selection and said hyperspectral image information.
12. The method for detecting object image using hyperspectral imaging by band selection of claim 1, wherein said cell is an esophageal cancer cell.
13. The method for detecting object image using hyperspectral imaging by band selection of claim 11, wherein said band is 415 nanometers and 540 nanometers.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0031]
[0032]
[0033]
[0034]
[0035]
[0036]
[0037]
[0038]
[0039]
[0040]
DETAILED DESCRIPTION
[0041] In order to make the structure and characteristics as well as the effectiveness of the present application to be further understood and recognized, the detailed description of the present application is provided as follows along with embodiments and accompanying figures.
[0042] To solve the problems of artificial errors or difficulty in image identification caused by complicated endoscope operations according to the prior art, the present application provides a method for detecting object image using hyperspectral imaging by band selection.
[0043] In the following, the properties and the accompanying system provided by the method for detecting object image using hyperspectral imaging by band selection according to the present application will be further illustrated.
[0044] First, please refer to
Step S05: Acquiring the hyperspectral image information formed by converting a reference image to a hyperspectral reference image;
Step S10: An image extraction unit extracting an input image to a host;
Step S15: Converting the input image according to the hyperspectral image information for giving a hyperspectral image;
Step S20: Analyzing the hyperspectral image for giving first hyperspectral characteristic values;
Step S25: Performing band selection on the first hyperspectral characteristic values according to a cell and performing principal component analysis to generate second characteristic values correspondingly;
Step S30: Performing convolution calculation on the second characteristic values according to kernels to give a convolution result, and giving a selected image according to the convolution result and an object-under-test image;
Step S35: The host generating a prediction box according to the edge of a selected image;
Step S40: The host extracting a bounding box of the input image for comparing a first central point of the prediction box with a second central point of the bounding box and giving a center displacement between the prediction box and the bounding box;
Step S45: The host performing a regression calculation according to the center displacement and giving a regression result;
Step S50: The host aligning the object-under-test image according to the regression result and the prediction box so that the selected image can move towards the second central point as the first central point moves towards the second central point;
Step S55: The host comparing the object-under-test image with a sample image for generating a comparison result; and
Step S60: The host classifying the input image as a target-object image or a non-target-object image according to the comparison result.
[0045] Please refer to
[0046] In the step S05, as shown in
[0047] As shown in
[0048] Firstly, the reference image REF and the spectrometer (Ocean Optics-QE65000) should be converted to the same XYZ color space. The conversion formula for the reference image REF is:
[0049] f(n) is the gamma function capable of converting sRGB to linear RGB values. T is the conversion matrix. [M.sub.A] is a color adaptation matrix. According to (1), linear RGB values con be converted to XYZ values (XYZ endoscope) defined by the XYZ color space.
[0050] The formulae of converting the reflection spectrum data extracted by the spectrometer to the XYZ color space are as follows:
X=k∫.sub.380 nm.sup.780 nmS(λ)R(λ)
Y=k∫.sub.380 nm.sup.780 nmS(λ)R(λ)
Z=k∫.sub.380 nm.sup.780 nmS(λ)R(λ)
where k is given by (8)
k=100/∫.sub.380 nm.sup.780 nmS(λ)
[0051]
[0052] In addition, the calibration matrix C in (9) can be used to calibrate endoscope images:
[C]=[XYZ.sub.Spectrum]×pinν([V]) (9)
where the variable matrix [V] is given by analyzing the possible error factors of an endoscope. The error factors include nonlinear response of endoscope, dark currents of endoscope, inaccurate filtering of color filters, and color offsets (for example, white balance). Thereby, the XYZ values (XYZ.sub.Spectrum) can be calibrated.
[0053] Since the result of third-order calculation for the narrow-band image and the white light image shows they are similar, the nonlinear response correction adopts a third-order equation. To calibrate nonlinear response of the endoscope, the following equation (10) is adopted:
V.sub.Non-linear=[X.sup.3Y.sup.3Z.sup.3X.sup.2Y.sup.2Z.sup.2X Y Z 1].sup.T (10)
[0054] The dark current of a general endoscope is fixed. It will not change significantly as the incoming light varies. Thereby, the influence of dark currents can be considered constant. By defining the calibration variable for dark currents as VDark, the influence of dark currents can be expressed as (11):
V.sub.Dark=[α] (11)
[0055] The calibration variable for inaccurate filtering of color filter and color offset can be defined as V.sub.Color.
V.sub.Color=[XYZ XY XZ YZ X Y Z].sup.T (12)
[0056] According to equations (10) to (12), the calibrated variable matric V can be deduced in (13):
V=[X.sup.3Y.sup.3Z.sup.3 X.sup.2Y X.sup.2Z Y.sup.2Z XY.sup.2 XZ.sup.2 YZ.sup.2 XYZ X.sup.2Y.sup.2Z.sup.2 XY XZ YZ XYZ α].sup.T (13)
[0057] By using the variable matrix V and the calibration matrix C, the calibrated X, Y, Z values [XYZ.sub.Correct] are given in (14):
[XYZ.sub.Ccorrect]=[C]×[V] (14)
[0058] The average error of the white light image in [XYZ.sub.Correct] and [XYZ.sub.Spectrum] is 1.40; the average error of the narrow-band image in [XYZ.sub.Correct] and [XYZ.sub.Spectrum] is 2.39.
[0059] Because the above calculations adopt the visible-light band with wavelength between 380 and 780 nanometers, the calibration result of the endoscope must be expressed in chromatic aberration. [XYZ.sub.Correct] and [XYZ.sub.Spectrum] are converted to the corresponding Lab color space of the CIE DE2000. The conversion functions are equations (15) to (17):
where f(n) is:
[0060] The average chromatic aberration of the white light image before calibration reaches 11.60. After calibration, the average chromatic aberration is 2.84. The average chromatic aberration of the narrow-band image before calibration reaches 29.14. After calibration, the average chromatic aberration is 2.58.
[0061] In the step S10, as shown in
[0062] In the step S20, the host 10 extracts a plurality of first hyperspectral characteristic values F1 correspondingly according to the hyperspectral image HYI. In the step S25, the host 10 performs band selection on the plurality of first hyperspectral characteristic values F1 acquired in the step S20 according to a band BND of a cell for further performing principal component analysis (PCA). To simplify the calculation result, the characteristic values with lower variations are filtered out. Then the hyperspectral image HYI is simplified and giving a plurality of second characteristic values F2.
[0063] The formula for PCA is:
y.sub.i=α.sub.j1(x.sub.1i−
[0064] x.sub.1i to x.sub.ni represent the spectrum intensity of the first to the n-th band BND.
[0065] The principle of selecting the band BND is as follows. When the red light is absorbed by the hemoglobins in blood vessels at different depth, the capillaries in shallow mucosa tissues appear brown while the blood vessels in submucosa tissues appear green, resulting in significant visual layering and thus facilitating identification of affections in mucosa tissues. Thereby, according to the present embodiment, the bands 415 nm and 540 nm for globins crucial to detection of esophageal cancer cells are selected from the hyperspectral image HYI. Nonetheless, the present application is not limited to the embodiment. To detect different cells, different bands can be selected. Besides, by using the calibration values [XYZ.sub.Correct] and the corresponding reflection spectrum data [R.sub.Spectrum] of the 24 color checkers as described above, the corresponding conversion matrix M is calculated. pinv[V] is the pseudo inverse matrix of the matrix M:
[M]=[Score]×pinν([V.sub.Color]) (20)
[S.sub.Spectrum].sub.380-780=[EV][M][V.sub.Color] (21)
[0066] By performing PCA on the reflection spectrum [R.sub.Spectrum], a plurality of principal components (EV) will be given. According to the adopted principal component, the principal component score [Score] will be given. According to the present embodiment, 10 principal components with stronger explanatory power (with weighting percentage 8.0417%, 8.2212%, 2.6422%, 0.609%, 0.22382%, 0.10432%, 0.054658%, 0.0472%, 0.02638%, 0.012184%) are adopted to perform dimension reduction operation and giving the simulation spectrum [S.sub.Spectrum].sub.380-780. The error between the simulation spectrum [S.sub.Spectrum].sub.380-780 and the corresponding [XYZ.sub.Spectrum] of the input image IMG is corrected from 11.60 to 2.85 in the white light image and from 29.14 to 2.60 in the narrow-band image, and hence achieving almost indistinguishable color errors by bare eyes. Thereby, when color re-rendering is required, the present application provides better performance, and superior hyperspectral images in the visible light band can be simulated.
[0067] In the step S30, as shown in
[0068] Please refer again to
[0069] In the step S33, as shown in
[0070] Meanwhile, the height and width can be calculated according to the side length s.sub.k:
h.sub.k=s.sub.k√{square root over (α.sub.r)} (23)
w.sub.k=s.sub.k/√{square root over (α.sub.r)} (24)
where h.sub.k represents the height of the anchor box of the k-th feature map; w.sub.k represents the width of the rectangular anchor box; α.sub.r represents the length-to-width ratio of the prediction box D; and α.sub.r is greater than 0.
[0071] In the step S40, as shown in
Location of Prediction box D: d=(d.sup.cx, d.sup.cy, d.sup.w, d.sup.h) (25)
Location of bounding box B: b=(b.sup.cx, b.sup.cy, b.sup.w, b.sup.h) (26)
Sizing factor: l=(l.sup.cx, l.sup.cy, l.sup.w, l.sup.h) (27)
b.sup.cx=d.sup.wl.sup.cx+d.sup.cx (28)
b.sup.cy=d.sup.hl.sup.cy+d.sup.cy (29)
b.sup.w=d.sup.wexp(l.sup.w) (30)
b.sup.h=d.sup.hexp(l.sup.h) (31)
[0072] First, the center of the bounding box B is aligned with the center of the prediction box D, meaning moving the central point of the bounding box B to the central point of the prediction box D. In other words, as shown in
[0073] In addition, to define the location of the object-under-test image O1 more accurately, a loss equation is further adopted:
L.sub.loc(x, l, g)=Σ.sub.i∈Pos.sup.NΣ.sub.m∈{cx,cy,w,h}x.sub.ij.sup.k smooth.sub.L1(l.sub.i.sup.m−ĝ.sub.j.sup.m) (32)
[0074] Thereby, the error between the location of the prediction box D and the location of the object-under-test image O1 can be verified.
[0075] In the step S35, as shown in
[0076]
[0077] According to the method for detecting object image using hyperspectral imaging by band selection according to the present application, band selection is performed on the hyperspectral image HYI. In the hyperspectral image HYI, the spectrum intensity of different bands of visible colors can be identified. The multi-dimensional hyperspectral image HYI can filtered to keep the required bands (415 nm and 540 nm are the two bands adopted according to the present embodiment). The required bands are processed by dimension reduction before subsequent calculations. Then the convolutional neural network is used to classify into target-object image or non-target-object image. 1780 results of finding lesions in esophagoscope images according to the present application is compared with the results using input images IMG with white light imaging (WLI) or with narrow-band imaging (NBI). After band selection, the accuracy of judging dysplasia regions in hyperspectral images HYI is raised to 98% (WLI: 96%; NBI: 85%); the accuracy of judging esophageal cancer (SCC) using band-selected hyperspectral images HYI is raised to 93% (WLI: 92%; NBI: 82%). Medical personnel can utilize the present application as supporting evidence for diagnoses. The present application does not adjust the colors of the input image IMG. Instead, the present application filters the light with colors.
[0078] According to another embodiment of the present application, the output of the band-selected hyperspectral image is a general image. After the step S25, the method further comprises a step:
Step S28: Converting the first hyperspectral characteristic values to a simulation image according to band selection and the hyperspectral image information.
[0079] The present application further provides an embodiment. As shown in
[0080] In the step S25, PCA and dimension reduction calculations are performed on the reflection spectrum data [R.sub.Spectrum] to give the simulation spectrum [S.sub.Spectrum].sub.380-780. In the step S28, according to the equations (5) to (7), the simulation spectrum [S.sub.Spectrum].sub.380-780 is converted to the XYZ values defined in the XYZ color space. Afterwards, (33) is used to convert the XYZ values to the simulation image SIM.
[0081] In the embodiment, the data after dimension reduction can be converted to a general image (RGB image). According to an embodiment, the selection band is 415 nm and 540 nm, which will generate simulation images SIM similar to narrow-band images. The penetration of light with different wavelengths into tissues differs. When the red light is absorbed by the hemoglobins in blood vessels at different depth, the capillaries in shallow mucosa tissues appear brown while the blood vessels in submucosa tissues appear green. The simulation images SIM according to the present application are generated by the two selection bands from the white light images with 401 bands. Consequently, the lesion region (the target-object image TA) originally with similar colors with the background appears prominent and hence improving the contrast with respect to the background image BG.
[0082]
[0083] To sum up, the present application provides a method for detecting object image using hyperspectral imaging by band selection. The host acquires the hyperspectral image information and converts the input image to the hyperspectral image according to the hyperspectral image information. Next, the convolution program is executed continuously on the hyperspectral image according to the band selection and enables the host to build the convolutional neural network for performing convolution on the input image from the image extraction unit. By filtering the region to be detected, the prediction box is generated on the input image. Then, by using the regression calculation, the bounding box is used to locate the location of the object-under-test image. Finally, comparison with the sample image is performed. By using the comparison result, the target-object image and the non-target-object image can be classified. Thereby, the purpose of identifying object image using the hyperspectral technology can be achieved.