METHOD FOR MULTI-CENTER EFFECT COMPENSATION BASED ON PET/CT INTELLIGENT DIAGNOSIS SYSTEM

20220399119 · 2022-12-15

    Inventors

    Cpc classification

    International classification

    Abstract

    Disclosed is a method for multi-center effect compensation based on a PET/CT intelligent diagnosis system. The method includes the following steps: estimating multi-center effect parameters of a test center B relative to a training center A by implementing a nonparametric mathematical method for data of the training center A and the test center B based on a location-scale model about additive and multiplicative multi-center effect parameters, and using the parameters to compensate the data of the test center B to eliminate a multi-center effect between the test center B and the training center A. According to the present disclosure, the multi-center effect between the training center A and the test center B can be compensated, so that the compensated data of the test center B can be used in the model trained by the training center A, and the generalization ability of the model is indirectly improved.

    Claims

    1. A method for multi-center effect compensation based on a PET/CT intelligent diagnosis system, comprising the following steps: step 1, importing PET/CT data of a training center A into a deep convolution neural network model of the training center A, obtaining a feature map of all the PET/CT data in the training center A, calculating an average value {circumflex over (α)}g, a covariant regression coefficient {circumflex over (β)}.sub.g and a variance {circumflex over (σ)}.sub.g of all features Y.sub.Ajg in the feature map of all the PET/CT data in the training center A, and carrying out Z-Score standardization thereon to obtain a standardized feature Z.sub.Ajg: Z Ajg = Y Ajg - α ˆ 9 - X β ˆ g σ ˆ g ( 1 ) where Z.sub.Ajg represents a feature g extracted from a case j in the training center A, and X is a design matrix of a relevant covariates; step 2, importing PET/CT data in a test center B into the deep convolution neural network model of the training center A, obtaining a feature map of all the PET/CT data in the test center B, and carrying out Z-Score standardization on the features Y.sub.Bjg in the feature map by using the average value and the variance of the training center A obtained in step 1, and obtaining a standardized feature Z.sub.Bjg: Z Bjg = Y Bjg - α ˆ g - X β ˆ g σ ˆ g ; ( 2 ) step 3, estimating multi-center effect parameters γ.sub.ig* and δ.sub.ig* according to the standardized features of the training center A and the test center B, and carrying out multi-center effect compensation for the feature of the test center B, so as to obtain a feature Y.sub.ijg* of the test center B after the multi-center effect compensation, which comprise that following sub-steps: (3.1) the standardized feature Z.sub.Ajg obtained in step 1 and the standardized feature Z.sub.Bjg obtained in step 2 conforming to a normal distribution, and calculating an average value {circumflex over (γ)}.sub.ig and a standard deviation {circumflex over (δ)}.sub.ig.sup.2 of the standardized features; obtaining an expected value E[γ.sub.ig] about a multi-center effect parameter γ.sub.ig by integration from a posterior distribution π(Z.sub.ig, γ.sub.ig, δ.sub.ig.sup.2) of standardized data Z.sub.ig; transforming the integral by a likelihood function according to the Bayesian theorem; γ ˆ ig = 1 n i .Math. j Z ijg ( 3 ) δ ˆ ig 2 = 1 n i - 1 .Math. j ( Z ijg - γ ˆ ig ) 2 ( 4 ) E [ γ ig ] = γ ig π ( Z ig , γ ig , δ ig 2 ) d ( γ ig , δ ig 2 ) ( 5 ) E [ γ ig ] = 1 C ( Z i g ) γ ig L ( Z ig .Math. "\[LeftBracketingBar]" γ ig , δ ig 2 ) π ( γ ig , δ ig 2 ) d ( γ ig , δ ig 2 ) ( 6 ) where n.sub.i represents a sample size of a center i; Z.sub.ijg represents the standardized features, in which i represents a center, j represents a case, and g represents a feature, C(Z.sub.ig)=∫L(Z.sub.ig, γ.sub.ig, δ.sub.ig.sup.2)π(γ.sub.ig, δ.sub.ig.sup.2)d(γ.sub.ig, δ.sub.ig.sup.2) L(Z.sub.ig|γ.sub.ig.sup.2)=Π.sub.jφ(Z.sub.ijg, γ.sub.ig, δ.sub.ig.sup.2), φ(Z.sub.ijg, γ.sub.ig, δ.sub.ig.sup.2) is a probability density function for evaluating a certain feature Z.sub.ijg about a random variable N(γ.sub.ig, δ.sub.ig.sup.2) that conforms to the normal distribution; (3.2) calculating an integral about the multi-center effect parameters by a Monte Carlo integral method, defining w.sub.ig″, =L(Z.sub.ig|{circumflex over (γ)}.sub.ig″, {circumflex over (δ)}.sub.ig″), so that an additive multi-center effect parameter γ.sub.ig*, can be estimated; using a same process to estimate a multiplicative multi-center effect δ.sub.ig*; γ ig * = E ^ ( γ ig ) = .Math. g w ig γ ˆ ig 2 .Math. g w ig ( 7 ) δ ig * = E ^ ( δ ig ) = .Math. g w ig δ ˆ ig 2 .Math. g w ig ( 8 ) (3.3) substituting the multi-center effect parameters γ.sub.Bg* and δ.sub.Bg* estimated in step (3.2) into the location-scale model to obtain the feature Y.sub.Bjg* of the test center B after the multi-center effect compensation; Y Bjg = * σ ˆ g δ ^ Bg * ( Z Bjg - γ ˆ Bg * ) + α ˆ g + X β ˆ g . ( 9 )

    Description

    BRIEF DESCRIPTION OF DRAWINGS

    [0017] FIG. 1 is a flow chart of the multi-center compensation method of the present disclosure;

    [0018] FIG. 2 is a flow chart showing the estimation of multi-center effect parameters.

    DESCRIPTION OF EMBODIMENTS

    [0019] The present disclosure will be described in detail below with reference to the attached drawings.

    [0020] FIG. 1 is a flow chart of a method for multi-center effect compensation based on a PET/CT intelligent diagnosis system of the present disclosure, which specifically includes the following steps:

    [0021] Step 1, importing PET/CT data of a training center A into a deep convolution neural network model of the training center A, obtaining a feature map of all the PET/CT data in the training center A, calculating an average value {circumflex over (α)}.sub.g, a covariant regression coefficient {circumflex over (β)}.sub.g and a variance {circumflex over (σ)}.sub.g of all features Y.sub.Ajg in the feature map of all the PET/CT data in the training center A, and carrying out Z-Score standardization thereon to obtain a standardized feature Z.sub.Ajg:

    [00006] Z Ajg = Y Ajg - α ˆ 9 - X β ˆ g σ ˆ g ( 1 )

    [0022] where Z.sub.Ajg represents a feature g extracted from a case j in the training center A, and X is a design matrix of a relevant covariates;

    [0023] Step 2, importing PET/CT data in a test center B into the deep convolution neural network model of the training center A, obtaining a feature map of all the PET/CT data in the test center B, and carrying out Z-Score standardization on the features Y.sub.Bjg in the feature map by using the average value and the variance of the training center A obtained in step 1, and obtaining a standardized feature Z.sub.Bjg:

    [00007] Z Bjg = Y Bjg - α ˆ g - X β ˆ g σ ˆ g ; ( 2 )

    [0024] Because the deep convolutional neural network model is trained based on the data of the training center A, the features of the test center B can be compensated to the level of the training center A through the above two steps.

    [0025] Step 3, estimating multi-center effect parameters γ.sub.ig* and δ.sub.ig* according to the standardized features of the training center A and the test center B, and carrying out multi-center effect compensation for the feature of the test center B, so as to obtain a feature Y.sub.ijg* of the test center B after the multi-center effect compensation, which comprise that following sub-steps:

    [0026] (3.1) the standardized feature Z.sub.Ajg obtained in step 1 and the standardized feature Z.sub.Bjg obtained in step 2 conforming to a normal distribution, and calculating an average value {circumflex over (γ)}.sub.ig and a standard deviation δ.sub.ig.sup.2 of the standardized features to be used as an observation value of the multi-center effect parameter; obtaining an expected value E[γ.sub.ig] about a multi-center effect parameter γ.sub.ig by integration from a posterior distribution π(Z.sub.ig, γ.sub.ig, δ.sub.ig.sup.2) of standardized data Z.sub.ig (formula 5); substituting the integral of formula 5 into the likelihood function to transform into formula 6 according to the Bayesian theorem, and obtaining an expected value of the calculation center parameter by the multi-center expected integral formula expressed by the likelihood function and prior probability.

    [00008] γ ˆ ig = 1 n i .Math. j Z ijg ( 3 ) δ ˆ ig 2 = 1 n i - 1 .Math. j ( Z ijg - γ ˆ ig ) 2 ( 4 ) E [ γ ig ] = γ ig π ( Z ig , γ ig , δ ig 2 ) d ( γ ig , δ ig 2 ) ( 5 ) E [ γ ig ] = 1 C ( Z i g ) γ ig L ( Z ig .Math. "\[LeftBracketingBar]" γ ig , δ ig 2 ) π ( γ ig , δ ig 2 ) d ( γ ig , δ ig 2 ) ( 6 )

    [0027] where, n.sub.i represents a sample size of a center i; Z.sub.ijg represents the standardized features, in which i represents a center, j represents a case, and g represents a feature, C(Z.sub.ig)=∫L(Z.sub.ig|γ.sub.ig, δ.sub.ig.sup.2)π(γ.sub.ig, δ.sub.ig.sup.2)d(γ.sub.ig, δ.sub.ig.sup.2) L(Z.sub.ig|γ.sub.ig, δ.sub.ig)=Π.sub.jφ(Z.sub.ig, γ.sub.ig, δ.sub.ig.sup.2, φ(Z.sub.ig, γ.sub.ig, δ.sub.ig.sup.2) is a probability density function for evaluating a certain feature Z.sub.ijg about a random variable N(γ.sub.ig, δ.sub.ig.sup.2) that conforms to the normal distribution;

    [0028] (3.2) calculating the multi-center effect parameter integral of formula 6 by a Monte Carlo integration method, defining w.sub.ig″=L(Z.sub.ig|{circumflex over (γ)}.sub.ig″, {circumflex over (δ)}.sub.ig″), so that an additive multi-center effect parameter γ.sub.ig*, can be estimated; using a same process to estimate a multiplicative multi-center effect δ.sub.ig*; a numerical integration method-Monte Carlo integration is used to estimate the integration result of formula 6, which avoids using an analytical method to calculate the integration.

    [00009] γ ig * = E ^ ( γ ig ) = .Math. g w ig γ ˆ ig 2 .Math. g w ig ( 7 ) δ ig * = E ^ ( δ ig ) = .Math. g w ig δ ˆ ig 2 .Math. g w ig ( 8 )

    [0029] (3.3) substituting the multi-center effect parameters γ.sub.Bg* and δ.sub.Bg* estimated in step (3.2) into the location-scale model to obtain the feature Y.sub.Bjg* of the test center B after the multi-center effect compensation;

    [00010] Y Bjg = * σ ˆ g δ ^ Bg * ( Z Bjg - γ ˆ Bg * ) + α ˆ g + X β ˆ g . ( 9 )

    Examples

    [0030] Taking the deep convolution neural network ResNet-152 as an example, how to apply this method for multi-center effect compensation to the deep convolution neural network will be explained.

    [0031] The network structure of ResNet-152 is shown in Table 1:

    TABLE-US-00001 TABLE 1 Resnet-152 Network Structure Layer name Detail Convolution layer 1 7 * 7, stride 2 Convolution layer 2 3 * 3 max pool, stride 2 [00011] [ 1 * 1 , 64 3 * 3 , 64 1 * 1 , 256 ] * 3 Convolution layer 3 [00012] [ 1 * 1 , 128 3 * 3 , 128 1 * 1 , 512 ] * 8 Convolution layer 4 [00013] [ 1 * 1 , 256 3 * 3 , 256 1 * 1 , 1024 ] * 36 Convolution layer 5 [00014] [ 1 * 1 , 512 3 * 3 , 512 1 * 1 , 2048 ] * 3 Mean value pooling layer + 1000-dimensional fully connected layer + softmax

    [0032] In an actual implementation process, a training center A used the PET/CT image data of this center to train an intelligent auxiliary diagnosis model of benign and malignant tumors based on ResNet-152. Now, this intelligent auxiliary diagnosis model was used in a test center B, a grass-roots hospital. However, due to the statistical differences between the imaging protocols used by the training center A and the test center B, if the model was directly used in the test center B, the diagnostic efficiency would be reduced to some extent. Therefore, the above problems were solved by the method for multi-center effect compensation of the present disclosure. The implementation steps were as follows:

    [0033] (1) All the image data used for training in the training center A were input into ResNet-152 after training, and a feature map of the last layer of a convolution layer 5 was extracted and saved as features.

    [0034] (2) All available image data of THE test center B were input into the trained ResNet-152, and the feature map of the last layer of the convolution layer 5 was extracted and saved as features.

    [0035] (3) Based on the features of the two groups obtained in step (1) and step (2), the additive and multiplicative multi-center effect parameters were estimated by using the method for multi-center effect compensation proposed by the present disclosure.

    [0036] (4) By using the multi-center effect parameters obtained in step (3), the features of the test center B were compensated for a multi-center effect.

    [0037] (5) The compensated features of the test center B were returned to the second half of ResNet-152, that is, the mean value pooling layer behind the convolution layer 5, the 1000-dimensional fully connected layer and the softmax layer, so as to obtain the classification results of benign and malignant tumors.

    [0038] After the estimated parameters of multi-center effect were saved, the steps (1) and (3) could be omitted in the subsequent diagnosis application as follow:

    [0039] (1) The image data to be tested in the test center B were input into the diagnosis model ResNet-152, and the feature map of the last layer of the convolution layer 5 was extracted as features.

    [0040] (2) By using the obtained additive and multiplicative multi-center effect parameters, the features obtained in step (1) were compensated for a multi-center effect.

    [0041] (3) The compensated features of test center B were returned to the second half of ResNet-152, that is, the mean value pooling layer behind convolution layer 5, the 1000-dimensional fully connected layer and the softmax layer, so as to obtain the classification results of benign and malignant tumors.

    [0042] Due to the multi-center effect compensation, the data of the test center B is applied to the diagnosis model of the training center A, which has a similar diagnosis efficiency to that of the training center A, and indirectly improves the generalization ability of the model.

    [0043] It should be noted that when the data compression apparatus provided in the foregoing embodiment performs data compression, division into the foregoing functional modules is used only as an example for description. In an actual application, the foregoing functions can be allocated to and implemented by different functional modules based on a requirement, that is, an inner structure of the apparatus is divided into different functional modules, to implement all or some of the functions described above. For details about a specific implementation process, refer to the method embodiment. Details are not described herein again.

    [0044] All or some of the foregoing embodiments may be implemented by using software, hardware, firmware, or any combination thereof. When the software is used for implementation, all or some of the embodiments may be implemented in a form of a computer program product. The computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on a server or a terminal, all or some of the procedures or functions according to the embodiments of this application are generated. The computer instructions may be stored in a computer-readable storage medium or may be transmitted from a computer-readable storage medium to another computer-readable storage medium. For example, the computer instructions may be transmitted from a web site, computer, server, or data center to another web site, computer, server, or data center in a wired (for example, a coaxial optical cable, an optical fiber, or a digital subscriber line) or wireless (for example, infrared, radio, or microwave) manner. The computer-readable storage medium may be any usable medium accessible by a server or a terminal, or a data storage device, such as a server or a data center, integrating one or more usable media. The usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, or a magnetic tape), an optical medium (for example, a digital video disk (DVD)), or a semiconductor medium (for example, a solid-state drive).

    [0045] The steps of the method or algorithm described combined with the embodiments of the present disclosure may be implemented in a hardware manner, or may be implemented in a manner in which a processor executes software instructions. The software instructions may consist of corresponding software modules, and the software modules can be stored in Random Access Memory (RAM), flash memory, Read Only Memory (ROM), Erasable Programmable ROM (EPROM), Electrically Erasable Programmable Read-Only Memory (EEPROM), registers, hard disks, removable hard disks, CD-ROMs or any other forms of storage media well-known in the art. An exemplary storage medium is coupled to the processor, such that the processor can read information from, and write information to, the storage medium. The storage medium can also be an integral part of the processor. The processor and storage medium may reside in an Application Specific Integrated Circuit (ASIC). Alternatively, the ASIC may be located in a node device, such as the processing node described above. In addition, the processor and storage medium may also exist in the node device as discrete components.