INFORMATION ENHANCING METHOD AND INFORMATION ENHANCING SYSTEM

20240054183 ยท 2024-02-15

    Inventors

    Cpc classification

    International classification

    Abstract

    Disclosed are an information enhancing method and an information enhancing system. The information enhancing method includes: sampling information to obtain a multi-view dataset labelled with feature and class; creating a fix function to represent quantity of fixes; creating a view sub-classifier to represent quality of fixes; unifying the quantity of fixes and the quality of fixes to create a quantity-quality balance model, and resolving the quantity-quality balance model to obtain a fixed multi-view dataset; computing weight of each view and weight of the feature of the fixed information; computing information entropy of a fixed labeled sample based on the weight of the view and the weight of the feature; and selecting a labeled sample based on the information entropy and the weights according to a selected generation manner to generate an unlabeled sample, thereby augmenting the sampled information and realizing information enhancement. By fixing and augmenting the sampled information, the disclosure effectively enhances the sampled information and improves application system performance, thereby offering a better guide to system design.

    Claims

    1. An information enhancing method, comprising steps of: sampling information to obtain a multi-view dataset labelled with feature and class; creating a fix function to represent quantity of fixes; creating a view sub-classifier to represent quality of fixes; unifying the quantity of fixes and the quality of fixes to create a quantity-quality balance model, and resolving the quantity-quality balance model to obtain a fixed multi-view dataset; computing weight of each view and weight of each feature of the fixed information; computing information entropy of a fixed labeled sample based on the weight of the view and the weight of the feature; and selecting a labeled sample based on the information entropy and the weights according to a selected generation manner to generate an unlabeled sample, thereby augmenting the sampled information and realizing information enhancement.

    2. The information enhancing method according to claim 1, wherein the fix function is:
    h(Z.sub.j?U.sub.jV.sub.j); where Z.sub.j denotes a hypothetical low-rank matrix, and the hypothetical low-rank matrix Z.sub.j corresponding to the feature information X.sub.j of each view is decomposed into a latent representation form U.sub.j and a coefficient matrix V.sub.j of the feature information, wherein U.sub.jV.sub.j denotes the fixed feature information.

    3. The information enhancing method according to claim 2, wherein the view sub-classifier is:
    g(S.sub.j,W.sub.j,V.sub.j,U.sub.j,Y.sub.j)=g(g(U.sub.jV.sub.j,W.sub.j)?Y.sub.jS.sub.j); where g(U.sub.jV.sub.j, W.sub.j) represents mapping U.sub.jV.sub.j to a corresponding predicted class using a mapping matrix W.sub.j, Y.sub.j denotes the class of each view, and S.sub.j is a coefficient matrix of classes.

    4. The information enhancing method according to claim 3, wherein an objective optimization function is formed using a metric function, and most values of the objective optimization function are resolved to form the quantity-quality balance model; the metric function is:
    ?(h,g)=?(h(Z.sub.j?U.sub.jV.sub.j)/g(S.sub.j,W.sub.j,V.sub.j,U.sub.j,Y.sub.j)) the objective function is f ( ) and the quantity-quality balance model is: f ( h , g , ? ) = min .Math. j = 1 m f ( h ( Z j - U j V j ) , g ( S j , W j , V j , U j , Y j ) , ? ( h ( Z j - U j V j ) / g ( S j , W j , V j , U j , Y j ) ) ) where m denotes the number of views.

    5. The information enhancing method according to claim 4, wherein the quantity-quality balance model is resolved using alternating minimization to obtain optimized form U.sub.j.sup.o of the latent representation form U.sub.j and optimized form V.sub.j.sup.o of the coefficient matrix V.sub.j of each view, wherein the information of each view through X.sub.j.sup.o=U.sub.j.sup.oV.sub.j.sup.o, a fixed multi-view data set.

    6. The information enhancing method according to claim 5, wherein weight ?.sub.j of each view and corresponding feature weight vector ?.sub.j are obtained using a multi-view clustering algorithm; each feature weight vector is ?.sub.j={?.sub.j1, . . . , ?.sub.jc, . . . , ?.sub.jd.sub.j}, where d.sub.j denotes the number of features of the view, and ?.sub.jc denotes the weight of the c.sup.th feature of the view.

    7. The information enhancing method according to claim 6, wherein the information entropy H.sub.l of each fixed labeled sample x.sub.l is computed using a distance weighted method.

    8. The information enhancing method according to claim 7, wherein an unlabeled sample x.sub.u nearest to or farthest from the labelled sample is selected to generate a Universum sample u.sub.l?u;
    custom-character(?.sub.1, . . . ,?.sub.j, . . . ,?.sub.m, . . . ,?.sub.1, . . . ,?.sub.j, . . . ,?.sub.m,x.sub.l, x.sub.u) where the generated Universum sample u.sub.l?u and the fixed multi-view dataset are unified into an information enhanced dataset.

    9. A memory, wherein a plurality of instructions are stored in the memory, the instructions being loadable and executable by a processor, the instructions including the information enhancing method according to claim 1.

    10. An information enhancing system, comprising: a processor, the memory according to claim 9, and a plurality of cameras; wherein the cameras are configured to sample information to obtain a multi-view dataset labelled with feature and class; the memory is configured to store instructions; and the processor is configured to load and execute the instructions in the memory.

    Description

    BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS

    [0032] FIG. 1 is a flow diagram of an information enhancing method based on a quantity-quality balance model and information entropy according to the present disclosure.

    [0033] FIG. 2 is a flow diagram of an information enhancing method based on a quantity-quality balance model and information entropy in an embodiment of the present disclosure.

    DETAILED DESCRIPTION

    [0034] Hereinafter, preferred embodiments of the present disclosure will be illustrated in detail with reference to FIGS. 1?2.

    [0035] As shown in FIG. 1, the present disclosure provides an information enhancing method based on a quantity-quality balance model and information entropy, comprising steps of: [0036] Step S1: sampling information to obtain a multi-view dataset labelled with sample feature X and class label Y; [0037] Step S2: decomposing a hypothetical low-rank matrix corresponding to feature information of each view into a latent representation form and a coefficient matrix of feature information, and creating a fix function to represent quantity of fixes; [0038] creating a view sub-classifier to represent quality of fixes; [0039] Step S3: unifying the quantity of fixes and the quality of fixes to build a quantity-quality balance model to ensure validity of the fixed information, wherein the quantity-quality balance model is usually resolved using alternating minimization, thereby realizing fix of missed information; [0040] Step S4: computing weight of each view and weight of each feature of the fixed information using a multi-view clustering algorithm;

    [0041] Step S5: computing information entropy of a fixed labeled sample based on the weight of the view and the weight of the feature to ensure validity of subsequent augmented information; and [0042] Step S6: selecting a high-certainty labeled sample based on the information entropy, weights, and a selected generation method to generate an appropriate unlabeled sample, thereby augmenting the sampled information and finally realizing information enhancement.

    [0043] As illustrated in FIG. 2, in an embodiment of the present disclosure, the information enhancing method based on a quantity-quality balance model and information entropy is implemented using an information sampling portion, an information fixing portion, and an information augmenting portion. The information sampling portion is configured to obtain an original multi-view dataset using a plurality of cameras, wherein the cameras refer to Hikvision ColorVu bullet network cameras, model #DS-2CD2T27F(D)WD-LS 2 mega-pixel 1/2.7 CMOS; the information fixing portion includes a quantity-quality balance model design submodule and an information fixing submodule, wherein the information fixing portion adopts a discrepancy ratio as a core to build a quantity-quality balance model and resolve the model using alternating minimization; the information augmenting portion includes a multi-view clustering algorithm submodule, an information entropy analyzing submodule, and a Universum sample selecting and generating submodule, wherein the information augmenting portion adopts a Universum sample generation algorithm with information entropy as the core.

    [0044] Further, the information enhancing method based on a quantity-quality balance model and an information entropy in this embodiment comprises: [0045] Step 1: capturing, by cameras, a series of samples, and manually labeling some of the samples, wherein corresponding sample feature is denoted as X, corresponding class label is denoted as Y, and for an unlabeled sample, its class label may be denoted as 0.

    [0046] Step 2: decomposing hypothetical low-rank matrix Z.sub.j corresponding to feature information X.sub.j of each view (hypothetically the j.sup.th view) into a latent representation form U.sub.j and a coefficient matrix V.sub.j of the feature information X.sub.j, wherein U.sub.jV.sub.j denotes the fixed feature information, and then the fix function expression h(Z.sub.j?U.sub.jV.sub.j) denotes the quantity of fixes, where the smaller the value, the more the information to be fixed.

    [0047] Step 3: for the fixed information U.sub.jV.sub.j, with map matrix W.sub.j as a bridge and S.sub.j representing coefficient matrix of classes, designing, with reference to the manner of mapping feature information X.sup.t to class information Y.sup.t by weight in the conventional pattern recognition field

    [00002] ( i . e . , X t .fwdarw. W t Y t ) ,

    respective view sub-classifiers to measure impact of the fixed information on the performance of the multi-view learning algorithm, wherein the impact denotes quality of fixes, wherein the smaller the value, the greater the fixed information enhances the performance of the multi-view learning algorithm.

    [0048] The view sub-classifier prefers to:


    g(S.sub.j,W.sub.j,V.sub.j,U.sub.j,Y.sub.j)=g(g(U.sub.jV.sub.j,W.sub.j)?Y.sub.jS.sub.j); [0049] where g(U.sub.jV.sub.j, W.sub.j) denotes mapping U.sub.jV.sub.j to a corresponding predicted class by W.sub.j; in actual applications, letting Y denote a class matrix, then S=Y?Y, i.e., representing the coefficient matrix of classes using similarities between classes.

    [0050] Step 4: forming an objective optimization function ? by unifying the quantity and quality portions of respective views and taking the relation between the quantity and quality portions as well as the balance metric into consideration by introducing a metric function ?, ?(h, g)=?(h(Z.sub.j?U.sub.jV.sub.j)/g(S.sub.j, W.sub.j, V.sub.j, U.sub.j, Y.sub.j)), and constructing most values of the objective optimization function ? so as to form a quantity-quality balance model, ?(h, g, ?)=min?.sub.j?1.sup.m?(h(Z.sub.j?U.sub.jV.sub.j)/g(S.sub.j, W.sub.j, V.sub.j, U.sub.j, Y.sub.j), ?(h(Z.sub.j?U.sub.jV.sub.j)/g(S.sub.j, W.sub.j, V.sub.j, U.sub.j, Y.sub.j)), [0051] where m denotes the number of views.

    [0052] The metric function ? is designed with discrepancy ratio as the core. Specifically, h(Z.sub.j?U.sub.jV.sub.j) denotes the quantity of fixes, where the smaller its outcome, the more the information to be fixed; while g(S.sub.j, W.sub.j, V.sub.j, U.sub.j, Y.sub.j) denotes the quality of fixes, where the smaller its outcome, the greater the fixed information enhances the performance of the multi-view learning algorithm. During the fix process, in order to prevent weighing too heavily on either quantity or quality, the metric function ?(h(Z.sub.j?U.sub.jV.sub.j)/g(S.sub.j, W.sub.j, V.sub.j, U.sub.j, Y.sub.j)) is introduced, where the function reflects a ratio (i.e., discrepancy ratio) between respective discrepancy measurement results with respect to quantity and quality. If the outcome of metric function a is greater than 1, it indicates that the fix process weighs more on quality; otherwise, the fix process weighs more on quantity; if the outcome of the metric function ? is equal to 1, it indicates that the quantity and the quality reach a balance. Therefore, with the discrepancy ratio and by introducing the metric function ?, the relationship between quantity and quality may be reflected by the outcome of the metric function ?. Additionally, since it is hard to reach exact 1 of the metric function value in actual scenarios; the range of the metric function value may be usually defined to be approximately 1 when designing the quantity-quality balance model, which may be regarded as reaching an equilibrium between quantity and quality. With the discrepancy ratio, the relation between quantity and quality and the balanced metric problem may be effectively resolved, and thus the missed information may be better fixed.

    [0053] Step 5: optimizing and resolving, by an information fixing submodule, the objective optimization function through alternating minimization to obtain optimizations (i.e., U.sub.j.sup.o and V.sub.j.sup.o) of the latent representation form U.sub.j and the coefficient matrix V.sub.j of respective views; then, fixing information of each view with X.sub.j.sup.o=U.sub.j.sup.oV.sub.j.sup.o to obtain a fixed multi-view dataset.

    [0054] Step 6: for the fixed multi-view dataset, analyzing, by a multi-view clustering submodule, contributions and impacts of different views and their feature information with respect to the multi-view clustering algorithm, to obtain the weight ?.sub.j of each view and corresponding feature weight vector ?.sub.j.

    [0055] Each feature weight vector may be written as ?.sub.j={?.sub.j1, . . . , ?.sub.jc, . . . , ?.sub.jd.sub.j}, where d.sub.j denotes the number of features of the view, and ?.sub.jc denotes the weight of the c.sup.th feature of the view.

    [0056] The feature weight refers to the weight of a feature, and the feature weight vector refers to a vector formed by unification of the weights of a plurality of features under one view.

    [0057] Step 7: computing and finding a plurality of neighbor samples near each fixed labeled sample x.sub.l using a weighed distance method based on the view weight and the feature weight vector, and obtaining, by an information entropy analyzing submodule, the information entropy H.sub.l of the labeled sample based on the class of the neighbor samples according to an information entropy computing equation H.

    [0058] The information entropy may reflect class decision certainty of the labeled sample, where a higher certainty indicates a higher validity of a Universum sample generated using priori knowledge of the labeled sample and may enhance class decision capability of the algorithm.

    [0059] Step 8: first selecting, by a Universum sample selecting and generating submodule, a high-certainty labeled sample x.sub.l based on the information entropy H.sub.l, and then selecting a corresponding unlabeled sample x.sub.u based on a selected generating manner (e.g., generating the Universum sample by computing and selecting an unlabeled sample closest to or farthest from the labeled sample using the distance weighted method), and generating a corresponding Universum sample u.sub.l?u according to a function expression custom-character(?.sub.1, . . . , ?.sub.j, . . . , ?.sub.m, . . . , ?.sub.1, . . . , ?.sub.j, . . . , ?.sub.m, x.sub.l, x.sub.u).

    [0060] Finally, the generated Universum samples u.sub.l?u and the fixed multi-view dataset in step 5 are unified into an information enhanced dataset.

    [0061] By fixing and augmenting the sampled information, the present disclosure effectively enhances the sampled information and improve application system performance, so as to offer a better guide to system design.

    [0062] Although the contents of the present disclosure have been described in detail through the foregoing preferred embodiments, it should be understood that the depictions above shall not be regarded as limitations to the present disclosure. After those skilled in the art having read the contents above, many modifications and substitutions to the present disclosure are all obvious. Therefore, the protection scope of the present disclosure should be limited by the appended claims.