LOYALTY EXTRACTION MACHINE

20220343204 · 2022-10-27

    Inventors

    Cpc classification

    International classification

    Abstract

    The present invention provides a loyalty extraction machine, wherein “quadratic multiform separation” (QMS) is modified and executed multiplicatively in an even generalized way. In each execution, the characteristic of one single membership is either enhanced or reduced. This process is performed in turn to each membership. Thus, every sample data (or element) receives multiple classification results. Then, the multiple classification results are collected and analyzed by an “eclectic classifier” to reach a final decision. The combination of the generalized QMS and the eclectic classifier therefore develops the loyalty extraction machine. Moreover, a label called “loyalty type” of the element is introduced to describe the effectiveness of membership recognition with respect to a training set.

    Claims

    1. A loyalty extraction machine, comprising: an input module configured to receive sample data (x); a data collection module connected to the input module and configured to store a collection of data (Ω) from the input module, the collection of data (Ω) including a training set (Ω.sub.tr) and/or a test set (Ω.sub.tt); a multiform separation engine connected to the data collection module and configured to generate 2m developed classifiers (ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m), m≥2; a classifier combination module connected to the multiform separation engine and the data collection module and configured to combine the 2m developed classifiers (ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m), the developed classifiers (ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m) being trained with the training set (Ω.sub.tr); and an output module connected to the classifier combination module and configured to derive an output result after the sample data (x) is processed through the classifier combination module.

    2. The loyalty extraction machine of claim 1, wherein the 2m developed classifiers are combined to form a vector function V(x)=(ŷ.sup.β,1(x), . . . , ŷ.sup.β,m(x), ŷ.sup.γ,1(x), . . . , ŷ.sup.γ,m(x)), wherein an outcome of the vector function (V) is a 2m-dimensional vector and named an identity (I).

    3. The loyalty extraction machine of claim 2, wherein, for generating the 2m developed classifiers (ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m), the multiform separation engine is configured to use m piecewise continuous functions to perform classification; the m piecewise continuous functions are respectively trained with the training set Ω.sub.tr through a training process.

    4. The loyalty extraction machine of claim 3, wherein, for a developed classifier, the multiform separation engine is configured to derive temporary evaluations f.sub.1(x), . . . , f.sub.m(x) from the m trained piecewise continuous functions f.sub.1, . . . , f.sub.m for a certain sample data x, and the sample data x is assigned to have a membership j by finding out a specific evaluation f.sub.j(x) after processing the temporary evaluations f.sub.1(x), . . . , f.sub.m(x) according to a specific operator.

    5. The loyalty extraction machine of claim 4, wherein the specific operator is a minimum operator.

    6. The loyalty extraction machine of claim 3, wherein each piecewise continuous function is a linear function, a quadratic function, a quartic function, a polynomial function, a rational function, an algebraic function, a transcendental function, or any other explicitly or implicitly defined suitable function.

    7. The loyalty extraction machine of claim 6, wherein each piecewise continuous function f.sub.j(x) is a q-dimensional member function having a form ∥A.sub.jx−b.sub.j∥.sup.2, for an integer q, a constant matrix A.sub.j, and a constant vector b.sub.j, j=1, . . . , m, and the m piecewise continuous functions become m member functions.

    8. The loyalty extraction machine of claim 7, wherein in the training process, m.sup.2 control parameters α.sub.jk, j=1, . . . , m, k=1, . . . , m are set between 0 and 1 so as to participate comparisons among the m member functions f.sub.1, . . . , f.sub.m, and the comparisons are performed according to intermediate functions φ.sub.jk:Ω.fwdarw.custom-character, j=1, . . . , m, k=1, . . . , m, defined by φ jk ( x ) = max { α jk , f j ( x ) f k ( x ) } .

    9. The loyalty extraction machine of claim 8, wherein in the training process, given the integer q, the m.sup.2 control parameters α.sub.jk, j=1, . . . , m, k=1, . . . , m, and the training set Ω.sub.tr, m weighted cost functions Φ.sup.β,1, . . . , Φ.sup.β,m with a lighter weight β where 0<β<1 and m weighted cost functions Φ.sup.γ,1, . . . , Φ.sup.γ,m with a heavier weight γ where γ>1 are defined depending on constant matrices A.sub.1, . . . , A.sub.m and constant vectors b.sub.1, . . . , b.sub.m that form the member functions f.sub.1, . . . , f.sub.m.

    10. The loyalty extraction machine of claim 9, wherein a cost c(j) contributed by a j-th training subset Ω.sub.tr(j) is defined by c ( j ) = c ( j ; f 1 , .Math. , f m ) = .Math. x Ω tr ( j ) .Math. k S , k j φ jk ( x ) , where S={1, 2, . . . , m} is a set of memberships.

    11. The loyalty extraction machine of claim 10, wherein weighted cost functions Φ.sup.β,j(f.sub.1, . . . , f.sub.m) are defined by Φ.sup.β,j(f.sub.1, . . . , f.sub.m)=c(1)+ . . . +c(j−1)+c(j)+c(j+1)+ . . . +c(m); and the training process is configured to solve min.sub.f.sub.1.sub., . . . , f.sub.m.sub.∈Θ(q)Φ.sup.β,j(f.sub.1, . . . , f.sub.m) in order to derive member functions f.sub.1.sup.β,j, . . . , f.sub.m.sup.β,j, j=1, . . . , m, where Θ(q) is a set of all member functions.

    12. The loyalty extraction machine of claim 11, wherein weighted cost functions Φ.sup.γ,j(f.sub.1, . . . , f.sub.m) are defined by Φ.sup.γ,j(f.sub.1, . . . , f.sub.m)=c(1)+ . . . +c(j−1)+γ.Math.c(j)+c(j+1)+ . . . +c(m); and the training process is configured to solve min.sub.f.sub.1.sub., . . . , f.sub.m.sub.∈Θ(q)Φ.sup.γ,j(f.sub.1, . . . , f.sub.m) in order to derive member functions f.sub.1.sup.γ,j, . . . , f.sub.m.sup.β,j, j=1, . . . , m, where Θ(q) is a set of all member functions.

    13. The loyalty extraction machine of claim 2, further comprising: a bucket creation module connected to the classifier combination module and configured to partition the training set (Ω.sub.tr) into a disjoint union of subsets, which are called buckets and denoted by B(I), wherein the respective buckets (B(I)) have respective identities (I) associated with characteristics of the data; and a bucket merger module connected to the bucket creation module and configured to merge empty buckets and/or small buckets into large buckets.

    14. The loyalty extraction machine of claim 13, wherein the bucket creation module is further configured to denote the cardinality of a subset of the bucket (B(I)) with membership (j) by (n.sub.B(I)(j)) and the cardinality of a subset of the training set (Ω.sub.tr) with membership (j) by (n.sub.tr(j)), and to perform merger such that the merged bucket (B) is sufficiently large that the condition max j { n B ( j ) n tr ( j ) } δ holds for a certain predetermined positive constant (δ).

    15. The loyalty extraction machine of claim 14, further comprising a membership assignment module connected to the bucket creation module and configured to assign respective memberships (j's) to the respective buckets (B(I)) if a ratio of the cardinality of sample data (x) with the membership (j) in a bucket (B(I)) to the cardinality of a subset (Ω.sub.tr(j)) of the training set (Ω.sub.tr) with the membership (j) is maximal among ratios of all memberships, the memberships (j's) referring to data categories of the training set (Ω.sub.tr); wherein in this assignment, Y(B(I))=j.

    16. The loyalty extraction machine of claim 15, wherein the membership of sample data (x) in the collection of data (Ω) is also the membership of the bucket (B(I)) to which the sample data (x) is distributed.

    17. The loyalty extraction machine of claim 2, further comprising a loyalty type indicator configured to determine loyalty type of a sample data (x) by confirming a location of a lighter weight (β) and/or a heavier weight (γ) in an identity (I).

    18. The loyalty extraction machine of claim 17, wherein the loyalty type indicator is configured to denote each identity (I) as a vector with 2m coordinates, for k=1, . . . , m, an identity (I) associated with a lighter weight (β) has a form of I ( β , k ) = ( * , .Math. , * , k the k - th coordinate , * , .Math. , * , the first m coordinates * , .Math. , * the second m coordinates ) ; and an identity (I) associated with a heavier weight (γ) has a form of I ( γ , k ) = ( * , .Math. , * , the first m coordinates * , .Math. , * , k the ( m + k ) - th coordinate , * , .Math. , * , the second m coordinates ) .

    19. The loyalty extraction machine of claim 18, wherein the loyalty type indicator is configured to determine that a sample data (x) has a membership (k) with strong loyalty if the vector function V(x)=I(β,k) and the eclectic classifier {tilde over (y)}(x)=k=Y(B(I(β,k))); or a sample data (x) has a membership (k) with weak loyalty if the sample data (x) is not strongly loyal to any membership, and the vector function V(x)=I(γ,k), and the eclectic classifier {tilde over (y)}(x)=k=Y(B(I(γ,k))); or an element has normal loyalty if it has neither strong loyalty nor weak loyalty.

    20. The loyalty extraction machine of claim 17, wherein the complete loyalty extraction machine is expressed by a single mapping
    ζ(x)=({tilde over (y)}(x),τ(x))∈S×T, x∈Ω, wherein a first component {tilde over (y)}(x) of the single mapping ζ(x) indicates the membership of the sample data (x), and a second component τ(x) of the single mapping ζ(x) indicates the loyalty type of the sample data (x).

    21. The loyalty extraction machine of claim 1, wherein it is implemented as hardware or software.

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    [0122] FIG. 1 shows a schematic diagram of a prior art classifier;

    [0123] FIG. 2 shows a schematic diagram of classification accomplished by one single inference function of a prior art classifier;

    [0124] FIG. 3 shows a problem that the prior art classifier may face;

    [0125] FIG. 4 shows a schematic block diagram of the loyalty extraction machine according to one embodiment of the present invention; and

    [0126] FIG. 5 shows a detailed block diagram of the (quadratic) multiform separation engine according to one embodiment of the present invention.

    DETAILED DESCRIPTION OF THE EMBODIMENT

    [0127] Different embodiments of the present invention are provided in the following description. These embodiments are meant to explain the technical content of the present invention, but not meant to limit the scope of the present invention. A feature described in an embodiment may be applied to other embodiments by suitable modification, substitution, combination, or separation.

    [0128] It should be noted that, in the present specification, when a component is described to have an element, it means that the component may have one or more of the elements, and it does not mean that the component has only one of the element, except otherwise specified.

    [0129] Moreover, in the present specification, the ordinal numbers, such as “first” or “second”, are used to distinguish a plurality of elements having the same name, and it does not means that there is essentially a level, a rank, an executing order, or a manufacturing order among the elements, except otherwise specified. A “first” element and a “second” element may exist together in the same component, or alternatively, they may exist in different components, respectively. The existence of an element described by a greater ordinal number does not essentially mean the existent of another element described by a smaller ordinal number.

    [0130] Moreover, in the present specification, the terms, such as “preferably” or “advantageously”, are used to describe an optional or additional element or feature, and in other words, the element or the feature is not an essential element, and may be ignored in some embodiments.

    [0131] Moreover, each component may be realized as a single circuit or an integrated circuit in suitable ways, and may include one or more active elements, such as transistors or logic gates, or one or more passive elements, such as resistors, capacitors, or inductors, but not limited thereto. Each component may be connected to each other in suitable ways, for example, by using one or more traces to form series connection or parallel connection, especially to satisfy the requirements of input terminal and output terminal. Furthermore, each component may allow transmitting or receiving input signals or output signals in sequence or in parallel. The aforementioned configurations may be realized depending on practical applications.

    [0132] Moreover, in the present specification, the terms, such as “system” “apparatus”, “device”, “module”, or “unit”, refer to an electronic element, or a digital circuit, an analogous circuit, or other general circuit, composed of a plurality of electronic elements, and there is not essentially a level or a rank among the aforementioned terms, except otherwise specified.

    [0133] Moreover, in the present specification, two elements may be electrically connected to each other directly or indirectly, except otherwise specified. In an indirect connection, one or more elements may exist between the two elements.

    [0134] FIG. 4 shows a schematic block diagram of the loyalty extraction machine 1 according to one embodiment of the present invention.

    [0135] The loyalty extraction machine 1 of the present invention is implemented by two parts, (quadratic) multiform separation and an eclectic classifier.

    [0136] (Quadratic Multiform Separation with Weights)

    [0137] As shown in FIG. 4, the loyalty extraction machine 1 of the present invention, provided in the context of machine learning, includes an input module 10, a data collection module 20, a (quadratic) multiform separation engine 80, a classifier combination module 30, a bucket creation module 40, a bucket merger module 50, a membership assignment module 60, and an output module 70.

    [0138] It can be understood that the modules or engines are illustrated here for the purpose of explaining the present invention, and the modules or engines may be integrated or separated into other forms as hardware or software in separated circuit devices on a set of chips or an integrated circuit device on a single chip. The royalty extraction machine 1 may be implemented in a cloud server or a local computer. The modules or engines of the present invention may be suitably converted into several steps of a method, and several steps of a method may be suitably converted into several modules or engines as well. The modules, engines, and steps of the present invention provide their respective functions of data processing that realize and/or optimize the algorithm of the present invention.

    [0139] The input module 10 is configured to receive sample data (or an element) x. The input module 10 may be a sensor, a camera, a speaker, and so on, that can detect physical phenomena, or it may be a data receiver.

    [0140] The data collection module 20 is connected to the input module and configured to store a collection of data Ω from the input module 10. The collection of data Ω⊂custom-character.sup.p includes a training set Ω.sub.tr and/or a test set Ω.sub.tt and/or a remaining set Ω.sub.th. Here custom-character is the set of real numbers and the expression Ω⊂custom-character.sup.p means that the collection of data Ω belongs to custom-character.sup.p, the space of p-dimensional real vectors. The collection of data Ω may also be regarded as a data structure.

    [0141] With supervised approach, a membership function y:Ω.fwdarw.S={1,2, . . . , m} can be found so that y(x) gives precisely the membership of the input data x. Accordingly, the collection of data Ω is composed of m memberships (or data categories), and the m memberships are digitized as 1, 2, . . . , m. To specifically explain the meaning of the data categories, for example, when a classifier is used to recognize animal pictures, membership “1” may indicate “dog”, membership “2” may indicate “cat”, . . . , and membership “m” may indicate “rabbit”. Herein, “dog”, “cat”, and “rabbit” are regarded as the data categories. For another example, when a classifier is used to recognize people's age by their faces, membership “1” may indicate “child”, membership “2” may indicate “teenage”, . . . , and membership “m” may indicate “adult”. Herein, “child”, “teenage”, and “adult” are regarded as the data categories.

    [0142] FIG. 5 shows a detailed block diagram of the (quadratic) multiform separation engine 80 according to one embodiment of the present invention.

    [0143] The (quadratic) multiform separation engine 80 includes a member function collector 82 and a member function trainer 84. The (quadratic) multiform separation engine 80 is connected to the data collection module 20 and configured to use m piecewise continuous functions f.sub.1, f.sub.2, . . . , f.sub.m to perform classification. The m piecewise continuous functions f.sub.1, f.sub.2, . . . , fin typically handle the same type of data, for example, they all handle image files for image recognition, all handle audio files for sound recognition, and so on, so that they can work consistently.

    [0144] The classification involves two stages (or modes): a training (or learning) stage and a prediction (or decision) stage.

    [0145] Loosely speaking, a function h:custom-character.sup.p.fwdarw.custom-character is called a piecewise continuous function if there exist finite disjoint subsets D.sub.1, . . . , D.sub.w such that D.sub.1 ∪ . . . ∪D.sub.w=custom-character.sup.p, and f is continuous on the interior of D.sub.j, j=1, . . . , w.

    [0146] In this embodiment, each piecewise continuous function f.sub.j(x) is set to be a quadratic function of the sample data x. In particular, let q∈custom-character be given, where N represents the set of natural numbers. A function f:custom-character.sup.p.fwdarw.custom-character is called a q-dimensional member function if it is of the form


    f(x)=∥Ax−b∥.sup.2

    for a constant matrix A∈custom-character.sup.q×p and a constant vector b∈custom-character.sup.q, where ∥⋅∥ denotes the Euclidean norm. In particular,

    [00014] A = ( a 11 .Math. a 1 p .Math. .Math. a q 1 .Math. a qp ) x = ( x 1 .Math. x p ) b = ( b 1 .Math. b q )

    [0147] Fix an integer q that is sufficiently large. Generate m q-dimensional member functions f.sub.j:custom-character.sup.p.fwdarw.custom-character, j=1, . . . , m, based on the training sets Ω.sub.tr. As will be discussed later, the constant matrices A.sub.1, . . . , A.sub.m and the constant vectors b.sub.1, . . . , b.sub.m of the m q-dimensional member functions are items to be solved.

    [0148] Accordingly, the member function collector 82 of the (quadratic) multiform separation engine 80 is configured to store a set of member functions, denoted by Θ(q). That is, Θ(q)={∥Ax−b∥.sup.2:constant matrix A∈custom-character.sup.q×p, constant vector b∈custom-character.sup.q}.

    [0149] The member function trainer 84 of the (quadratic) multiform separation engine 80 is configured to perform the training process.

    [0150] According to the present invention, in the training process, m.sup.2 control parameters α.sub.jk, j=1, . . . , m, k=1, . . . , m are set to participate comparisons among the m member functions, and the comparisons are performed according to a specific operator. Preferably, the m.sup.2 control parameters α.sub.jk, j=1, . . . , m, k=1, . . . , m are set between 0 and 1, and they are not necessarily distinct.

    [0151] With the m q-dimensional member functions f.sub.1, . . . , f.sub.m, and the m.sup.2 control parameters α.sub.jk, j=1, . . . , m, k=1, . . . , m, intermediate functions φ.sub.jk:Ω.fwdarw.custom-character, j=1, . . . , m, k=1, . . . , m, are defined by

    [00015] φ jk ( x ) = max { α jk , f j ( x ) f k ( x ) } ,

    [0152] Obviously, φ.sub.jk(x)<1 if and only if f.sub.j(x)<f.sub.k(x), j∈S, k∈S, k≠*j. It is noted that S={1, 2, . . . , m} is the set of memberships.

    [0153] The training process will be more efficient with the introduction of the control parameters.

    [0154] It is to be understood that the goal of the training process according to the present invention is to match the property “x has membership j” for j=1, . . . , m, with the algebraic relations φ.sub.jk(x)<1, k∈S, k≠j.

    [0155] In order to reach the goal, for j∈S, define

    [00016] c ( j ) = c ( j ; f 1 , .Math. , f m ) = .Math. x Ω tr ( j ) .Math. k S , k j φ jk ( x )

    where c(j) represents the cost contributed by Ω.sub.tr(j), j∈S.

    [0156] Now we explain the case of membership 1 carefully. In this case, a lighter weight β is applied to c(1). The rest cases can be carried out similarly and therefore will be skipped. In those cases, the lighter weight β is applied alternatively to c(2), . . . , c(m).

    [0157] Fix a suitable constant β∈(0,1), that is, 0<β<1, to be a lighter weight and construct a weighted cost function in which the lighter weight β is placed on the cost contributed by Ω.sub.tr(1):


    β.Math.c(1)+c(2)+ . . . +c(m)=Φ.sup.β,1(f.sub.1, . . . ,f.sub.m).

    [0158] The quantity of the weighted cost function Φ.sup.β,1 provides a performance measure for separating the training subsets Ω.sub.tr(1), . . . . , Ω.sub.tr(m), by the given member functions f.sub.1, . . . , f.sub.m, while the influence by the elements of membership 1 is weakened by a lighter weight β∈(0,1).

    [0159] With the integer q, the m.sup.2 control parameters α.sub.jk, j=1, . . . , m, k=1, . . . , m, the lighter weight β, and the training set Ω.sub.tr given, and q sufficiently large, the weighted cost function Φ.sup.β,1 defined above therefore depends only on the constant matrices A.sub.1, . . . , A.sub.m and the constant vectors b.sub.1, . . . , b.sub.m that define the member functions f.sub.1, . . . , f.sub.m.

    [0160] The minimizers f.sub.1.sup.β,1, . . . , f.sub.m.sup.β,1 are generated by solving

    [00017] min f 1 , .Math. , f m Θ ( q ) Φ β , 1 ( f 1 , .Math. , f m ) = min f 1 , .Math. , f m Θ ( q ) { β .Math. c ( 1 ; f 1 , .Math. , f m ) + .Math. k = 2 m c ( k ; f 1 , .Math. , f m ) }

    [0161] The minimizers f.sub.1.sup.β,1, . . . , f.sub.m.sup.β,1 are the objectives pursued in the training process performed by the member function trainer 84. They are used to construct the classifier ŷ.sup.β,1 of the present invention.

    [0162] As shown in FIG. 5, with the lighter weight β∈(0,1) placed on c(1), we can obtain the developed classifier ŷ.sup.β,1. Similarly, with β placed on c(2), . . . , c(m), we can correspondingly obtain developed classifiers ŷ.sup.β,2, . . . , ŷ.sup.β,m.

    [0163] On the other hand, fix a suitable constant γ>1 to be a heavier weight and consider sums of costs with the heavier weight γ placed on the respective costs contributed by Ω.sub.tr(j), j∈S, these weighted cost functions are denoted as


    Φ.sup.γ,j(f.sub.1, . . . ,f.sub.m)=c(1)+ . . . +c(j-1)+γ.Math.c(j)+c(j+1)+ . . . +c(m), j∈S.

    [0164] By redoing everything similarly as described above including solving minimizers and generating their corresponding classifiers, we have m more classifiers ŷ.sup.γ,1, ŷ.sup.γ,2, . . . , ŷ.sup.γ,m.

    [0165] In summary, 2m classifiers ŷ.sup.β,1, ŷ.sup.β,2, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, ŷ.sup.γ,2, . . . , ŷ.sup.γ,m have been developed by the quadratic multiform separation with weights of the present invention.

    [0166] (Eclectic Classifier)

    [0167] Referring back to FIG. 4, in the loyalty extraction machine 1, the classifier combination module 30 is connected to the (quadratic) multiform separation engine 80 and the data collection module 20 and configured to combine 2m developed classifiers ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m, m≥2, trained with the training set Ω.sub.tr, wherein 2m is the number of developed classifiers. Each of the developed classifiers ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m.

    [0168] However, it is not necessary that all of the 2m developed classifiers in the loyalty extraction machine 1 come from the aforementioned quadratic multiform separation with weights. In the creation of buckets, we may add or remove any one or more particular developed classifiers. One or more developed classifiers may employ one model from convolutional neural network (CNN), recurrent neural network (RNN), long short-term memory (LSTM) network, YOLO, ResNet, ResNet-18, ResNet-34, Vgg16, GoogleNet, Lenet, MobileNet, decision trees, or support vector machine (SVM), but not limited thereto.

    [0169] The developed classifiers ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m should be adjusted or trained to have different architectures (regarding the number of neurons, their connections, weights, or bias) even if they employ the same module from the aforementioned models. However, the developed classifiers ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m typically handle the same type of data, for example, they all handle image recognition, all handle sound recognition, and so on.

    [0170] In particularly, the developed classifiers ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m are combined to form a vector function defined as


    V(x)=(ŷ.sup.β,1(x), . . . ,ŷ.sup.β,m(x),ŷ.sup.γ,1(x), . . . ,ŷ.sup.γ,m(x)))∈S.sup.2m, x∈Ω.

    [0171] Here, each V(x) is a preliminary result given by the developed classifiers ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m, and it is a 2m-dimensional real vector, and S.sup.2m collects the preliminary results for x∈Ω. The preliminary results will be further processed as follows.

    [0172] The bucket creation module 40 is connected to the classifier combination module 30 and configured to partition the training set Ω.sub.tr into buckets B(I) with identities I. That is,

    [00018] Ω tr = .Math. I S 2 m B ( I )

    where, for any identity I∈S.sup.2m,


    B(I)={x∈Ω.sub.tr:V(x)=I}.

    [0173] When an element x is said distributed to B(I), it means that V(x)=I. The identities I are associated with characteristics of the data.

    [0174] It can be understood that the buckets are also data sets created to realize the classification according to the present invention. To specifically explain the meaning of the bucket B(I) and its identity I, for example, in case of m=3 and k=4, a possible form of the identity may be I=(1,2,2,3), and a possible form of the bucket may be B(I)=B((1,2,2,3))={x∈Ω.sub.tr:ŷ.sup.β,1(x)=1,ŷ.sup.β,2(x)=2,ŷ.sup.β,3(x)=2,ŷ.sup.β,4(x)=3}.

    [0175] The bucket merger module 50 is connected to the bucket creation module 40 and configured to merge empty buckets and/or small buckets into large buckets, for example, according to their cardinalities, so as to reduce the bias caused by the rareness of data therein.

    [0176] In particular, it is possible to define n.sub.B(I)=|B(I)| and n.sub.B(I)(j)=|B(I)∩Ω.sub.tr(j)|, and obviously, n.sub.B(I)=Σ.sub.j=1.sup.mn.sub.B(I)(j). The bucket creation module 40 is then further configured to define (or denote) the cardinality n.sub.B(I)(j) of a bucket B(I) with a membership j and the cardinality n.sub.tr(j) of a subset of the training set Ω.sub.tr with the membership j, and to perform merger such that

    [00019] max j { n B ( j ) n tr ( j ) } δ

    holds for certain predetermined positive constant δ between 0 and 1. The choice of the constant δ may be problem dependent, so a specific value of δ will not be given in the present description.

    [0177] The membership assignment module 60 is indirectly connected to the bucket creation module 40 through the bucket merger module 50 and configured to assign respective memberships j's to the respective buckets B(I), for example, according to their cardinalities. The memberships j's refer to data categories of the training set Ω.sub.tr.

    [0178] One possible approach is that: let a bucket B(I) in the training set Ω.sub.tr with identity I be given. Assign the bucket B(I) a membership j if the ratio of the number of sample data (or elements) x's with membership j in B(I) to the cardinality |Ω.sub.tr(j)| of a subset Ω.sub.tr(j) of the training set Ω.sub.tr with membership j is maximal among ratios of all memberships. This defines a function Y on the collection of buckets B(I) to S that

    [00020] Y ( B ( I ) ) = j if max 1 k m { .Math. "\[LeftBracketingBar]" B ( I ) .Math. Ω tr ( k ) .Math. "\[RightBracketingBar]" .Math. "\[LeftBracketingBar]" Ω tr ( k ) .Math. "\[RightBracketingBar]" } = .Math. "\[LeftBracketingBar]" B ( I ) .Math. Ω tr ( j ) .Math. "\[RightBracketingBar]" .Math. "\[LeftBracketingBar]" Ω tr ( j ) .Math. "\[RightBracketingBar]" ,

    [0179] It should be emphasized that there are many ways to determine the membership of a bucket, and which then result in different functions Y.

    [0180] Referring back to FIG. 4, the output module 70 is indirectly connected to the classifier combination module 30 through the bucket creation module 40, the bucket merger module 50, and the membership assignment module 60, and configured to derive an output result after the sample data x is processed through the classifier combination module 30. It should be emphasized that the classifier combination module 30 has combined 2m classifiers ŷ.sup.β,1, . . . , ŷ.sup.β,m, ŷ.sup.γ,1, . . . , ŷ.sup.γ,m developed by the multiform separation engine 40.

    [0181] The output result may be directly the membership j, or converted to the data category, such as “dog” “cat”, or “rabbit” indicated by the membership. The output module 70 may include a display device to show the output result.

    [0182] The eclectic classifier in combination with the “QMS with weights” of the present invention can be expressed by the following formal definition:


    {tilde over (y)}(x)=Y(B(V(x))), x∈Ω.

    [0183] (Loyalty Type)

    [0184] With the aforementioned implementation, the loyalty extraction machine 1 of the present invention can further include a loyalty type indicator 90 connected between the membership assignment module 60 and the output module 70 and configured to produce a label of loyalty type of a sample data (x) by confirming a location of a lighter weight (β) and/or a heavier weight (γ) used to form the vector function V(x).

    [0185] Recalling that each identity I is a vector with 2m coordinates, for k=1, . . . , m, the identity with value k at the k-th coordinate is denoted by:

    [00021] I ( β , k ) = ( * , .Math. , * , k the k - th coordinate , * , .Math. , * , the first m coordinates * , .Math. , * the second m coordinates ) ,

    and the identity with value k at the (m+k)-th coordinate is denoted by

    [00022] I ( γ , k ) = ( * , .Math. , * , the first m coordinates * , .Math. , * , k the ( m + k ) - th coordinate , * , .Math. , * , the second m coordinates ) .

    [0186] Let x∈Ω. Strong loyalty, weak loyalty, and normal loyalty are defined below, respectively.

    [0187] An element x has a membership k with strong loyalty if V(x)=I(β,k) and {tilde over (y)}(x)=k=Y(B(I(β,k))).

    [0188] An element x has a membership k with weak loyalty if x is not strongly loyal to any membership, and V(x)=I(γ,k), and {tilde over (y)}(x)=k=Y(B(I(γ,k))).

    [0189] An element is said to have normal loyalty if it have neither strong loyalty nor weak loyalty.

    [0190] It is noted that an element x has weak loyalty if it does not satisfy the condition of strong loyalty for any j=1, . . . , m, and the conditions for weak loyalty hold for x for some k∈S.

    [0191] For convenience, let T={Strong, Normal, Weak} be the set of possible loyalty types. Also let τ:Ω.fwdarw.T be the “loyalty type function” such that τ(x) gives the loyalty type of x. Combining the “QMS with weights” and the “eclectic classifier” to one single mapping, define ζ:Ω.fwdarw.S×T by


    ζ(x)=({tilde over (y)}(x),τ(x))∈S×T, x∈Ω.

    [0192] In this way, the first component {tilde over (y)}(x) of ζ(x) indicates the membership of x, and the second component τ(x) of ζ(x) indicates the crucial information expressed as the loyalty type of x.

    [0193] Accordingly, the aforementioned formal definition can be regarded as the complete loyalty extraction machine 1 of the present invention.

    [0194] In this embodiment, there are three loyalty types: strong loyalty, weak loyalty, and normal loyalty. It can be understood that, more loyalty types can be established. Supposedly, by varying the weights continuously, a continuum loyalty type can be realized which corresponds to the notion of “level of confidence” proposed by the same applicant. On the other hand, it is also possible to employ only two loyalty types: strong loyalty and normal loyalty, when only the training process with lighter weight β is performed.

    [0195] (Test Results of the Loyalty Extraction Machine on Fashion-MNIST Samples)

    [0196] In order to explain the importance and the application of the loyalty type derived according to the present invention, the test results of the loyalty extraction machine of the present invention is provided on Fashion-MNIST samples.

    [0197] Fashion-MNIST is a dataset of Zalando's article images, consisting of a training set of 60,000 samples and a test set of 10,000 samples. Each sample is a 28×28 grayscale image, associated with a label from 10 categories (which are represented by “memberships” in the present invention). Fashion-MNIST is intended to serve as a direct drop-in replacement for the original MNIST dataset for benchmarking machine learning algorithms. It shares the same image size and structure of training and testing splits.

    [0198] The loyalty extraction machine 1 of the present invention is implemented by machine learning algorithm that produces a classifier providing the following information (referring to FIG. 4) for each image in Fashion-MNIST:

    [0199] (i) one loyalty type, such as strong, normal, or weak; and

    [0200] (ii) one predicted category label from 10 categories.

    [0201] It is noted that more loyalty types and more categories may be defined according to other databases in other applications.

    [0202] The following Table 1 summarizes the results for the training set. Let n be the total number of images in the data set. For each loyalty type, n.sub.1 is the number of images assigned to this loyalty type, n2 is the number of images with correct category label prediction, and LPA is the so-called local prediction accuracy, which is the prediction accuracy for this loyalty type. The overall prediction accuracy, i.e., sum of all numbers in Column n.sub.2 divided by n, is 95.09%. Equivalently, the overall prediction accuracy is equal to the inner product of two vectors, see column n.sub.1/n and column LPA.

    TABLE-US-00001 TABLE 1 Loyalty type n.sub.1 n.sub.1/n n.sub.2 LPA Strong 47,052 78.42% 46,859 99.59% Normal 10,653 17.76% 8,757 82.20% Weak 2.295  3.83% 1,436 62.57%

    [0203] The following Table 2 summarizes the results for the test set. The overall prediction accuracy is 89.44%.

    TABLE-US-00002 TABLE 2 Loyalty type n.sub.1 n.sub.1/n n.sub.2 LPA Strong 7,683 76.83% 7,419 96.56% Normal 1,884 18.84% 1,329 70.54% Weak 433  4.33% 196 45.27%

    [0204] While the loyalty extraction machine of the present invention produces an overall prediction accuracy comparable to that of several leading methods, it produces much higher prediction accuracy for images with strong loyalty type. It is important to notice that the loyalty type of an image can be predetermined. Also, for the Fashion-MNIST samples, images with strong loyalty type exceeds 76% of the total images, both in training test and in test set

    [0205] In conclusion, the present invention provides multiple multiform separation classifiers, each of them appropriately utilizes multiple functions, so as to produce better solutions, in terms of accuracy, robustness, complexity, speed, dependency, cost, and so on. The multiple multiform separation classifiers are used as developed classifiers in an eclectic classifier. The eclectic classifier combines the results from the developed classifiers that can give a maximal ratio or a majority of predictions or decisions regarded as an optimal answer. In this way, the extreme influences of the disadvantages of the developed classifiers can be avoided, and the advantages of the developed classifiers can be jointly taken into consideration.

    [0206] Although the present invention has been explained in relation to its preferred embodiment, it is to be understood that many other possible modifications and variations can be made without departing from the spirit and scope of the invention as hereinafter claimed.