METHODS FOR CORRELATED HISTOGRAM CLUSTERING FOR MACHINE LEARNING
20230119704 · 2023-04-20
Assignee
Inventors
Cpc classification
G06N7/01
PHYSICS
International classification
Abstract
A methodology for correlated histogram clustering for machine learning which does not require a priori knowledge of cluster numbers, which extends beyond bimodal scenarios to multimodal scenarios, and does not need iterative optimization methods nor require powerful data processing.
Claims
1. A method in a machine learning system for generating correlated histogram clusters, comprising the steps of: 1) generating n histograms for n-dimensional data set, D; 2) selecting a subset (D′) of the histogram data based on a frequency greater than a threshold; 3) generating n histograms for n-dimensional data set D′ with optimal bin size; 4) identifying m histogram peaks (modes) for each dimension; 5) for the i.sup.th peak of the j.sup.th dimension, m.sub.ij, identify an index, p, in the data by finding the value in dimension j of D′ closest to m.sub.ij, and set value C.sub.i of centroid C equal to m.sub.ij; 6) identify the associated data value D′.sub.pk for another one of the dimensions, k, and identify the nearest peak from the histogram of k.sup.th dimension to D′.sub.pk, and assign value C.sub.k of centroid C to that peak; 7) repeat step 6 for every dimension of data k through n, k≠I; 8) save centroid C and repeat steps 5-7 for all histogram peaks of the j.sup.th dimension; and, 9) repeat steps 5-8 for all dimensions j through n.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0035] For a more complete understanding of the present disclosure, reference is now made to the following detailed description taken in conjunction with the accompanying drawings, in which:
[0036]
[0037]
[0038]
[0039]
[0040]
[0041]
[0042]
[0043]
[0044]
[0045]
[0046]
[0047]
[0048]
[0049] Corresponding numerals and symbols in the different figures generally refer to corresponding parts unless otherwise indicated and, in the interest of brevity, may not be described after the first instance.
DETAILED DESCRIPTION
[0050] The following detailed description discloses a methodology, for use in or training of a machine learning system, for generating correlated histogram clusters. The methodology does not require a priori knowledge of cluster numbers, which extends beyond bimodal scenarios to multimodal scenarios, and does not need iterative optimization methods nor require powerful data processing. With so much effort spent on the various machine learning techniques of unsupervised learning, a relatively simple yet unobvious approach is to leverage statistics and correlate histogram data.
[0051]
Centroids from Coarse Histograms
[0052] Selecting a threshold frequency (number of counts) results in a coarse histogram. From coarse histograms, the optimal number of bins can be determined for both data sets. Selecting the larger value results in both histograms having equal resolution. For each data set, multiple extrema may be identified from the counts; and for each extrema, the midpoint of the edge width determines the centroid.
[0053] For the histograms illustrated in
Centroids from Density Estimates
[0054] In cases where a histogram is not expressive of the underlying modality, a density estimate that is sensitive to modality may be employed. The Harrell-Davis Density Estimator (Harrell, 1982) is one such density estimate, though there are many, that can aid in the identification of peaks. To interpret modes in the density estimates, another method is required—the Lowland Modality Method using Quantile Respective Density Estimates (QRDE) can be used to find modes from a density estimate (Akinshin, 2020). Using this method, modes are defined as the highest peak, M, between two other peaks, P1 and P2, such that the proportion of the bin area between M and P.sub.i and the total rectangular area between the M and P.sub.i is greater than some threshold value, called the sensitivity.
[0055]
[0056] The embodiment described hereinafter will reference centroids found in
Optimal Bin Counts
[0057] Any method for constructing a histogram will require the choice of a bin count. There are simple rules-of-thumb for obtaining a bin count like taking the square root of N, taking 1+log(N) (e.g., using the Sturges Method as known to those skilled in the art), or taking 2+cube root of the N where N for all of these is the number data points. These methods rely on the number of data points rather than the underlying statistics of the data. One such approach is to minimize the following function of the mean and variance of the frequencies (Shimazaki and Shinomoto, 2007):
As is typically done, histograms, as well as density estimates, are sorted as shown in
Indexing
[0058]
[0059] If the data is indexed, then one simply looks for an index corresponding to a particular centroid (from A's data set) and then uses that same index to locate the other centroid (from B's data set). For example, data set A has one of many indexes that match the value 4.73 (within a few values of the second decimal place), one of which happens to be the index 77. Looking at data set B, index 77 leads one to find a corresponding value of 0.43. Recognizing 0.43 is near 0.48 (within a few values of the second decimal place), one concludes that one of the cluster centroids (A, B) is the pair (4.73, 0.48). It turns out that this just happens to be the same as the second elements in the histogram order for A and B.
[0060] Repeating the methodology, data set A has one of many indexes that match the value 8.43 (within a few values of the second decimal place), one of which happens to be the index 79. Looking at data set B, index 79 leads one to find a corresponding value of 0.26, which just happens to coincide with the centroid. Thus, one concludes that another of the cluster centroids (A, B) is the pair (8.43, 0.26). This is not in the order of the histogram data. The third centroid value of A corresponds to the first centroid values of B.
[0061] The methodology can be repeated for the last pair or deduced by elimination that is it must be (1.77, 1.25). Of course, this is not in the order of the histogram data. The first mode value of A corresponds to the third mode value of B.
[0062] The final set of three correlated centroids are (1.77, 1.25), (4.73, 0.48), and (8.43, 0.26). A visual embodiment of the final result (a two-dimensional histogram) is shown in
[0063] The foregoing methodology can be extended beyond this tri-modal example embodiment of two data sets to an embodiment of a multi-modal, n-dimensional data set without the need for knowing the cluster number a priori and performed rapidly without having to apply advance algorithmic techniques. The correlated histogram clustering (“CHC”) methodology is illustrated by the flowchart 900 in
The CHC methodology results in a novel approach to clustering that does not require a priori knowledge of cluster number, extends to multimodal scenarios, and does not need iterative optimization methods nor require powerful data processing.
Comparing Correlated Histogram Methodology to Existing Approaches
[0081] As discussed previously, there exist other approaches to finding clusters in a dataset.
[0082]
[0083]
[0084] Finally,
[0085] The foregoing has disclosed a novel methodology for generating correlated histogram clusters which can be used to advantage in machine learning systems and the training thereof. Although the embodiments and the advantages have been described in detail, it should be understood that various changes, substitutions, and alterations can be made herein without departing from the spirit and scope thereof as defined by the claims. For example, many of the features and functions discussed above can be implemented in software, hardware, firmware, or a combination thereof. Also, many of the features, functions, and steps of operating the same may be reordered, omitted, added, etc., and still fall within the scope of the claims and equivalents of the elements thereof.