Systems and Methods for Privacy Preserving Accurate Analysis of Network Paths

20230254144 · 2023-08-10

    Inventors

    Cpc classification

    International classification

    Abstract

    Anonymizing systems and methods comprising a native configurations database including a set of configurations, a key management database including a plurality of private keys, a processor in communication with the native configurations database and the key management database, and a memory coupled to the processor. The set of configurations includes one or more textual descriptions and one or more ranges, wherein each range includes a contiguous sequence comprised of IP addresses, port numbers, or IP addresses and port numbers. The processor is configured to retrieve the set of configurations from the native configurations database, wherein the set of configurations includes a plurality of objects; retrieve a private key from the key management database; assign a unique cryptographically secure identity to each object; and anonymize the plurality of objects based on the cryptographically secure identities and the private key. The present system prevents retrieving the textual descriptions and the ranges of the configuration files of the native configuration database from the anonymized configuration database

    Claims

    1. A system for anonymizing a group of configuration files to provide an anonymized configuration database, the system comprising: a native configuration database including sets of configuration files, each configuration file including one or more textual descriptions and one or more ranges, wherein each range includes a consecutive sequence of IP addresses, port numbers, or IP addresses and port numbers, wherein the ranges may be contiguous; and a key management database including a key assigned to each set of configuration files of the native configuration database; wherein the program instructions, when executed by the processor, permit the processor to execute steps for anonymizing the textual descriptions and the ranges of the native configuration database to provide an anonymized configuration database, and wherein retrieving the textual description and the ranges of the configuration files of the native configuration database from the anonymized configuration database is prevented.

    2. The system of claim 1, wherein each range includes a consecutive sequence of IP addresses and port numbers.

    3. The system of claim 1, wherein a rule-table is assigned to the set of configuration files, wherein the rule-table includes a plurality of rules that are position-insensitive.

    4. The system of claim 3, wherein each range of the anonymized configuration database representation includes the same number of members as the corresponding original range.

    5. The system of claim 1, wherein the ranges may be fragmented.

    6. The system of claim 1, wherein the step of anonymizing the ranges comprises creating a set of ranges that contain an endpoint but do not begin or end at the endpoint, wherein this set of ranges is created by the steps of: creating a list of range endpoints; sorting the list of range endpoints; and computing, for each endpoint, a set of ranges that contain the endpoint but do not begin or end at the endpoint.

    7. The system of claim 6, wherein the memory includes further program instructions executable by the processor that, when executed, cause the processor to offset the set of ranges.

    8. The system of claim 1, further comprising a remote anonymized systems database, and wherein the memory includes further program instructions executable by the processor that, when executed, cause the processor to: build a representation including a plurality of objects; assign a unique cryptographically secure identity to each object in the representation, each identity derived from the private key; and transmit the anonymized representation to the remote anonymized systems database.

    9. The system of claim 8, further comprising: a remote anonymized analysis results database; and a remote oblivious analysis engine in communication with remote anonymized systems database and the remote database of anonymized analysis results; and a further processor and a further memory coupled to the further processor configured to store further program instructions executable by the further processor, wherein the further program instructions, when executed by the further processor, cause the further processor to: receive user input to generate an anonymized analysis of the anonymized representation; retrieve, by the remote oblivious analysis engine, the anonymized representation from the remote anonymized systems database; analyze the anonymized representation to develop anonymized analysis results; and store the anonymized analysis results in the remote anonymized analysis results database.

    10. The system of claim 1, further comprising a remote anonymized configurations database, and wherein the memory includes further program instructions executable by the processor that, when executed, cause the processor to transmit the anonymized plurality of objects to the remote anonymized configurations database.

    11. The system of claim 10, further comprising: a remote anonymized analysis results database; a remote oblivious analysis engine in communication with remote anonymized systems database and the remote database of anonymized analysis results; and a further processor and a further memory coupled to the further processor configured to store further program instructions executable by the further processor, wherein the further program instructions, when executed by the further processor, cause the further processor to: receive user input to generate an anonymized analysis of the anonymized representation; retrieve, by the remote oblivious analysis engine, the anonymized plurality of objects from the remote anonymized configurations database; build a representation based on the anonymized plurality of objects; assign a unique cryptographically secure identity to each object in the representation, each identity derived from the private key; and transmit the anonymized representation to the remote anonymized systems database.

    12. The system of claim 1, wherein the memory includes further program instructions executable by the processor that, when executed, cause the processor to: generate a de-anonymization key; and store the de-anonymization key in the key management database.

    13. The system of claim 14, wherein the memory includes further program instructions executable by the processor that, when executed, cause the processor to: receive user input to de-anonymize the anonymized analysis results; receive, from the remote anonymized analysis results database, the anonymized analysis results and the de-anonymization key from the key management database; invert the anonymization performed on the representation; and present the de-anonymized results to the user.

    14. A method of anonymizing a group of configuration files for secure transmission from a remote-based analysis system to a user, each configuration file comprising an object, each object containing one or more textual descriptions and one or more ranges of consecutive IP addresses, port numbers, or IP addresses and port numbers, said method comprising the steps of: creating a native configuration database comprising sets of configuration files, each configuration file including one or more textual descriptions and one or more ranges of consecutive IP addresses, port numbers, or IP addresses and port numbers; assigning a key to the set of configurations, and creating a key management database comprising the keys; assigning a unique cryptographically secure identity to each object, each cryptographically secure identity being derived from the key; and anonymizing objects with a cryptographically secure identity derived from the key to provide an anonymized group of configuration files, wherein retrieving the textual descriptions and the ranges of the native configuration database from the anonymized configuration database is prevented.

    15. The method of claim 14, wherein access to decrypted groups of the anonymized groups of configuration files by a third party is prevented during transmission to and processing by a remote-based analysis system.

    16. The method of claim 15 wherein the remote-based analysis system is a cloud based remote analysis system.

    17. The method of claim 16 wherein a range-intersection within an anonymized collection of ranges of the IP address, port numbers or IP addresses and port numbers, is identified, and wherein the range-intersection is securely preserved in the anonymized groups of configurations processed by the remote-based analysis system, and the anonymized groups of configurations are processed by the remote-based analysis system to provide a processed anonymized group of configurations to be securely transmitted to the user.

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    [0028] The drawing figures depict one or more implementations in accord with the present concepts, by way of example only, not by way of limitations. In the figures, like reference numerals refer to the same or similar elements.

    [0029] FIG. 1 is a schematic representation of an example of components of a first anonymizing system.

    [0030] FIG. 2 is a schematic representation of the anonymizing system of FIG. 1 anonymizing a system representation.

    [0031] FIG. 3 is a schematic representation of the anonymizing system of FIG. 1 performing an oblivious analysis of the anonymized system representation.

    [0032] FIG. 4 is a schematic representation of the anonymizing system of FIG. 1 de-anonymizing the results of the oblivious analysis of the anonymized system representation.

    [0033] FIG. 5 is a schematic representation of the anonymizing system of FIG. 1 performing an oblivious merge of two anonymized systems into a larger aggregate system.

    [0034] FIG. 6 is a schematic representation of an example of components of a second anonymizing system.

    [0035] FIG. 7 is a schematic representation of the anonymizing system of FIG. 6 anonymizing a configuration.

    [0036] FIG. 8 is a schematic representation of the anonymizing system of FIG. 6 creating an anonymized system representation.

    [0037] FIG. 9 is a schematic representation of the anonymizing system of FIG. 6 performing an oblivious analysis of the anonymized system representation.

    [0038] FIG. 10 is a is a schematic representation of the anonymizing system of FIG. 6 de-anonymizing the results of the oblivious analysis of the anonymized system representation.

    [0039] FIG. 11 is a schematic representation of the anonymizing system of FIG. 6 performing an oblivious merge of two anonymized systems into a larger aggregate system.

    DETAILED DESCRIPTION OF THE INVENTION

    [0040] FIGS. 1-5 illustrate components of an anonymization system 100. In the schematic shown in FIG. 1, the anonymization system 100 includes a client-side system 102 and a remote server system 104 that communicate via the internet 106, although any suitable communication means may be used.

    [0041] Generally, all contiguous ranges of IP addresses and contiguous ranges of ports referenced in the file can be identified by analyzing a configuration or system representation. The anonymization technique of the anonymization system (described in greater detail below) identifies all connected ranges, and then constructs a translation of each connected range by finding an integer value used to shift the range, which can be easily inverted. By describing the matching of a rule-table to an IP packet header in terms of intersections, it is easily shown that a given IP packet header matches the rangers described in a rule if and only if the values of the packet header, when anonymized, match the ranges described in the rule under the same anonymization.

    [0042] In the illustrated embodiment, the anonymization system 100 includes a native configurations database 108 including a set of configurations, a key management database 110 including a plurality of private keys, and a remote anonymized systems database 112. A processor 114 is in communication with the native configurations database 108 and the key management database 110, and includes a memory 116 configured to store program instructions executable by the processor 114 that, when executed cause the processor 114 to undertake certain steps to anonymize a set of data.

    [0043] More specifically, the anonymizing system 100 performs a method of anonymizing data. Referring to FIG. 2, the system 100 first retrieves a set of configurations 118 from the native configurations database 108 and a private key 120 from the key management database 110. In the illustrated embodiment, a system representations module 122 retrieves the set of configurations 118 and the private key 120 in order then build a representation 124 comprised of a plurality of objects and assigns a unique cryptographically secure identity to each object in the representation 124, each identity being derived from the private key. In the illustrated embodiment, the representation 124 is transmitted to a system anonymizer module 126, which then anonymizes the representation 124 based on the cryptographically secure identities and the private key. The system 100 finally transmits the anonymized representation 128 to the remote anonymized systems database 112. The system anonymizer also generates a de-anonymization key 130 that is stored in the key management database 110 and used by a de-anonymization module 131 to de-anonymize the final results.

    [0044] The remote server system 104 includes the anonymized systems database 112, an oblivious analysis engine 132, an anonymized analysis results database 134, and an oblivious systems merger 136. The remote server system 104 may also include a second processor 138 and a second memory 140 including program instructions executable by the second processor 138.

    [0045] Referring to FIG. 2, the system representations module 122 retrieves a set of set of configurations 118 from the native configurations database 108 and a private key 120 from the key management database 110 upon receipt of instructions 142 from a user via a graphical user interface 144 within the client-side system 102. The set of configurations 118 may include a plurality of ranges including one or more of a set of potential hosts, a destination IP address, a source port, and a destination port. In some embodiments, the plurality of ranges may be fragmented. The systems representation module 122 builds a representation 124 of the set of configurations 118, the representation 124 including a plurality of objects. Each object in the representation 124 is provided a unique cryptographically secure identity derived from the object specification and private key 120. The representation 124 is transmitted to the system anonymizer module 126, which anonymizes the representation 124 based on the cryptographically object identities and the relevant private key 120.

    [0046] By analyzing a configuration 118 or system representation 124, all contiguous ranges of IP addresses and contiguous ranges of ports referenced in the file can be identified. A range includes contiguous sequences of IP address and contiguous sequences of port numbers. The system 100 identifies a selected source subnetwork or host and finds all flows the configuration allows whose origin is in the source subnetwork. Therefore, a source range includes a set of potential hosts, typically Classless Inter-Domain Routing blocks, which can be viewed as a range. The system generates a meta-packet based on a source subnetwork description, where the source IP address is the range of IP addresses in the originating subnetwork, and wild-card integer ranges for the destination IP address, the source port, and the destination port. In some embodiments, special cases of IP ranges that have special meaning and should not be anonymized, such as well-known private address spaces and multicast address spaces.

    [0047] As the meta-packets are pushed through the network description, the logic applied is similar to logic applied to individual packets. However, each step may fragment a meta-packet into multiple meta-packets that include address and port ranges that are subsets of the original ranges. Each meta-packet that reaches the destination is referred to as a “flow.”

    [0048] For an anonymization of a meta-packet that produces a set of flows, the de-anonymization of the ranges in those flows will produce exactly the same set of flows as are computed from the same non-anonymized meta-packet, which each flow having followed exactly the same sequence of devices, having been admitted by exactly the same set of access control rules, and having had exactly the same routing decisions made and have exactly the same NAT rules applied.

    [0049] The routing of packet information typically involves the application of a rule-table to an IP packet header. The IP packet header specifies integer values for the source and destination IP addresses, the source and destination port numbers, and the protocol. Each firewall applies a rule-table to the IP packet header in order to determine whether to admit the packet and transform part of the packet through Network or Port Address Translation. Similarly, each router applies a rule-table to determine whether to route the packet. For a five-dimensional IP packet header, a rule-table includes five-dimensional matching conditions. The matching conditions can be exact values or ranges in terms of specifying a union of intervals. Example applications of rule-tables to two IP packet headers are provided below to illustrate matching and non-matching conditions.

    TABLE-US-00001 packet header condition matching? (130.126.151.36, 21790, (*, *, 192.17.90.0/ match 192.17.90.136, 443, 24 443, 6) (130.126.151.36, 21790, (*, *, 192.17.90.0/ no match 192.17.90.136, 443, 24, 110, 6)

    [0050] The range field of a meta-packet resulting from a match is marked as “shaped” if the range in that field of the rule is from a known space is and is not a wildcard. The distinction is important in the logic for computing precisely which flows are presented to a rule. For a list of rules R1, R2, . . . Rn, and a meta-packet P, the logic is to take the intersection of P's fields with R1's fields, and siphon off the meta-packet which intersects R1 in all dimensions. The residual flow P/R1 (i.e., P excluding R1) is logically the set of meta-packets which are intersected with R2. These reflect the exclusion of R1, and the set of meta-packets presented to R3 will represent the exclusion of both R1 and R2. To support anonymization, the exclusions applied to unshaped ranges are not computed, but instead are carried along symbolically. When a meta-packet is discovered to be a path, one or more of its dimensions may be unshaped. If there are exclusions, these are necessarily from known (anonymized) space and are represented symbolically, which means that to correctly identify the ranges of unknown space represented as a course or destination, de-anonymization needs to be applied only to the exclusions.

    [0051] The action applied to a packet depends on which rules match, their position in the table, and sometimes the size of the ranges encoded in the rule's matching conditions. For example, in many firewalls the action applied to a packet is that action associated with the rule among all possible matches which appears earliest in a sequential representation of the rule-table. In some other cases such as a BSD firewall, it is the last rule among all possible matches in the table that determines the action, unless there is a match with an earlier rule designated as being “quick.” In the case of a router's forwarding table, the matching rule which defines the routing applied may be the rule whose matching condition is smallest.

    [0052] In the anonymization system 100, the rule which defines the action applied to the packet is (1) among a set of matching rules whose matching conditions have non-empty intersection with the packet in the corresponding dimensions, and (2) the algorithm for selecting among all matching rules depends in no way on the specific values of the ranges in those rule's matching conditions. Therefore, each rule-table and the algorithm for selecting among matching rules is position-insensitive. For example, if P is a computer program whose inputs include ranges of integers and I is a subset of those ranges, P is “position-insensitive” on I if any branching decision that depends on I is the same for multiple positionings of the range in the domain.

    [0053] Within the IP space, a range [a, b] is related to range [c, d] if those ranges have a non-empty intersection. Ranges may be aggregated into disjoint “connected” sets where all ranges in a connected set are related by the transitive closure of the “intersects” relations. A connected set can be described by a “super-range’ [A, B] where any member of the connected set is a subset of [A, B]. Obviously, if each of [a, b] and [c, d] is contained in [A, B], then the intersection of [a, b] and [c, d] will be either empty or be in [A, B]. By definition, the connected sets do not intersect. Therefore, if [a, b] is in one connected set and [c, d] is in a different connected set, their intersection is empty.

    [0054] In some cases, the native configurations contain object group names that are suggestive of geographic region or role of the IP addresses in the group. These group names can be anonymized by using a key to initialize a cryptographically secure random number generator, and either to create completely randomized strings as anonymizations, or to use the random number stream to draw without replacement from a very large dictionary of words (initially randomized itself). The private de-anonymization key provides the inverse mapping.

    [0055] The anonymizing steps of the present anonymizing system 100 are all position-insensitive, at least with respect to IP ranges being consistent with CIDR alignment. The decisions involved in matching access control rules, routing decisions, and application of NAT transformations are determined by taking intersections of IP and/or port ranges in a position-insensitive fashion. In this way, the anonymization preserves the structure of intersections.

    [0056] In one embodiment, the anonymization system 100 draws elements from a domain D to anonymize. While a conventional domain D may include non-negative integer representation of IP addresses, it is presumed that all elements of D are ordered with respect to the numerical values of the IP addresses they represent. To define a range on D, Let a, b∈D. Define [a, b] as a range on D:


    [a,b]={x∈D|a≤x≤b}

    [0057] Not every element in D needs to be anonymized. A subset of ranges I={[s.sub.1, t.sub.1], [s.sub.2, t.sub.2], . . . , [s.sub.n, t.sub.n]} have been extracted or inferred from the configuration to be anonymized, and their union is D.sub.I. The anonymization system will anonymize only those x∈D.sub.I.

    [0058] Presuming an anonymization function φ: D.sub.I.fwdarw.D on individual elements, ranges on D, can be mapped into subsets of D, by defining


    A.sub.φ([u,v])={φ(x)/for all x∈[u,v]}

    [0059] In order to anonymize ranges, analyze the anonymized ranges, and de-anonymize the results, the anonymization range must have the same structure as the original range. More specifically, the image under anonymization is itself a range and has exactly the same number of members as the original. These conditions are necessary for the anonymization steps to be position-insensitive, as the anonymization system and modules within are expecting the structure of ranges and may make control decisions based on the number of elements in that range (even if decisions are not based on the specific values in that range). In this way, the anonymization function φ preserves structure.

    [0060] More specifically, the anonymization function φ preserves structure if it is one-to-one, and for every [u, v] on D.sub.I,


    A.sub.φ([u,v])=[φ(u),p(v)] and card([u,v])=card([φ(u),φ(v)])

    where card([s, t])=v(t)−v(s)+1 is the cardinality function.

    [0061] A consequence of an anonymization function φ preserving structure is that ranges which do not overlap in D.sub.I have images under φ which do not overlap. Otherwise φ would not be one-to-one. Furthermore, preservation of structure implies that ranges which are nested in D.sub.I have images that are nested. That is, if [s, t]⊂[u, v], then


    A.sub.φ([s,t])=[φ(s),φ(t)]⊂[φ(u),φ(v)]=A.sub.φ([u,v])

    [0062] This means the effect of φ can be viewed globally, in terms of its effect on canonical ranges. In one example, [s, t] on D.sub.I is said to be canonical if there exists no s.sup.l∈D.sub.I with s.sup.l<s such that [s.sup.l, t] is a range on D.sub.I, nor is there any t.sup.l∈D.sub.I with t<t.sup.l such that [s, t.sup.l] is a range on D.sub.I. A canonical range is “as big as possible” in D.sub.I. Where φ changes a canonical range, it changes every one of its subranges. If φ preserves structure, its effect on D.sub.I has to be that of shifting the canonical ranges around in D so that they do not overlap.

    [0063] According to one theorem, I is a set of ranges on set D and D.sub.I is their union. R(D.sub.I) is the set of canonical ranges on D.sub.I, and the anonymization function φ: D.sub.I.fwdarw.D preserves structure. Then, for each [s.sub.i, t.sub.i]∈R(D.sub.I), there exists an integer β.sub.i such that:


    φ(x)=v−1(v(x)+βi) for all xcustom-character[si,ti]

    [0064] To prove the theorem, choose any [s.sub.i, t.sub.i]∈R(D.sub.I), and define β.sub.i=v (φ(s.sub.i))−v(s.sub.i). Now:

    [00001] v ( φ ( si ) ) = v ( si ) + ( v ( φ ( si ) ) - v ( si ) = v ( si ) + β i

    [0065] Since v is one-to-one, v.sup.−1 is applied to both sides of this equation to see that:


    v.sup.−1(v(φ(s.sub.i)))=v.sup.−1(v(s.sub.i)+β.sub.i),

    or equivalently φ(s.sub.i)=v.sup.−1(v (s.sub.i)+β.sub.i). Now choose any x∈[s.sub.i, t.sub.i] with s.sub.i<x. Because φ preserves structure, A.sub.100 ([s.sub.i,x])=[φ(s.sub.i), φ(x)]. Furthermore, v(φ(x))−v(φ(s.sub.i))=v(x)−v(s.sub.i) because the image of [s.sub.i, x] under φ is a range with the same number of elements as [s.sub.i, x]. It follows that:

    [00002] v ( φ ( si ) ) = v ( x ) + ( v ( φ ( si ) ) - v ( si ) = v ( x ) + β i

    Applying v.sup.−1 to both sides gives φ(x)=v.sup.−1(v (x)+β.sub.i) as required. □

    [0066] As shown above, any anonymization function that preserves structure necessarily has the form of “shifting” canonical ranges within D. If [s.sub.i, t.sub.i]∈R(D.sub.I) is shifted by an offset β.sub.i, the numerical range [v(s.sub.i), v(t.sub.i)] is shifted by |β.sub.i| values, meaning forward if β.sub.i is positive and backward if β.sub.i is negative.

    [0067] This proves that if one is going execute the same computer program on anonymized ranges as one would with original ranges, this form is a necessary condition for the anonymization. To be a sufficient condition for proper execution and ability to de-anonymize, the structure preserving anonymization must also preserve range intersections. This means that the anonymization of intersections is the same as the intersection of anonymizations.

    [0068] Given that φ:D.sub.I.fwdarw.D is an anonymization function, φ preserves range intersections if, for every pair of ranges [s, t], [u, v] on D.sub.I,


    A.sub.φ([s,t]∩[u,v])=A.sub.φ([s,t])∩A.sub.φ([u,v]),

    taking A.sub.φ(∅)=∅.

    [0069] The importance of the idea is clear that when A.sub.φ is one-to-one, the anonymization is invertible. Presume that ranges on D.sub.I are initially anonymized and, in the course of things, take the intersection B=A.sub.100 ([s, t])∩A.sub.100 ([u, v]). If φ is range-intersection preserving, then we can recover from the expression of B what the intersection of the pre-anonymized ranges would be, by A.sub.φ.sup.−1(B)=A.sub.φ.sup.−1(A.sub.φ([s, t]∩[u, v]))=[s, t]∩[u, v]. This property enables us to de-anonymize the results that the position-insensitive computation on the anonymized ranges produced. The next step is to show that if φ preserves structure, then it also preserves range intersection.

    [0070] According to a second theorem, if p: D.sub.I.fwdarw.D is an anonymization function which preserves structure, then φ preserves range intersection.

    [0071] To proof this theorem, let [s, t] and [u, v] be ranges on D.sub.I. If [s, t] and [u, v] have non-empty intersection values, they must be subranges of the same canonical range. Since φ preserves structure there must exist β such that φ shifts all members of the image of the canonical range under v by β. The elements of D are globally ordered, and the shifting of the numerical values corresponding to [s, t] and [u, v] can be envisioned as shifting elements of D along that global ordering, by |β| units, towards larger values if P is positive, towards smaller values if β is negative. With respect to that order, Aφ([s, t]) shifts [s, t] and Aφ([u, v]) shifts [u, v] exactly |β| positions on D. Likewise, if [x,y]=[s, t]∩[u, v], then A.sub.φ([x,y]) shifts [x,y] exactly |β| positions to align exactly with Aφ([s, t])∩Aφ([u, v]), as required. Now if [s, t] and [u, v] are in the same canonical range but do not intersect, by the same logic, shifting them both in D by the same shift offset β implies they will not intersect. Finally, if [s, t] and [u, v] come from different canonical ranges, then because φ necessarily keeps the images of canonical ranges from overlapping, the intersection of their respective anonymizations is also empty.

    [0072] If D.sub.I is anonymized in such a way that the computer program operating on the anonymized ranges can be insensitive to the anonymization, the anonymization must preserve structure, otherwise the anonymized ranges differ in attributes other than just value. Since the anonymization preserves structure, it has a particular form, of shifting canonical ranges about without overlap. It also preserves range intersections, meaning that the results produced by the position-insensitive computation can be de-anonymized. Any anonymization such that the results of position-insensitive computation can be precisely de-anonymized must have the shifting form.

    [0073] D may comprise an integer or a natural number representation of IP addresses.

    [0074] In the anonymization of range sets, anonymization consists of assigning offsets β.sub.i to the canonical ranges. The offsets are selected to ensure that the anonymized canonical ranges remain separated and leak no information about the values of the pre-anonymized canonical ranges.

    [0075] Given a set of ranges I, the canonical ranges can be identified in time proportional to n log n, where n is the number of ranges, as follows: [0076] 1. In linear time create a list L of range endpoints, r.sub.1, r.sub.2, . . . , r.sub.m, where a set B.sub.i of ranges which begin with r.sub.i and a set T.sub.i of ranges that terminate at r.sub.i are associated with each i. [0077] 2. Sort L. As there are no more than 2n endpoints, this is accomplished in O(n log n) time. [0078] 3. With a linear scan of sorted L that uses the B.sub.i and T.sub.i sets to track precisely which ranges encompass each endpoint r, compute the set of ranges C(r) that contain r but do not begin or end at r for each r∈L. Using a range tree data structure, the computation at each r is O(log n). [0079] 4. A canonical range is found as [s, t] where C (s) and C (t) are both empty, but C(r) is not empty for every r∈L with s<r<t.

    [0080] Certain details of the algorithm depend on whether the anonymization offsets β.sub.i are constrained (as they are with CIDR block addresses) or unconstrained.

    [0081] In an unconstrained anonymization, taking D to be the natural numbers, let N be the largest value in the IP space to be anonymized (e.g. N=2.sup.32−1 for IPv4 and N=2.sup.64−1 for a user domain space under IPv6), let I be a set of ranges on D to anonymize, and let k be the number of canonical ranges on D.sub.I.

    [0082] Without loss of generality, the least value to which an address can be mapped is 0. There may well be a minimum value below which assignment should not be made, but the descriptions below are easily amended to accommodate this constraint. The goal is to pack all of the canonical ranges starting at the low end of [0, N] in such a way that provides no indication to an observer of what the original location may have been. If canonical ranges were packed in order of their pre-anonymized appearance, the observer would have knowledge that the native position of a canonical range could not appear earlier than its anonymized position. To defeat that kind of information leakage, the canonical ranges may be randomly permutated, with each permutation having the same probability 1/k! of being chosen. Using the Knuth shuffle, this is accomplished in O(k) time. Re-indexing the range endpoints to reflect the permutation, the first canonical range [s.sub.1, t.sub.1] is mapped to [0, (t.sub.2−s.sub.1 s.sub.1 with anonymization offset β.sub.1=−s.sub.1. This defines φ(s.sub.2)=0 and φ(t.sub.2)=(t.sub.1−s.sub.1). The second t range [s.sub.2, t.sub.2] is mapped to [φ(t.sub.2)+1, φ(t.sub.1)+1+(t.sub.2−s.sub.2)] with offset β.sub.2=φ(t.sub.1)+1−s.sub.2, so that φ(s.sub.2)=φ(t.sub.1)+1 and φ(t.sub.2)=φ(s.sub.2)+(t.sub.2−s.sub.2), and so on.

    [0083] In an unconstrained anonymization, IPv4 address ranges usually derive from a CIDR block (e.g. 172.80.8.0/24), and are expressed in the device configurations as CIDRblocks, or as IP addresses with net-masks which are semantically equivalent to CIDR blocks. This constrains the possible anonymization offsets. To anonymize the configuration itself, then a “/24” network in the configuration needs to be transformed into a “/24” network under the anonymization. All CIDR blocks must be remapped to locations that accommodate them, because the starting address of a /d network must have zeros in its 32−d least significant bits. The larger the network, the more zeros the first address must have. With a canonical range, the system finds the CIDR blocks of largest size and ensures that the canonical range is placed to accommodate them. Therefore, if the largest CIDR block with a canonical range is a /d network, then the starting address for the anonymized placement of that CIDR block must have 0 in its 32−d least significant bits. Since the canonical range is a contiguous sequence of IP addresses, ensuring that one of the largest blocks is properly aligned ensures that all the blocks of largest size are aligned and that any smaller blocks are also aligned.

    [0084] To avoid placing smaller networks in ways that inhibit the placement of larger networks, networks are placed in decreasing severity of constraint. Within a set of canonical ranges with the same maximal CIDR block size, the order of their placement is uniformly at random. Once all constrained placements are made, any remaining unconstrained placements are then made. [0085] 1. Classify each canonical range as being constrained if any of its constituent ranges is a CIDR block (two or more addresses), or otherwise unconstrained. [0086] 2. Order the constrained canonical ranges in decreasing size of largest constituent CIDR block, and group together canonical ranges whose largest constituent CIDR blocks have the same size. Denote by G.sub.n a group whose defining CIDR block is a /n network. [0087] 3. In order of increasing i, randomly permute the members of G.sub.i, and pack them one-by-one onto [0, N], beginning at the bottom, in such a way that a network from G.sub.i is aligned so that its /i CIDR sub-ranges are mapped to addresses whose least significant 32−i bits are zero. [0088] 4. Left now with only the unconstrained canonical ranges, use the unconstrained algorithm described earlier, packing the unconstrained canonical ranges beginning with the end of the packed constrained canonical ranges.

    [0089] Anonymization and de-anonymization is very efficient, each operation taking only O(log k) time where k is the number of canonical ranges. Space requirements are also low, just O(k).

    [0090] In one embodiment, the anonymization key for the anonymization function φ includes a table with two columns, with as many rows as there are canonical ranges. The ranges may appear in order in the table. To anonymize a value or a subrange, one uses binary search on the rows to efficiently find the canonical range which contains it. The second column gives the shift offset for that canonical range, and the anonymization is simply the addition of the offset to the value or to the endpoints of the range.

    [0091] The de-anonymization is similar, although using a separate key. There number of rows in this key corresponds to the number of canonical ranges, and the anonymized positions of the ranges are given in the first column of each row. Those rows are sorted with respect to those anonymized values. For a given row, the second column is the negated value of the shift offset for the associated canonical range. To determine the de-anonymization of a value or a range, a binary search is performed on the first column to find the corresponding canonical range, then add the negated offset in the second column to the anonymized value, or endpoints of the anonymized range.

    [0092] A path is represented in terms of the source and destination IP ranges, the source and destination port ranges, set of protocols, and sequence of devices and rules which define the path. All of the ranges involved in the path are either from known space or are unshaped with exclusions from known space. To de-anonymize a range from known space, the anonymizing system retrieves the de-anonymization key for the anonymized range of the connected range that must contain it. The shift offset is associated with the connected range, which when negated and added to the endpoints of an anonymized range translates the anonymized range back to the original range.

    [0093] Protocols are identified by well-known identifications in an IP header. In some embodiments, a protocol may be substituted with another integer and de-anonymized by reversing the substitute.

    [0094] The possible state of an observer's knowledge about the non-anonymized ranges with an a priori joint probability distribution. Denoted by γ.sub.i, the original starting location for the i.sup.th range observed in the anonymized configurations. The joint distribution is the set of non-zero probabilities:


    Pr(a.sub.1, a.sub.2, . . . ,p.sub.n)=Pr{γ.sub.1=a.sub.1, γ.sub.2=a.sub.2, . . . ,γ.sub.n=a.sub.n}.

    [0095] Denote by δ.sub.i the starting position of the i.sup.th anonymized range, and define Δ=(δ.sub.1, δ.sub.2, . . . , δ.sub.n). The question of how knowledge of Δ impacts the posterior probability of where the canonical ranges are located in the original configurations can be answered by Bayes Theorem.

    [00003] Pr ( a 1 , a 2 , .Math. , p n | Δ ) = Pr ( Δ | γ 1 = a 1 , γ 2 = α 2 , .Math. , γ n = a n ) Pr ( a 1 , a 2 , .Math. , p n ) Pr ( Δ ) = Pr ( a 1 , a 2 , .Math. , p n )

    [0096] because Pr(Δ|γ.sub.1=a.sub.1, γ.sub.2=a.sub.2, . . . , γ.sub.n=a.sub.n)=Pr(Δ). This last equivalence follows because placement of the anonymized canonical ranges depend only on their size of their constituent ranges (and this only in the case of constrained ranges); the anonymization is completely independent of the original range's placement. This shows that the anonymization yields no more information about the pre-anonymized ranges beyond what would otherwise be know about their likely distributions.

    [0097] In the course of anonymizing the representation, a de-anonymization key is created and stored in the key management component. The de-anonymization key identifies for each range object in the original range of integers and the transformed range of integers. A public key that describes the range sizes of each object, indexed by its cryptographically secure identity, is also stored for later use.

    [0098] Referring back to FIG. 2, the anonymized representation 128 is transmitted to the remote anonymized systems database 112 on the remote server system 104 once the representation 124 is anonymized. Referring to FIG. 3, input data 146 received by the client-side GUI 144 triggers the specification of the system 100 and other parameters to be analyzed by the oblivious analysis engine 132, which obtains the specific anonymized representation 128 from the remote anonymized systems database 112. The oblivious analysis engine 132 computes the analysis and stores an anonymized analysis results 148 in the anonymized analysis results database 134.

    [0099] To de-anonymize the results 148 as shown in FIG. 4, a command 150 is received from the user GUI 144 to perform the de-anonymization. The results de-anonymizer module 131 on the client-side system 102 retrieves the anonymized system representation 128 from the remote anonymized systems database 112, the anonymized analysis results 148 from the anonymized analysis results database 134, and the private de-anonymization key 130 associated with the system representation 124 from the key management database 110. Using the de-anonymization key 130, the results de-anonymizer module 131 inverts the anonymization performed on the ranges of that system representation 124 and presents the de-anonymized results 151 to the user via the GUI 144.

    [0100] In a further embodiment, the anonymizing system 100 merges two anonymized systems into a larger aggregate system. Upon the receipt of instructions 152 to merge two systems from the GUI, the first and second private keys 154, 156 of first and second sets of configurations, respectively, are merged to produce a private anonymization key 158 for the merged system and a public key 160 that identifies the size of each object. The instructions emitted from the GUI also include a command 162 to an oblivious systems merger module 136 to merge the first and second anonymized representations of the first and second sets of configurations, respectively. The oblivious systems merger module 136 retrieves the first and second anonymized representations 163 and the public key 160 from the key management database 110 and produces a merged anonymized system 164 that is stored on the remote anonymized systems database.

    [0101] FIGS. 6-11 illustrate a further embodiment of the anonymized system 200 including a client-side system 202 and a remote server side system 204 communicating through the internet 206 or other suitable system. In contrast to the anonymized system 100, the anonymized system 200 directly anonymizes the set of configurations 218 on the client-side server 202, and the step of generating the system representation occurs on the remote server system 204 instead of the client-side system 202. For ease of reference, like reference numbers are used with respect to the embodiment of FIGS. 1-5 and the embodiment of FIGS. 6-11.

    [0102] Similar to the client-side system 102 within the anonymized system 100, the client-side system 202 includes a native configurations database 208, a key management database 210, a results de-anonymizer module 231, and a first processor 214 as shown in FIG. 6. The client-side server 202 also includes a configuration anonymizer module 226, which retrieves a set of configurations 218 from the native configurations database 208 and a private key 220 from the key management database 210, as shown in FIG. 7. The configuration anonymizer module 226 creates cryptographically secure range object identities (similar to the system anonymizer module 126 of the anonymizing system 100) and applies the transformation directly to the object identities to generate an anonymized configuration 229. The anonymized configuration 229 is transmitted to an anonymized configuration database 227 on the remote server system 104. The configuration anonymizer module 226 also generates a de-anonymization key 230 for storage in the key management database 210.

    [0103] The remote server system 104 includes a system representation module 222, an anonymized configurations database 227, an anonymized systems database 212, an oblivious analysis engine 232, an anonymized analysis results database 234, and an oblivious systems merger module 236. The remote server system 204 may also include a second processor 238 and a second memory 240 including program instructions executable by the second processor 238.

    [0104] In FIG. 8, the anonymized system 200 generates an anonymized system representation 228 for a set of anonymized configurations 229. Upon receipt of instructions 242 from the graphical user interface 244 on the client-side server 202, the system representation module 222 on the remote server system 204 retrieves the anonymized configuration 229 from the anonymized configurations database 227 and builds an anonymized representation 228 based on the anonymized configuration 229 using object identities that are obtained from the meta-data associated with the anonymized configurations 229. The anonymized system representation 228 is then stored in the anonymized systems database 212 on the remote server system 204.

    [0105] FIG. 9 illustrates the steps undertaken to anonymize the analysis of the anonymizing system 200. Input data 246 received by the client-side GUI 244 triggers the specification of the system 200 and other parameters to be analyzed by the oblivious analysis engine 232, which obtains the specific anonymized representation 228 from the remote anonymized systems database 212. The oblivious analysis engine 232 computes the analysis and stores an anonymized analysis results 248 in the anonymized analysis results database 234.

    [0106] To de-anonymize the results 248 as shown in FIG. 10, a command 250 is received from the user GUI 244 to perform the de-anonymization. The results de-anonymizer module 231 on the client-side system 202 retrieves the anonymized system representation 228 from the remote anonymized systems database 212, the anonymized analysis results 248 from the anonymized analysis results database 234, and the private de-anonymization key 230 associated with the anonymized configuration 229 from the key management database 210. Using the de-anonymization key 230, the results de-anonymizer module 231 inverts the anonymization performed on the ranges of the anonymized configuration 229 and presents the de-anonymized results 251 to the user via the GUI 244.

    [0107] In a further embodiment, the anonymizing system 200 merges two anonymized systems into a larger aggregate system. Upon the receipt of instructions 252 to merge two systems from the GUI, the first and second private keys 154, 156 of first and second sets of configurations, respectively, are merged to produce a private anonymization key 258 for the merged system and a public key 260 that identifies the size of each object. The instructions emitted from the GUI also include a command 262 to an oblivious systems merger module 236 to merge the first and second anonymized representations of the first and second sets of configurations, respectively. The oblivious systems merger module 236 retrieves the first and second anonymized representations 263 and the public key 260 from the key management database 110 and produces a merged anonymized system 264 that is stored on the remote anonymized systems database 212.

    [0108] It is the nature of the anonymized configuration or system representation that the sizes of ranges are publicly known. A netmask for an IP address identifies a CIDR block of a given size containing the IP address. The anonymization system includes an offset that shifts native ranges without creating overlaps. Information about the original offsets is contained if, for example, it is known that if [a,b] is strictly less than [c,d] and the anonymized [a,b] is strictly less (or strictly greater) than the anonymized [c,d]. Information leakage about the original set of range values will be minimized if knowledge of relative positions of the anonymized ranges gives no better information than random chance about the position of the original range. The cryptographically secured identification of an object range gives no better information than random about the original location of the range, which implies that a placement of connected range that is deterministic function of the identifications of the objects it contains will not leak original position information. Cryptographically secure random permutations may also be used.

    [0109] Aspects of the systems and methods provided herein encompass hardware and software for controlling the relevant functions. Software may take the form of code or executable instructions for causing a processor or other programmable equipment to perform the relevant steps, where the code or instructions are carried by or otherwise embodied in a medium readable by the processor or other machine. Instructions or code for implementing such operations may be in the form of computer instruction in any form (e.g., source code, object code, interpreted code, etc.) stored in or carried by any tangible readable medium.

    [0110] It should be noted that various changes and modifications to the presently preferred embodiments described herein will be apparent to those skilled in the art. Such changes and modifications may be made without departing from the spirit and scope of the present invention and without diminishing its attendant advantages.