SYSTEM AND METHOD FOR SUPERVISED NETWORK CLUSTERING
A method (and system) for supervised network clustering includes receiving and reading node labels from a plurality of nodes on a network, as executed by a processor on a computer having access to the network, the network defined as a group of entities interconnected by links. The node labels are used to define densities associated with the nodes. Node components are extracted from the network, based on using thresholds on densities. Smaller components having a size below a user-defined threshold are merged.
Latest IBM Patents:
- INTERACTIVE DATASET EXPLORATION AND PREPROCESSING
- NETWORK SECURITY ASSESSMENT BASED UPON IDENTIFICATION OF AN ADVERSARY
- NON-LINEAR APPROXIMATION ROBUST TO INPUT RANGE OF HOMOMORPHIC ENCRYPTION ANALYTICS
- Back-side memory element with local memory select transistor
- Injection molded solder head with improved sealing performance
1. Field of the Invention
The present invention relates generally to mining and learning network data. More specifically, the present invention describes methods and systems for supervised network clustering using densities associated with nodes and extracting node components from the network, based on using thresholds on densities.
2. Description of the Related Art
Network data has become increasingly popular, because of the increasing proliferation of social and information networks. A significant amount of research has been devoted to the problem of mining and learning network data. In many scenarios, a subset of the nodes in the network may have labels associated with them, and this information can be effectively used for a variety of clustering and classification applications.
In the context of the present invention, a network generally refers to a group of entities connected by links. This is a useful abstraction for many real-world scenarios, such as computer routers, pages on a website, or the participants in a social network. The nodes refer to the individual entities (e.g., routers, pages, participants) which are connected by links, which could either be communication links, hyperlinks, or social network friendship links.
The useful properties of such nodes can be captured by labels, which are essentially drawn from a small set of keywords describing the node. For example, in a social network of researchers, the label on the node could correspond to their topic area of interest. Such labels can provide useful background knowledge for a variety of applications, including directing a clustering process in different ways, depending upon the nature of the underlying application.
On the other hand, the available labels may often be noisy, incomplete, and are often partially derived from unreliable data sources. Many of the underlying clusters in the network may also not be fully described from such information, and even when the labels for a particular kind of desired category are available, they may represent an extremely small subset of the nodes. Nevertheless, such noisy, sparse, and incomplete information can also be useful in some parts of the network, and should therefore not be ignored during clustering.
SUMMARY OF THE INVENTIONIn view of the foregoing and other exemplary problems, drawbacks, and disadvantages of the conventional methods and systems, an exemplary feature of the present invention is to provide a new method and structure for supervised network clustering.
Another exemplary feature of the present invention is to provide a highly adaptive network clustering algorithm.
Another exemplary feature of the present invention is to provide an approach for constraining the nodes to belong to specific clusters, which is particularly useful in cases where prior knowledge is available for directing the clustering process.
In a first exemplary aspect of the present invention, described herein is a method for supervised network clustering, including receiving and reading node labels from a plurality of nodes on a network, the network being defined as a group of entities interconnected by links; using the node labels to define densities associated with the nodes; extracting node components from the network, based on using thresholds on densities; and merging smaller components having a size below a user-defined threshold.
In a second exemplary aspect of the present invention, also described herein is a method of clustering, including receiving and reading node labels from a plurality of nodes on a network, as executed by a processor on a computer having access to the network, the network defined by a group of entities connected by links; calculating a random-walk-based probability for each node on the network, to define densities associated with the nodes; and defining clusters of nodes in the network based on the densities.
In a third exemplary aspect of the present invention, also described herein is a method of clustering, including calculating a density associated with a plurality of nodes on a network, as executed by a processor on a computer having access to the network, the network defined by a group of entities connected by links; and defining clusters of nodes in the network based on the densities.
In a fourth exemplary aspect, also described herein is an apparatus, including a processor, and a memory device, the memory device storing therein a set of machine-readable instructions permitting the processor to execute a method of supervised network clustering, the method including receiving and reading node labels from a plurality of nodes on a network, as executed by a processor on a computer having access to the network; using the node labels to define densities associated with the nodes, extracting node components from the network, based on using thresholds on densities, and merging smaller components having a size below a user-defined threshold.
In a fifth exemplary aspect, also described herein is a server including an input port to receive information concerning nodes on a network and a processor, wherein the processor receives, via the input port, and reads node labels from a plurality of nodes on the network, the network defined by a group of entities connected by links, calculates a random-walk-based probability for each node on the network, to define densities associated with the nodes, and defines clusters of nodes in the network based on the densities.
In a sixth exemplary aspect, also described herein is a computer including a processor; and a memory device, the memory device storing a set of computer-readable instructions for the CPU to execute a method of clustering, the method including calculating a density associated with each of a plurality of nodes on a network, as executed by a processor on a computer having access to the network, the network defined by a group of entities connected by links, and defining clusters of nodes in the network based on the densities.
Other aspects, features and advantages of the invention will be more fully apparent from the ensuing disclosure and appended claims.
These and other advantages may be achieved with the present invention.
The foregoing and other exemplary purposes, aspects and advantages will be better understood from the following detailed description of an exemplary embodiment of the invention with reference to the drawings, in which:
The learning process for a network, as meaning in the context of the present invention a group of entities interconnected by links, can take on many forms, depending on the level of supervision in the learning process:
-
- At one end of the spectrum, no supervision may be available in the form of node labels. This problem is equivalent to the unsupervised network clustering problem.
- In many cases, some labels may me available at the nodes, which provide the partial supervision necessary for the clustering process. However, many other node clusters may also exist, which are not necessarily related to the labels on the nodes. This problem has remained largely unexplored for the case of structural data. The other end of the spectrum consists of the fully supervised learning or collective classification scenario, in which all the unlabeled nodes need to be classified, based on the current pattern of labeling of a small subset of the nodes.
For a given network application, the most suitable level of supervision may depend upon the underlying data and the task at hand. In most scenarios involving very large networks, a plethora of partial labels may be available in order to supervise the cluster creation process. Some examples are as follows:
-
- In a scientific community network, it may be possible to label a small subset of nodes, depending upon the area of interest of the particular academic.
- In a movie information network such as the Internet Movie Database (IMDb), containing both movies and actors, the genre of the movie may be labeled, whereas the pre-dominant genre of an actor may not be available. This information can be used to direct the clustering process towards a scenario in which actors are clustered together not just by their linkage to one another, but also by their similarity in terms of the genre of the movie in which they may act.
- In a social network application, it may be desirable to cluster actor nodes based on their affinity to some set of products. While such labels may not be known across all the nodes, they may be available for some small subset of the nodes.
Furthermore, the social network for a given user may be defined in a variety of ways, depending upon their professional contacts, family, school, or alma-mater contacts. Interesting formation about the user (such as their school info) could be considered labels. These different communities of the same user are often quite different from one another and represent different segments of the user social life. The supervision process can help in focusing the community detection approach in a particular direction.
It is clear that such labels can be very useful for directing the clustering process in different ways, depending upon the nature of the underlying application. On the other hand, the available labels may often be noisy, incomplete, and are often partially derived from unreliable data sources. Many of the underlying clusters in the network may also not be fully described from such information, and even when the labels for a particular kind of desired category are available, they may represent an extremely small subset of the nodes. Nevertheless, such noisy, sparse, and incomplete information can also be useful in some parts of the network, and should therefore not be ignored during clustering.
In contrast, the present invention addresses these issues by providing a fully adaptive approach, in which the level of supervision for network clustering can be controlled adaptively depending upon the application at hand. The two extreme versions of this scheme can perform either purely unsupervised clustering or fully supervised collective classification.
One challenge which has recently been observed with network clustering is that different regions of the data have different levels of density in the network, as a result of which homogeneous clustering algorithms tend to create unusually large and incoherent clusters containing a significant percentage of the nodes from the network. This means that the link densities in different regions of the social network may be quite different. In such a scenario, the use of global analysis can either construct very small communities in sparse local regions, or report large and incoherent communities in dense regions. Therefore, it is important to use local structural analysis for determining the relevance of communities in a social network. Furthermore, the topological shape of the clusters in a graph may be arbitrary, and is not necessarily spherical, as is implied by distance-based clustering algorithms in networks.
The present invention demonstrates a density-based approach, which is able to satisfy both goals. Thus, it provides a highly adaptive network clustering algorithm, which can discover clusters of varying shape and density, and also incorporate different levels of supervision in the clustering process.
Referring now to the drawings, and more particularly to
Initially, a summary of the present invention on supervised network clustering is provided, including a discussion of how a density-based model can be used for supervised network clustering. In this discussion, an undirected network G=(N, A) is assumed, in which N is the set of nodes, and A is the set of edges. It is assumed that the number of nodes in N is n. In many applications, the edges in the network may be associated with a weight, which indicates the strength of the relationship.
We further assume that the weight of edge (i,j) is denoted by w_{ij}. For example, in an author-relationship network, the weights might represent the number of papers authored by a pair of individuals. In many network applications, the weight w_{ij} is assumed to be 1, though the present invention allows the use of a weight for greater generality, if needed. It is assumed that the subset N_s of nodes are labeled, and that there are 1 different labels denoted by {1 . . . 1}. Therefore, all nodes in N_s are labeled with one of the values drawn from {1 . . . 1}, whereas the remaining nodes in N-N_s) are unlabeled.
An exemplary goal of the present invention is to partition the nodes in $N$ into $k$ different sets of nodes C—1 . . . C_k (i.e., clusters). In addition, we have a set of small subgraphs O which are referred to as the outlier set. Thus, we have N=C—1 U C—2 U . . . U C_k U O. We note that the labeling of the set of nodes in N_s can be used to supervise the creation of the clusters C—1 . . . C_k at a variety of different levels depending upon the application at hand. For a given node i, we assume that the edges incident on it are denote by I(i).
An exemplary overall idea of the supervised clustering approach of the present invention is to design a density-based method in which clusters are defined in terms of density-connected sets of nodes. A density-connected pair of nodes is one in which a path of nodes exists between the pair, such that each node has density above a pre-defined threshold. One natural advantage of density-based methods is that they are not restricted to particular topological shapes of the clusters, as are implied by distance-based methods. On the other hand, the concept of density is much more challenging to define in the context of structural data.
The density at a node is defined in terms of random walk probabilities. A more intuitive way of understanding the clustering process is in the context of a page-rank style random-walk process in which a surfer traverses the different nodes in the network by randomly picking any neighbor of a node during the walk. The page-rank random-walk concept has been used in the Google search engine and is described in more detail below. The density of a node, as used in the present invention, is essentially defined in an identical way to the page-rank computation.
Specifically, the density of a node i is defined as the steady-state probability that a random surfer on the network with a pre-defined set of reset probabilities visits node $i$ at any given transition. Intuitively, a random surfer on the network (upon entering a cluster), tends to get trapped in the cluster because the nodes in this dense region tend to have much higher visit probability than the surrounding nodes with lower visit probability. Therefore, a natural way of demarcating the boundaries of this cluster would be to exclude nodes from the cluster for which the density is below a given threshold, and then considering only connected regions of these high density nodes as candidates for a cluster.
Before discussing the clustering process in more detail, we will introduce the fundamentals of random-walk computation. In the random-walk process, at any given step, the random surfer either transitions to any node j adjacent to i with probability proportional to p_{ij}=w_{ij}Λsum_{j\in I(i)} w_{ij}, or it resets to a random node in the network with probability bias vector (or personalization vector)\gamma. Thus, the conditional probability of transition to node i (in case of a reset) is denoted by gamma(i). By picking the value of gamma(i) to be consistent with the different class labels, it is possible to perform the supervision process effectively.
The CPU processes the graph continuously over time, and uses the main memory 20, for intermediate book keeping of the statistics. These statistics may eventually be sometimes stored on the disk. The clustering of the stream, as discussed in
The first step of density estimation is performed in step 210. This step is performed with the use of the labels from the underlying data. The labels provide help in supervising the density estimation process. This step will be discussed in detail in
The next step is to find connected components from these densities. This is performed in step 220 in
Finally in step 230, the smaller connected components are merged, in order to create the larger clusters from the underlying network.
Therefore, in one exemplary embodiment, our approach generates the different thresholds in an iterative process. In order to generate the threshold for the density connected components, we use the average density over all the nodes as a threshold for the estimation process. In step 320, we remove all the nodes in the network based on this density-threshold qualification. The removal of these nodes reduces the number of remaining nodes in the network. In step 330, we check if a sufficient number of nodes are still remaining in the network. If this is indeed the case, then we update the density threshold in step 340. This updated density threshold is the density across the remaining nodes in the network. After updating the density threshold, we remove the nodes based on the density threshold qualification, and go back to step 320. This process is continually repeated, until the remaining network does not contain a sufficient number of nodes.
This paper describes the PageRank algorithm of the Google search engine, described in this paper by its two cofounders, as follows:
“2.1 PageRank: Bringing Order to the Web
The citation (link) graph of the web is an important resource that has largely gone unused in existing web search engines. We have created maps containing as many as 518 million of these hyperlinks, a significant sample of the total. These maps allow rapid calculation of a web page's “PageRank”, an objective measure of its citation importance that corresponds well with people's subjective idea of importance. Because of this correspondence. PageRank is an excellent way to prioritize the results of web keyword searches. For most popular subjects, a simple text matching search that is restricted to web page titles performs admirably when PageRank prioritizes the results (demo available at google.stanford.edu). For the type of full text searches in the main Google system, PageRank also helps a great deal.
2.1.1 Description of PageRank Calculation
Academic citation literature has been applied to the web, largely by counting citations or backlinks to a given page. This gives some approximation of a page's importance or quality. PageRank extends this idea by not counting links from all pages equally, and by normalizing by the number of links on a page.
PageRank is defined as follows:
We assume page A has pages T1 . . . Tn which point to it (i.e., are citations). The parameter d is a damping factor which can be set between 0 and 1. We usually set d to 0.85. There are more details about d in the next section. Also C(A) is defined as the number of links going out of page A. The PageRank of a page A is given as follows:
PR(A)=(1−d)+d(PR(T1)/C(T1)+ . . . +PR(Tn)/C(Tn))
Note that the PageRanks form a probability distribution over web pages, so the sum of all web pages' PageRanks will be one.
PageRank or PR(A) can be calculated using a simple iterative algorithm, and corresponds to the principal eigenvector of the normalized link matrix of the web. Also, a PageRank for 26 million web pages can be computed in a few hours on a medium size workstation. There are many other details which are beyond the scope of this paper.
2.1.2 Intuitive Justification
PageRank can be thought of as a model of user behavior. We assume there is a “random surfer” who is given a web page at random and keeps clicking on links, never hitting “back” but eventually gets bored and starts on another random page. The probability that the random surfer visits a page is its PageRank.
And, the d damping factor is the probability at each page the “random surfer” will get bored and request another random page. One important variation is to only add the damping factor d to a single page, or a group of pages. This allows for personalization and can make it nearly impossible to deliberately mislead the system in order to get a higher ranking. We have several other extensions to PageRank, again see [Page 98].
Another intuitive justification is that a page can have a high PageRank if there are many pages that point to it, or if there are some pages that point to it and have a high PageRank. Intuitively, pages that are well cited from many places around the web are worth looking at. Also, pages that have perhaps only one citation train something like the Yahoo! homepage are also generally worth looking at. If a page was not high quality, or was a broken link, it is quite likely that Yahoo's homepage would not link to it.
PageRank handles both these cases and everything in between by recursively propagating weights through the link structure of the web.”
Thus, from the above-recited passages describing the Google search engine, it can be seen that the prior art teaches to calculate the random-walk-based probabilities of nodes for purpose of ranking pages. In contrast, the present invention uses the random-walk-based probabilities for purpose of calculating clusters of nodes.
Returning now to the present invention, we note that many of the component clusters found by the present invention's algorithm may be smaller than a user-defined threshold. Such clusters need to be merged into larger clusters, as discussed in step 230 of
This is performed in step 610. In step 620, we check if at least one merge did occur in the last iteration. If at least one merge did occur, we go back to step 610. Otherwise, we report the current components as relevant clusters in step 630, and terminate.
Exemplary Hardware Implementation
The CPUs 711 are interconnected via a system bus 712 to a random access memory (RAM) 714, read-only memory (ROM) 716, input/output (I/O) adapter 718 (for connecting peripheral devices such as disk units 721 and tape drives 740 to the bus 712), user interface adapter 722 (for connecting a keyboard 724, mouse 726, speaker 728, microphone 732, and/or other user interface device to the bus 712), a communication adapter 734 for connecting an information handling system to a data processing network, the Internet, an Intranet, a personal area network (PAN), etc., and a display adapter 736 for connecting the bus 712 to a display device 738 and/or printer 739 (e.g., a digital printer or the like).
In addition to the hardware/software environment described above, a different aspect of the invention includes a computer-implemented method for performing the above method. As an example, this method may be implemented in the particular environment discussed above.
Such a method may be implemented, for example, by operating a computer, as embodied by a digital data processing apparatus, to execute a sequence of machine-readable instructions. These instructions may reside in various types of signal-bearing storage media.
Thus, this aspect of the present invention is directed to a programmed product, comprising signal-bearing storage media tangibly embodying a program of machine-readable instructions executable by a digital data processor incorporating the CPU 711 and hardware above, to perform the method of the invention.
This signal-bearing storage media may include, for example, a RAM contained within the CPU 711, as represented by the fast-access storage for example. Alternatively, the instructions may be contained in another signal-bearing storage media, such as a magnetic data storage diskette 800 (
Whether contained in the diskette 800, the computer/CPU 711, or elsewhere, the instructions may be stored on a variety of machine-readable data storage media, such as DASD storage (e.g., a conventional “hard drive” or a RAID array), magnetic tape, electronic read-only memory (e.g., ROM, EPROM, or EEPROM), an optical storage device (e.g. CD-ROM, WORM, DVD, digital optical tape, etc.), paper “punch” cards, or other suitable signal-bearing storage media including memory devices in transmission hardware, communication links, and wireless, and including different formats such as digital and analog. In an illustrative embodiment of the invention, the machine-readable instructions may comprise software object code.
As is readily apparent from the above description, the present invention discusses a new method for supervised network clustering which can be useful for constraining the nodes to belong to specific clusters. This is particularly useful in cases where prior knowledge is available for directing the clustering process.
Although the present invention has been described as an exemplary embodiment, it should be apparent that variations of this exemplary embodiment are possible and considered as included in the present invention.
For example, rather than using the server 5, the invention could also be implemented as a user-interactive application, in which the user interactively labels nodes as relevant to a particular group.
As another example of a possible modification, if desired, the user may even look at the output of the clustering and further modify the labels.
Therefore, it is noted that, Applicant's intent is to encompass equivalents of all claim elements, even if amended later during prosecution.
Claims
1. A method of supervised network clustering, said method comprising:
- receiving and reading node labels from a plurality of nodes on a network, as executed by a processor on a computer having access to said network, said network being defined as a group of entities interconnected by links;
- using said node labels to define densities associated with said nodes;
- extracting node components from the network, based on using thresholds on densities; and
- merging smaller components having a size below a user-defined threshold.
2. The method of claim 1, wherein a random walk process is used to define the densities.
3. The method of claim 2, wherein a restart vector associated with the random walk is defined on a basis of node labels.
4. The method of claim 1, wherein density-connected nodes above a given threshold are determined as initial components.
5. The method of claim 4, wherein the density-connected nodes are defined as nodes between a path that exists in which all nodes have densities greater than the threshold.
6. The method of claim 1, wherein smaller components having a size less than the user-defined threshold are merged with larger components.
7. The method of claim 6, wherein each smaller component is merged to the component with which it has a largest number of connections.
8. The method of claim 4, wherein said clustering is iterative and continues until no further merging of small clusters occurs.
9. The method of claim 8, wherein a threshold for an iteration comprises an average density over all remaining nodes.
10. The method of claim 1, as embodied as a set of computer-readable instructions tangibly embodied on a non-transitory storage medium.
11. A method of clustering, said method comprising:
- receiving and reading node labels from a plurality of nodes on a network, as executed by a processor on a computer having access to said network, said network defined as a group of entities interconnected by links;
- calculating a random-walk-based probability for each said node on said network, to define densities associated with said nodes; and
- defining clusters of nodes in said network based on said densities.
12. The method of claim 11, wherein said clusters are extracted based on using a threshold on said densities.
13. The method of claim 12, further comprising merging smaller clusters with sizes below a user-defined threshold.
14. The method of claim 12, wherein said cluster extraction comprises an iterative process.
15. The method of claim 14, wherein said threshold initially comprises an average density of said network.
16. The method of claim 11, as embodied as a set of computer-readable instructions tangibly embodied on a non-transitory storage medium.
17. The method of claim 16, wherein said non-transitory storage medium comprises one of:
- a Random Access Memory (RAM) device of a computer, as storing said computer-readable instructions for a program currently executing on said computer;
- a Read Only Memory (ROM) device of a computer, as storing said computer-readable instructions for a program that can selectively be executed by said computer;
- a standalone memory device storing said computer-readable instructions for a program that can selectively be uploaded onto a memory device in a computer; and
- a memory device associated with a computer on the network, as storing said computer-readable instructions for a program that can selectively be downloaded to a memory device of another computer on said network.
18. A method if clustering, said method comprising:
- calculating a density associated with a plurality of nodes on a network, as executed by a processor on a computer having access to said network, said network defined as a group of entities interconnected by links; and
- defining clusters of nodes in said network based on said densities.
19. The method of claim 18, wherein said densities are calculated as a random-walk-based probability for each said node on said network.
20. The method of claim 18, further comprising merging smaller cluster components having a size below a user-defined threshold.
Type: Application
Filed: Aug 10, 2012
Publication Date: Feb 13, 2014
Applicant: International Business Machines Corporation (Armonk, NY)
Inventor: Charu C. Aggarwal (Yorktown Heights, NY)
Application Number: 13/572,179
International Classification: G06F 15/173 (20060101);