Spelling suggestions: "subject:"probabilistic cographs"" "subject:"probabilistic bigraphs""
1 |
Dense subgraph mining in probabilistic graphsEsfahani, Fatemeh 09 December 2021 (has links)
In this dissertation we consider the problem of mining cohesive (dense) subgraphs in probabilistic graphs, where each edge has a probability of existence. Mining probabilistic graphs has become the focus of interest in analyzing many real-world datasets, such as social, trust, communication, and biological networks due to the intrinsic uncertainty present in them.
Studying cohesive subgraphs can reveal important information about connectivity, centrality, and robustness of the network, with applications in areas such as bioinformatics and social networks. In deterministic graphs, there exists various definitions of cohesive substructures, including cliques, quasi-cliques, k-cores and k-trusses. In this regard, k-core and k-truss decompositions are popular tools for finding cohesive subgraphs. In deterministic graphs, a k-core is the largest subgraph in which each vertex has at least k neighbors, and a k-truss is the largest subgraph whose edges are contained in at least k triangles (or k-2 triangles depending on the definition). The k-core and k-truss decomposition in deterministic graphs have been thoroughly studied in the literature. However, in the probabilistic context, the computation is challenging and state-of-art approaches are not scalable to large graphs. The main challenge is efficient computation of the tail probabilities of vertex degrees and triangle count of edges in probabilistic graphs. We employ a special version of central limit theorem (CLT) to obtain the tail probabilities efficiently. Based on our CLT approach we propose peeling algorithms for core and truss decomposition of a probabilistic graph that scales to very large graphs and offers significant improvement over state-of-the-art approaches. Moreover, we propose a second algorithm for probabilistic core decomposition that can handle graphs not fitting in memory by processing them sequentially one vertex at a time. In terms of truss decomposition, we design a second method which is based on progressive tightening of the estimate of the truss value of each edge based on h-index computation and novel use of dynamic programming. We provide extensive experimental results to show the efficiency of the proposed algorithms.
Another contribution of this thesis is mining cohesive subgraphs using the recent notion of nucleus decomposition introduced by Sariyuce et al. Nucleus decomposition is based on higher order structures such as cliques nested in other cliques. Nucleus decomposition can reveal interesting subgraphs that can be missed by core and truss decompositions. In this dissertation, we present nucleus decomposition for probabilistic graphs. The major questions we address are: How to define meaningfully nucleus decomposition in probabilistic graphs? How hard is computing nucleus decomposition in probabilistic graphs? Can we devise efficient algorithms for exact or approximate nucleus decomposition in large graphs?
We present three natural definitions of nucleus decomposition in probabilistic graphs: local, global, and weakly-global. We show that the local version is in PTIME, whereas global and weakly-global are #P-hard and NP-hard, respectively. We present an efficient and exact dynamic programming approach for the local case. Further, we present statistical approximations that can scale to bigger datasets without much loss of accuracy. For global and weakly-global decompositions we complement our intractability results by proposing efficient algorithms that give approximate solutions based on search space pruning and Monte-Carlo sampling. Extensive experiments show the scalability and efficiency of our algorithms. Compared to probabilistic core and truss decompositions, nucleus decomposition significantly outperforms in terms of density and clustering metrics. / Graduate
|
2 |
Analysis and Optimization of Communication Networks with Flow RequirementsLange, Thomas 24 April 2019 (has links)
In this thesis, we will study the concept of k-edge connected and k-connected reliability.
There, vertices are modelled as fail-safe and edges fail stochastically independent. For a fixxed k, the network is then considered operational when each pair of vertices has k edge disjoint or internally disjoint paths, respectively, connecting them in the surviving subnetwork. Thus, the property of being operating covers the connectivity of the surviving graph together with some minimum bandwidth.
We study essential and irrelevant edges for those reliability measures. Further, we study a splitting approach to transform the reliability of the graph into the probability that subgraphs have a certain connectivity. We also extend an approximation algorithm of Karger from the All-Terminal Unreliability to k-edge connected Unreliability and study the k-edge connected Reliability for some special graph classes, namely graphs with restricted treewidth, edge-transitive graphs and the complete graph.:1 Introduction
2 Monotone Systems
2.1 Monotone Binary Systems
2.2 Monotone Multistate Systems
3 Graphs and Graph Operations
4 Higher Connectivity
4.1 Connectivity Number and Edge-Connectivity Number
4.2 Algorithms
5 Essential and Irrelevant Edges
6 Probabilistic Graphs and Reliability Measures
7 Reductions
8 Splitting
8.1 Some Special Cases for Small Separators/Cuts
8.2 Generalization to Arbitrary Separators
8.3 Constructing the Splitting Classes for 2-ec, 3-ec and 2-vc
8.4 Minimality
8.5 A Lattice-based Approach
9 An Approximation Scheme
9.1 Definition of Approximation Algorithms
9.2 The FPRAS for All-Terminal-Unreliability
9.3 Improved Bound for the Number of alpha-cuts
9.4 Extension to k-edge-connected Unreliability
10 Special Graph Classes
10.1 Graphs with Bounded Treewidth
10.2 Edge-Transitive Graphs
10.3 The Complete Graph
11 Future Research
12 Summary
|
Page generated in 0.0757 seconds