The critical node detection problem (CNDP) aims to fragment a graph G=(V,E) by removing a set of vertices R with cardinality |R|≤k, such that the residual graph has minimum pairwise connectivity for user-defined value k. Existing optimization algorithms are incapable of finding a good set R in graphs with many thousands or millions of vertices due to the associated computational cost. Hence, there exists a need for a time- and space-efficient approach for evaluating the impact of removing any v∈V in the context of the CNDP. In this paper, we propose an algorithm based on a modified depth-first search that requires O(k(|V|+|E|)) time complexity. We employ the method within in a greedy algorithm for quickly identifying R. Our experimental results consider small- (≤250 nodes) and medium-sized (≤25,000 nodes) networks, where it is possible to compare to known optimal solutions or results obtained by other heuristics. Additionally, we show results using six real-world networks. The proposed algorithm can be easily extended to vertex- and edge-weighted variants of the CNDP.
Hinweise
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
MV devised and analyzed experiments, and drafted the initial manuscript. DA devised experiments and edited the manuscript. Both authors read and approved the final manuscript.
Introduction
Detecting important or critical vertices in a graph/network has many important applications. Depending on the context, these critical vertices/nodes may be used to promote or mitigate a diffusive process that is acting upon the network. If promoting a spreading process, such as to spread market advertisements or public health warnings, the notion of ‘critical’ refers to the identification of individuals who are most likely to be influential spreaders and maximally permit information spread through the network. In such cases, the selected individuals may be targeted for demonstrations and promotions or invited to public events. For mitigation contexts, such as stopping the spread of a computer virus or for the construction of stable power delivery networks, the identified vertices are those whose removal from the graph will maximally limit diffusive spread.
Numerous definitions of what a critical node ‘is’ have been previously investigated, including junctions in cell-signaling or protein-protein networks [1], highly influential individuals [2], smart grid vulnerability [3], targeted vaccination for pandemic prevention [4,5], or keys to decipher brain functionality [6]. In some contexts, an accurate mathematical definition for a critical node, particularly for highly complex systems such as the brain [7], may not yet exist. It is important to note that both promotion and mitigation can often be defined in a mathematically similar manner. In this paper, we focus on the context of mitigation; however, the results are typically applicable for problems where the goal is to maximally aid the diffusive process [8-11].
Anzeige
More specifically, in this paper, we consider the critical node detection problem (CNDP) as defined by [12]. Given graph G=(V,E) where |V|=n and |E|=m, ascertain (a typically small) subset of vertices, R⊆V, |R|≤k, whose removal leaves the residual graph G∖R with minimum pairwise connectivity, i.e.,
where the sum is over all connected components \(\mathcal {C}_{i}\) of the residual graph, and \(|\mathcal {C}_{i}|\) indicates the number of vertices in component \(\mathcal {C}_{i}\). The optimal network is therefore one that is maximally fragmented and simultaneously minimizes the variance among the number of vertices in the connected components. That is, the residual network will contain a relatively large set of connected components, each containing a similar number of vertices. This problem has been shown to be \(\mathcal {NP}\) hard [12,13]. Figure 1 highlights an example graph before and after removing all of its cut vertices (as an illustrative and simple strategy for detecting critical nodes) and the associated CNDP solution.
×
Related work
In general, graph partitioning has been an extremely active area of research for decades and we do not attempt a comprehensive review of those works here. Rather, in this section, we focus on the most related previous research to the CNDP and only highlight related works on graph partitioning for completeness.
The case where G is a tree structure has been examined and proven to be \(\mathcal {NP}\) complete when considering non-unit edge costs [14]. A polynomial-time dynamic programming algorithm with worst-case complexity \(\mathcal {O}\left (n^{3}k^{2}\right)\) for solving the problem with unit edge costs was also provided and applied to variants of the CNDP [15]. In [16], an integer linear programming model with a non-polynomial number of constraints is given and branch-and-cut algorithms were proposed. A reformulation of the CNDP that requires Θ(n2) constraints was recently shown and optimal solutions for small networks were ascertained [17,18].
Anzeige
Heuristic solutions without provable approximation bounds have also been investigated, but computation time for very large networks remains an important issue in these methods as well. The CNDP work of [12] utilizes a solution to the maximum-independent set (MIS) problem as a starting point for a local search, repeating the process until a desired termination criteria is reached. The algorithm is tested on a limited number of network structures with promising results. Two stochastic search algorithms are employed in [19] that permit solutions to significantly larger networks (having up to a few thousand vertices) to be solved within reasonable time and without significant resources. Randomized rounding-based algorithms have been also proposed in [20,21] but without approximation bounds (although, an instance-specific bound was derived). A \(\mathcal {O}(\log n \log \log n)\) pseudo-approximation algorithm was proposed in [13].
The CNDP is related to a variety of other graph partitioning problems in the literature; for instance, the minimum multi-cut problem, which aims to separate a set of source-sink pairs by removing a subset of minimum weighted edges. An \(\mathcal {O}(\log n)\) approximation for general graphs is provided in [22]. Another well-studied graph partitioning problem is the k-cut problem. Given an undirected weighted graph, the goal is to find a minimum cost set of edges that separates the graph into at least k connected components. An \(\mathcal {O}(2-2/k)\)-approximation algorithm has been devised for this problem [23]. Classical multi-way cut, multi-cut, and k-cut problems that include a budget constraint to limit the number edges or vertices that can be cut have been studied in [24], where the authors also propose the problem of maximizing the number of connected components. In [25,26], an \(\mathcal {O}(\sqrt {n})\)-approximation algorithm is presented for the sparsest cut, edge expansion, balanced separator, and graph conductance problems, all of which are based on the notion of graph partitioning. These approaches also minimize the size of the interface between the resulting components. Their approach is based on semi-definite relaxation to these problems in concert with expander flows and has influenced much subsequent research.
Among the many problems that have been defined, some of the most similar include the following. The goal of the minimum contamination problem is to minimize the expected size of contamination by removing a set of edges of at most a given cardinality [27]. A variant of this problem is also proposed with the goal of minimizing the proportion of vertices in the largest resulting network, and a bi-criteria algorithm is given that is able to achieve an \(\mathcal {O}\left (1+\epsilon,\frac {1+\epsilon }{\epsilon }(\log n)\right)\) approximation. In [28], a game-theoretic analysis is conducted that requires a solution to a generalization of the sum-of-squares partitioning problem [4]. Exact methods for link-based vulnerability assessment using edge disruptors have also recently been investigated [29,30].
The remainder of this paper is organized as follows. The proposed greedy algorithm and its motivation are provided in Section ‘The proposed algorithm’. Section ‘Experimental results’ provides experimental results on small- and medium-sized benchmark networks, as well as six real-world networks, and a comparison to the greedy MIS-based algorithm of [12] is conducted. Conclusions are then presented in Section ‘Conclusion’.
The proposed algorithm
In this section, we propose a greedy algorithm for the CNDP. Greedy algorithms are typically less computationally intensive than other strategies such as dynamic programming but usually sacrifice solution quality to attain this speed. One exception occurs in the case of maximizing a submodular function, where it was shown in [31] that unless \(\mathcal {P}=\mathcal {NP}\), a greedy algorithm will yield the optimal approximation. It is easy to verify that the CNDP is not a submodular function. However, the greedy approach is still an appealing framework in the CNDP context, especially for very large networks.
We first describe a linear-time algorithm to evaluate the impact of removing each v∈V and then use this within a greedy algorithm for determining a solution to Equation 1. The speed of the greedy algorithm is further enhanced by a priority-queue-based implementation that yields significant practical performance increases over a naive implementation. Our algorithm has a running-time complexity of \(\mathcal {O}(k(n+m))\) and is similar to the \(\mathcal {O}\left (kn^{2}\right)\) approach used for the maximum cascading algorithm in [3]. Both algorithms are based on Tarjan [32].
For large networks with many thousands or millions of vertices (and edges), a computationally efficient approach to minimizing the CNDP objective is required. Selecting a critical subset R⊂V from a single observation of the network may be easily deceived due to the influence of cut vertices, as indicated in Lemma 1 [12]. That is, selecting R in a sequential fashion may better allow for the discovery of a set that is more likely to fragment the network by detecting cut vertices that are not obvious unless a sequential approach is taken.
Lemma1.
Let M1 and M2 be two sets of partitions obtained by deleting D1 and D2 sets of vertices from graph G=(V,E), where |D1|=|D2|=k. Let L1 and L2 be the number of components in M1 and M2 respectively and L1≥L2. If \(\mathcal {C}_{h}=\mathcal {C}_{\ell }, \quad \forall h, \ell \in M_{1}\), then we obtain a better objective function value by deleting the set D1 (where \(\mathcal {C}_{h}\) is the number of vertices in connected component h).
Thus, we propose the sequential greedy approach shown in Algorithm 1. At each iteration, the vertex whose removal will have the largest decrease on the objective function (Equation 1) is selected for removal and added to set R, where f() computes the CNDP objective value. Computation of line 3 is a bottleneck to solving large-scale problems. Naively, it implies removal of each v∈V∖R and re-evaluation of the objective function, which is too computationally intensive.
Instead, we provide an \(\mathcal {O}(k(n+m))\) algorithm based on a modified depth-first search (DFS). On a practical note, the iterative (versus recursive) algorithm implementation of DFS should be used because sufficiently large networks will quickly encounter stack overflow errors during the search. Performing a DFS on G will construct an equivalent graph representation called a DFS-tree DFS(v) rooted at arbitrary v. Figure 2 provides an example of the conversion between an original graph G to a DFS tree rooted at v=0. Our subsequent solution is derived from observations of the DFS tree.
×
Observation1.
Ignoring back edges, a leaf vertex v∈V of a DFS tree has no children. Hence, the subtree rooted at v contains a single vertex whose deletion will not create any new connected components in the residual graph. That is, G and G∖{v} contain the same number of connected components.
Observation2.
Let δ(v) be the set of children vertices of v in the DFS tree, ignoring back edges. Then, the total number of vertices descendant from v through these children can be recursively defined as
$$ s(v) = \sum_{w \in \delta(v)} \begin{cases} s(w), & \text{if \textit{w} is an internal or root vertex} \\ 1, & \text{if \textit{w} is a leaf} \end{cases} $$
(2)
So, upon removing a vertex v∗∈V, the residual graph will contain at least the same number of connected components as the previous graph. Now, let T(v) denote the set of subtrees of v in a DFS tree, as represented by the immediate descendant of v. For instance, in Figure 2b, T(3)={4} and for ti∈T(v), |ti|=5. We can then make the following observations.
Observation3.
Each internal vertex v of the DFS tree will either be a cut vertex or not. Removing v∗=v will obviously result in an updated objective value, but if v∗ is a cut vertex then the residual graph G(V∖{v})will contain a nonempty T(v∗) because at least one new connected component will be split from G. Ignoring back edges from v∗, the contribution of the children subtrees to the new objective value is computed as
where |ti| is the number of vertices in DFS subtree ti∈T(v∗). As will be shown below, this sum can be straightforwardly computed for each vertex during the backtracking stage of DFS under the presumption that the current vertex being explored may be the next vertex removed.
Observation4.
If v is a cut vertex, then it will be identified as being so after visiting the entire subtree of each of its children. However, the order in which vertices are visited during the DFS does not guarantee that all non-descendant vertices in the graph will be explored before reaching v. Hence, the number of vertices in the ancestor DFS tree of v must also be recorded. This is accomplished by computing the difference between the total number of vertices in the graph and those descendant vertices of v, i.e. |V(v)|−s(v), where V(v) indicates the set of vertices reachable from v. V(v) is the size of the connected component to which v belongs and can be easily monitored at run time.
The above four observations imply that v∗ of Algorithm 1 can be computed in linear time by augmenting a DFS for identifying cut vertices to additionally calculate the impact of removing any vertex v∈V. That is,
which is accomplished during the backtracking phase of DFS. Pseudocode for implementing the approach is given in A Evaluating objective function in \(\mathcal {O}(|V|+|E|)\). Since DFS has running time complexity of O(n+m) and Equation 4 can be executed in constant time per node during the search, then the proposed greedy algorithm requires O(k(n+m)) complexity to remove k vertices from G.
We make two further observations that will yield significant practical improvements by storing each connected component in a priority queue, indexed by the vertex whose removal in the component will most minimize Equation 1.
Observation5.
Let Q⊆V∖{v∗} be the subset of vertices not reachable in graph G from vertex v∗. Then, it is not necessary to recompute the impact of removing any w∈Q from G∖{v∗} since v∗ and each w belong to different network components. That is, only vertices u∈V(v∗) must be re-examined if v∗ is deleted from G.
Observation6.
Each connected component \(\mathcal {C}_{i}\) of graph G can be identified by a root vertex associated with a DFS tree. For each \(\mathcal {C}_{i}\), there will exist a vertex \(v^{\prime }_{i}\) whose removal maximally decreases the objective function value. Let \(v^{\prime }_{i}\) be the root of the ith DFS tree associated with \(\mathcal {C}_{i}\). This requires no significant computational or memory overhead since upon deletion of v∗, the subgraph to which it belongs must be re-evaluated with respect to the objective function. The proposed algorithm in A Evaluating objective function in \(\mathcal {O}(|V|+|E|)\) will successfully determine \(v_{i}^{\prime }\).
Observations 5 and 6 indicate that further practical improvements are possible. Specifically, a priority queue can be utilized to store the set of connected components C, which are represented and ordered by their respective root vertices and their impact on the objective value if removed, respectively. For each Ci, removing its root vertex will most significantly decrease the CNDP objective value versus any other vertex in the same component. After the component to which v∗ belongs is removed from the queue, it will be re-evaluated using the modified DFS search and any newly resulting connected components will be added to the priority queue with an appropriate root node. Depending on the queue implementation, maintaining priority should require no more than \(\mathcal {O}(\log |C|)\). The per-iteration run time will significantly improve as the number of vertices in each connected component decreases. Effectively, the expected computation time will be \(\approx k \left (\frac {(n+m)}{|C|} + \log |C| \right)\), although the worst case remains \(\mathcal {O}(k(m+n))\). Algorithm 2 outlines the priority queue-based approach.
Experimental results
We evaluate the proposed algorithm on three sets of data sets: (1) small networks where optimal solutions or bounds are known [20], (2) medium-sized benchmark networks [19], (3) six real-world networks. The real-world benchmark networks and their properties are given in Table 1. All networks are unweighted and simplified before use (no self-loops or multi-edges).
|V| and |E| are the number of vertices and edges, ρ is the global clustering coefficient, δ is the diameter, and ξ is the degree assortativity.
We compare results to the most similar approach in literature [12], which is an \(\mathcal {O}\left (n^{2}m\right)\) greedy algorithm based on MIS and local search. Unless stated otherwise, the MIS-based algorithm is run for k iterations (for similarity to the proposed algorithm) with one iteration during the local search phase. For medium-sized real-world networks, our experimental results also compare Algorithm 1 to three centrality measures used in greedy sequential fashion such as in [38]. These are used only as a base-level comparison for the quality of the greedy approach. We consider node degree, PageRank [39], and authority score [40] centrality attacks. The computer used for simulations was a 3.4 GHz Intel i7 processor with 16 GB RAM, running Linux Mint Debian Edition.
Small networks
Even though the purpose of the proposed algorithm is to provide a means of determining a set of critical nodes in very large networks, an analysis of its performance on small networks is useful to gauge approximation ability. The running time in all cases in negligible (less than 1 s). The MIS-based approach is also implemented for comparison and required up to 30 s to complete. There are four network types, and the number of nodes to remove is varied: k={5,10,15,…,50}. Gurobi Solver 5.6 [41] is used to determine the upper and lower bound, within a 3,000-s time limit.
The results are summarized in Table 2 and Figure 3. The main difference between the two greedy algorithm results is that for highly connected networks, the MIS-based approach seems to be more useful; whereas the proposed algorithm is better suited for sparse networks.
Table 2
Results of the proposed greedy algorithm (SEQ) versus the MIS-based algorithm and upper (UB) and lower bound (LB) as determined by Gurobi 5.6 solver within 3,000 s
Table 3 presents results using 16 benchmark instances found in [19]. The table indicates the results of the PBIL algorithm [19], which is a population-based search, as well as sequential node removal using maximum node degree and highest PageRank as a heuristic objective, respectively. Of course, for these latter two experiments, the reported result indicates the CNDP objective of Equation 1. Values indicated in bold are the best results among those observed for each problem instance.
Table 3
Summary benchmark results comparing PBIL [19], degree and PageRank-based sequential algorithms, and the proposed greedy approach
Problem
|V|
|E|
K
PBIL
Degree
PageRank
Greedy
ErdosRenyi
235
350
50
6,700
1,086
1,953
3,011
ErdosRenyi
466
700
80
44,255
9,299
25,892
28,994
ErdosRenyi
941
1,400
140
225,576
123,947
113,752
116,135
ErdosRenyi
2,344
3,500
200
2,009,132
1,851,950
1,708,603
1,395,584
BarabasiAlbert
500
499
50
892
202
236
199
BarabasiAlbert
1,000
999
75
3,057
622
689
559
BarabasiAlbert
2,500
2,499
100
28,044
4,258
4,808
3,726
BarabasiAlbert
5,000
4,999
150
146,753
13,038
13,971
10,216
WattsStrogatz
250
1,246
70
13,768
16,110
16,110
16,110
WattsStrogatz
500
1,496
125
53,779
70,125
62,149
69,751
WattsStrogatz
1,000
4,996
200
308,596
319,600
319,600
319,600
WattsStrogatz
1,500
4,498
265
703,241
761,995
700,474
761,995
ForestFire
250
514
50
886
547
403
217
ForestFire
500
828
110
1,904
902
490
293
ForestFire
1,000
1,817
150
9,594
7,796
2,609
1,414
ForestFire
2,000
3,413
200
12,569
27,451
11,419
5,002
Italic values indicate best solution per problem instance. Problem instances are those in [19].
The benchmark results in Tables 2 and 3 reveal two insights. Firstly, sequential algorithms may perform poorly when compared to non-sequential algorithms in the instance of many potential solutions of equal quality. This is especially observed from the Watts-Strogatz network results. In these cases, the networks are highly connected and so there is unlikely to be many cut vertices. Moreover, as the networks are more sparse, the greedy approach becomes increasingly desirable. These two insights are founded in the same observation.
Observation7 (The Problem of Ties).
Highly connected graphs with few cut vertices will admit numerous candidate solutions, each with similar objective value. Due to sequential-based approaches lacking ability to investigate sets of potential solutions, these algorithms are best suited for sparse graphs.
The first consequence of The Problem of Ties concerns an explanation for the observed behavior, while the second leads to a conjecture.
Theorem1.
Assume G is sufficiently large and connected such that randomly removing k vertices is, with very high probability, unlikely to reveal v∗. Then, the worst-case problem instance for the proposed greedy algorithm will result when G=(V,E) contains no cut vertices initially but a single vertex v∗ exists whose removal immediately uncovers a sequence of k−1 residual graphs that each contain a cut vertex and whose union forms the optimal choice for set R.
Proof.
Suppose there exists an optimal solution for the CNDP on G with cut-set R, |R|=k. Moreover, assume the residual graph H=G∖R contains z connected components of equal size. Let a solution obtained by the greedy approach be composed of y components. Then, the objective function can be written as, for z≤y,
$$ z \left(\begin{array}{c}|\mathcal{C}_{i}|\\{2}\end{array}\right) \leq \sum_{i=1}^{z} \left(\begin{array}{c}|\mathcal{C}_{i}|\\{2}\end{array}\right) \leq \sum_{j=1}^{y} \left(\begin{array}{c}|\mathcal{C}_{j}|\\{2}\end{array}\right) $$
(5)
If k vertices are removed in both the optimal and greedy solutions, then if greedy never encounters v∗ or reveals any cut vertices,
$$ z \left(\begin{array}{c}|\mathcal{C}_{i}|\\{2}\end{array}\right) = z \left(\begin{array}{c}\frac{|V|-K}{z}\\{2}\end{array}\right) \quad \text{and } \quad \sum_{j=1}^{y} \left(\begin{array}{c}|\mathcal{C}_{j}|\\{2}\end{array}\right) = \left(\begin{array}{c}|V|-K\\{2}\end{array}\right) $$
(6)
The exact value of z will depend on the number of new connected components created as each vertex is removed (after the initial v∗). Of course, no better optimal solution can be constructed in this circumstance. Moreover, the greedy algorithm will attain the worst possible objective value since no connected components will be created as a result of removing k vertices.
Equation 6 also reveals insight into why highly connected networks are more difficult. That is, why the problem of ties is confounding. The difference between these two values will be smallest if many cut vertices exist, with few ties between solutions. Hence, we conjecture that reducing the Problem of Ties will result in higher quality solutions.
Real-world networks
To compare the quality of the greedy approach we vary k as 0.01,0.05,0.10,0.15, 0.20, and 0.25 proportion of each network, respectively. As with medium-sized networks, we compare results to methods of network attack (degree, PageRank, and Kleinberg’s authority score) in a similar greedy sequential approach. These strategies have been recognized as potentially useful for network fragmentation when considering other robustness measures such as minimizing the largest network component [38]. It should be noted that betweenness and closeness centrality, which are often also employed to test network vulnerability, are too computationally inefficient to be considered for these networks. We also compare the results to the MIS-based greedy heuristic. However, due to excessive computation time, we only report the result after one iteration of the MIS-based algorithm.
Table 4 compares the objective value for k={0.10n,0.20n} of the vertices in each network, respectively. The greedy approach outperforms the centrality-based strategies in all cases. The MIS-based approach is competitive with SEQ, but the computation time requirements limited its ability to discover a solution for the conmat network (see Table 5). Figure 4 plots the same experimental results over the entire range of k values for each network where the significance of the greedy solution quality is better highlighted versus the centrality-based approaches. The proposed algorithm is especially destructive for K<0.15n. All of the networks except the ego network exhibit a power-law degree distribution, which seems to be a major influence on the ability of fragmenting the networks for centrality-based approaches. The greedy algorithm significantly outperforms in these situations. Moreover, the global clustering coefficient, diameter, or degree assortativity do not seem to have such an obvious impact as the degree distribution does.
Table 4
Comparison of the CNDP objective value after removing 10% and 20% of vertices from each network in Table1
Problem
K
SEQ
Degree
PageRank
Authority
MIS
Comnat
2313
58,796,393
103,398,683
87,630,163
126,804,602
NA
4627
83,686
90,610
92,242
7,399,785
NA
Ego
404
2,717,347
5,339,614
3,816,109
6,320,816
2,192,636
808
1,848,740
2,070,535
2,886,709
3,438,031
903,441
Flight
294
322,527
484,331
467,962
1,014,305
77,777
588
1,457
1,698
1,715
1,567
2,626
Powergrid
494
22,182
51,508
212,369
56,815
25,253
988
3,639
4,580
14,744
3,771
5,378
Relativity
524
224,010
1,628,337
302,309
3,382,195
23,620
1,048
4,089
4,896
9,023
6,390
6,163
Oclinks
190
637,936
785,662
758,328
835,297
746,085
380
218,215
258,277
246,876
306,289
402,824
SEQ-based results are those obtained by the proposed algorithm. The MIS-based approach was unable to arrive at solutions within 40 h for the conmat networks.
Table 5
Comparing the run times (in milliseconds) of each approach
Problem
Slow greedy
SEQ
Degree
PageRank
Authority
MIS
Conmat
1,732,841
21,229
379
16,810
26,027
126,101,000
Ego
7,036
2,129
92
1,480
3,750
111,931,000
Flight
10,199
207
25
144
231
3,273,000
Powergrid
29,896
252
41
288
3,470
22,292,000
Relativity
43,679
421
44
180
601
43,016,000
Oclinks
2,301
143
11
114
360
1,734,000
The proposed greedy approach is considered for both cases of using the queue-based strategy (fast greedy) or not (slow greedy).
×
Running time (in milliseconds) for these networks was also investigated and shown in Table 5. In order to highlight the benefit of the priority queue-based solution, we sequentially remove vertices using both a naive greedy method that operates over the entire graph at each iteration (termed slow greedy) and the proposed fast greedy approach (SEQ), which will only consider vertices in the component that contained the most recently removed vertex. The significant improvement of the priority queue is obvious. The greedy approach requires similar time to run as PageRank, although optimizing our implementation may further reduce or surpass this gap. As expected, sequentially removing vertices based on node degree is by far the fastest method. The MIS-based approach is significantly more time consuming, which is expected based on its \(\mathcal {O}\left (n^{2}m\right)\) running time behavior.
Conclusion
In this paper, we proposed an efficient greedy heuristic for identifying critical vertices in networks whose removal leaves the residual network with minimum pairwise connectivity. We provided arguments for an upper-bound running time of \(\mathcal {O}(k(n+m))\), although the practical performance is significantly improved using a priority-queue-based strategy for storing connected components. We utilized both benchmark and larger networks with many thousands of nodes, where finding solutions using current approaches typically requires a significant amount of time. The resulting greedy algorithm is shown to yield better results than common centrality measures for large graphs while being computationally competitive with degree-based greedy vertex removal. The results on benchmark graphs led to the abstract construction of a worst-case input graph, which was a consequence of identifying The Problem of Ties. Interesting future work may aim to reduce the impact of this issue by incorporating additional information that is highly correlated with the objective in order to better identify potentially interesting candidate nodes for removal.
The algorithm proposed in this paper was given without any proof of approximation quality, only indicating the extreme circumstances of problem instance. Future work will prove this bound. Moreover, experimentation on different network types and much larger sized networks, including run time, should also be conducted. Potential improvements in run time may be attainable if within-connected component objective function evaluation was parallelized or a relationship between the nodes can be identified so that only an \(\mathcal {O}(n)\) process is required to update the impact of a vertex removal. Extensions to vertex- and edge-weighted variants of the CNDP are also possible.
Appendix
A Evaluating objective function in \(\mathcal {O}(|V|+|E|)\)
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0), which permits use, duplication, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
MV devised and analyzed experiments, and drafted the initial manuscript. DA devised experiments and edited the manuscript. Both authors read and approved the final manuscript.