Abstract
In the paper we consider the well-known Influence Maximization (IM) and Target Set Selection (TSS) problems for Boolean networks under Deterministic Linear Threshold Model (DLTM). The main novelty of our paper is that we state these problems in the context of pseudo-Boolean optimization and solve them using evolutionary algorithms in combination with the known greedy heuristic. We also propose a new variant of (1 + 1)-Evolutionary Algorithm, which is designed to optimize a fitness function on the subset of the Boolean hypercube comprised of vectors of a fixed Hamming weight. The properties of this algorithm suit well for solving IM. The proposed algorithm is combined with the greedy heuristic for solving IM and TSS: the latter is used to construct initial solutions. We show that the described hybrid algorithms demonstrate significantly better performance compared to the computational scheme combining the greedy heuristic with the classic variant of (1 + 1)-EA. In the experiments, the proposed algorithms are applied to both real-world networks and the random networks constructed with respect to well-known models of random graphs. The results show that the new algorithms outperform the competition and are applicable to TSS and IM under DLTM for networks with tens of thousands of vertices.
Introduction
Network science is a very fruitful and intensively developed area which features a lot of high quality publications. They can be broadly categorized into studies that analyze the properties related to network topology and papers that consider the properties of functions defined by networks. The former typically consider the so-called critical phenomena in networks and in this context they study various qualitative properties such as clustering, various consequences of the connectivity, such as a small world effect, degrees distributions, etc., as well as the models of random graphs that can be used to generate networks with desired properties. Since these questions do not constitute the main object of our study, let us cite in this context the reviews [1,2] and the book [3].
The scope of the present paper belongs to the second class of network science studies: the one that considers an existing network and analyzes various dynamic processes which occur inside said network. Thus the problems considered in the paper can be viewed as simple examples of agent-based modeling. Namely, we consider a network which models some collective: the network vertices interpret the members of collective (agents) and the edges interpret the connections between agents. With each network agent there is associated a special function that defines the degree of activity of this agent at each discrete time moment. The process in question models information dissemination in such a network, which can often be viewed as an activation process.
Among the central problems in the formulated context are the Influence Maximization problem (IM) and the related Target Set Selection problem (TSS). In just a few recent years hundreds of papers on these problems have been published, see e.g. [4–25]. In Sect 5 we will briefly review the main results achieved in these works. We would like to specifically note that the seminal paper [4] remains one of the main works on IM for more than 20 years. In that paper they performed the analysis of a number of key properties of activation processes in networks, and proved the combinatorial nature of both IM and TSS (e.g. showed that both are NP-hard). Also, in [4] it was demonstrated that under some models of information dissemination there exists an approximation algorithm for solving IM with constant multiplicative approximation factor of 1–e−1. This is the case for the so-called Linear Threshold Model and Independent Cascade Model. Both models are not deterministic in the base case, as the value of the objective function (influence in terms of [4]) is defined using some random variables. What is interesting, however, is that it is the non-deterministic nature of these models that is the reason for the existence of a 1–e−1-approximation algorithm for IM. In addition to that, [4] showed that under the deterministic variant of Linear Threshold Model (to which we further refer as to DLTM) IM already has tight arguments for inapproximability: specifically, if there exists an approximation algorithm with multiplicative constant factor for IM under DLTM, then P = NP.
Recall, that under DLTM the thresholds of vertices are fixed and therefore the influence spread follows a completely deterministic scenario. It means that its value is determined as the number of active vertices in the network at the moment when the activation process stops, or more precisely, when the activation algorithm does not make any inactive vertices active. DLTM looks to be quite reasonable since in applied scenarios the vertices’ thresholds usually have entirely practical interpretation and thus can be measured, similar to how it is done in the known threshold models of collective behavior [26].
Taking into account the high structural complexity of IM and TSS, the design of computational algorithms to solve these problems is a relevant direction of research. In the present paper, we propose several novel algorithms for finding an approximate solution for IM and TSS. In particular, we use as the basis of the proposed approaches the evolutionary algorithms which are applicable to solving optimization problems from a wide class, including the search for effective combinations of hyperparameters in Artificial Intelligence [27–29], etc. To justify the good accuracy of the proposed methods we find the exact solutions of network activation problems for networks of small size by reducing them to combinatorial problems, in particular to the Boolean satisfiability problem (SAT). The latter is solved using state-of-the-art SAT solvers which are actively used in combinatorial optimization and in the recent years find increasing use in the area of Explainable AI, see [30,31], etc.
The main novelty of our approach to IM and TSS solving is the combination of greedy algorithms with evolutionary algorithms and some additional heuristics. Specifically, the main contributions of our approach are the following:
We consider IM and TSS under DLTM in the context of the general pseudo-Boolean optimization problem which makes it possible to apply a wide variety of algorithms for their solving, including evolutionary algorithms;
We describe a new variant of (1 + 1)-Evolutionary algorithm ((1 + 1)-EA), which is designed to optimize a pseudo-Boolean fitness function on a subset of the Boolean hypercube of dimension n formed by the vectors of fixed size and apply this algorithm to IM and TSS;
Using the networks of small size and reduction of TSS to the Boolean Satisfiability problem (SAT) we experimentally show that the metaheuristic algorithms for IM and TSS often make it possible to find solutions that are close to exact solutions;
We propose the hybrid strategy that combines the well-known greedy heuristic with evolutionary algorithms and show that it is applicable to solving IM and TSS under DLTM for real-world networks with tens of thousands of vertices.
Let us give a brief outline of our paper. The following Sect 1 is the preliminaries which contains the main notions, some auxiliary results, and also describes basic principles according to which evolutionary algorithms can be employed to optimize pseudo-Boolean fitness functions. Next Sect 2 contains the description of the (1 + 1)-Weighted Evolutionary Algorithm designed to optimize arbitrary pseudo-Boolean fitness functions over the subset of the Boolean hypercube formed by vectors of fixed Hamming weight. Further this algorithm together with several other evolutionary algorithms is used to solve IM and TSS. In Sect 3 we state IM and TSS under DLTM in form of pseudo-Boolean optimization problems, and describe a number of auxiliary heuristics used together with evolutionary algorithms when solving IM and TSS. The following Sect 4 presents the results of two series of computational experiments. In the first one we use networks of small size and reduction of TSS to SAT to construct exact solutions and show that the proposed metaheuristic algorithms often find either exact solutions or get very close to them. In the second series, we demonstrate the practical applicability of proposed algorithms to solving IM and TSS under DLTM for real-world networks from the SNAP repository [32] with thousands and tens of thousands of vertices. The paper is concluded with Sect 5 containing a brief review of related works and Sect 6 with conclusions and discussion of future work.
1 Preliminaries
It is convenient to consider DLTM in the context of the so-called Synchronous Boolean Networks (SBN), introduced by Stuart Kauffman in his seminal paper [33].
1.1 Synchronous Boolean networks and DLTM
Let us consider an arbitrary network as a directed labeled graph , where V is the set of vertices, A is the set of directed edges called arcs, and L is the set of labels associated with vertices and arcs. Assume that G does not have loops or multiple arcs. For an arbitrary vertex we say that the vertex is a neighbor of v if , i.e. if there is an arc in G that goes from u to v. The set of all neighbors of the vertex v is called the neighborhood of this vertex, which we denote as .
Hereinafter, we suppose that each vertex from G is associated with a symbol from the set of size 2, which we denote by . If a vertex v is associated with the number 1 then we say that v is active, otherwise v is assigned with the number 0 and is inactive. Apparently, it was S. Kauffman who proposed for the first time in his work [33] to associate with G the parameter t which takes the values in the set and represents discrete time. The set of numbers from connected to each network vertex at some time t forms a Boolean vector of length n, . In this case, as it was shown by S. Kauffman, such Boolean vectors associated with the considered network at moments can be viewed as the states of a Discrete Dynamical System (DDS), in which the transitions between the states are specified by the deterministic rules, that do not depend on the specific time moment. S. Kauffman viewed the value from associated with vertex v at moment t + 1, t ≥ 0 as the value of some Boolean function of arity which takes as an input the values assigned to vertices from at time moment t. Therefore, all coordinates of the Boolean vector of length n representing the network state at time moment are recomputed synchronously as the functions of the coordinates of the vector corresponding to the network state at time t. The described model is known as Synchronous Boolean network or Kauffman network. We will refer to the Boolean functions associated with network vertices the way outlined above as weight functions.
Since the rules that define the transitions between the states of Boolean network are deterministic, then the DDS defined this way, to which we further refer as DDS G, has a finite number of states. It means that, given the countable number of time states, such a DDS will traverse the same states multiple times. The situations of such a sort are defined as attractors or cycles. More strictly, let us represent the state of DDS G at an arbitrary time moment t by a Boolean vector of length n denoted as . Consider the set of states such that j is the minimal possible number for which it holds that . In this case, we say that this set of states forms the cycle (attractor) of length j. If the states form a cycle of length 1, i.e. , then the state is called a fixed point of a considered DDS.
Hereinafter, by {0,1}n denote the set comprised of all binary vectors of length n. This set essentially is a Boolean hypercube of dimension n. Due to the deterministic nature of the weight functions of the vertices, we can view all 2n different Boolean vectors from {0,1}n as the states of DDS G. These states can be represented by the vertices of a directed graph called State Transition Graph (STG) and denoted by . An arc (a,b) in interprets the transition from the state of DDS G to the state . For two arbitrary states , , , such that there exists a path from to in we will say that is reachable from . All these notions are illustrated on Fig 1. For example, state (0110) is a fixed point. Coincidentally, it is reachable from all other states. On the other hand, state (0000) is not reachable from any other state.
Fig 1. Example of a Synchronous Boolean Network with 4 vertices and its State Transition Graph.

It is possible to specify the weight functions of the vertices using different ways: by truth tables (as in [33]), by Boolean formulas (as in Fig 1) or by some predicates that express the relations between the corresponding elements (typically, in numeric form).
Further, we will mostly deal with DDS G, in which each arc (u,v) is assigned a non-negative rational number . For correctness, assume that for each v at least one of the numbers is positive. Let us define the number as follows: . Let us connect with an arbitrary vertex v some fixed number θv ∈ [0, 1], called threshold of v. Now, we define the weight function of vertex :
| (1) |
The state of DDS G at the moment t = 0 is called initial and is specified separately. One can use any vector from {0,1}n as an initial. A vertex v is called active at moment t, if , otherwise it is called inactive at this moment.
DDS G defined this way corresponds to the Deterministic Linear Threshold Model (DLTM) in the sense of [4].
1.2 IM and TSS w.r.t. DLTM
The following theorem is widely known and it follows from the fact that at time moment the number of active vertices in the network is greater or equal than the number of active vertices at moment t. The latter is, in turn, ensured by the properties of weight functions (1).
Theorem 1. For an arbitrary DDS G defined above with weight functions of the kind (1) it holds that for any initial state there exists such that , reachable from is a fixed point of DDS G.
Hereinafter, for an arbitrary we will be interested in the smallest t such that is a fixed point reachable from . Let us denote the corresponding fixed point as . We will refer to the transition of DDS G from state to a fixed point as to activation process.
Definition 1. Consider an arbitrary DDS G w.r.t. DLTM defined above. For any vector define the influence as the number of active vertices in G at fixed point reachable from the initial state specified by α.
For an arbitrary vector denote by the Hamming weight [34] of the considered vector. For fixed denote by the set of all s.t. .
Definition 2. Influence Maximization problem (IM) w.r.t. DLTM for specific value of parameter k consists in finding the following vector:
| (2) |
In other words, with each vector α from one associates the value equal to the number of active vertices the network has in a fixed point reachable from α. We denote by such a vector that for any it holds that .
In the context of IM the set of vertices which are active in initial state α is often referred to as a target set. Thus, any vector from {0,1}n specifies some target set and vice versa. Taking this into account, we will use the term “target set" for an arbitrary considered as an initial state of DDS G.
The Target Set Selection (TSS) problem [4,35] is closely related to IM.
Definition 3. The TSS w.r.t. DLTM for the predefined number R, consists in finding a target set α of minimal Hamming weight such that .
Note, that while in IM one searches for a vector of particular weight k which maximizes the value of σ over , in TSS one is looking for a vector α of minimal weight in the whole hypercube {0,1}n, such that .
1.3 Greedy heuristics for IM and TSS w.r.t. DLTM
As said above, there exist effective approximate algorithms for solving IM w.r.t. non-deterministic variant of linear threshold model (to which we further refer as LTM), such as the one considered in [4].
The weight functions in the framework of LTM are the same as functions (1) but the thresholds of vertices are randomly chosen. In more detail, we consider a probabilistic experiment in which the sample space is [0,1]n (n-th Cartesian degree of the segment [0,1]). Each replica of [0,1] is connected to a separate vertex of network G and each [0,1] has a uniform distribution, according to which the threshold for the corresponding vertex is selected. For an arbitrary on [0,1]n a random variable is defined, the value of which for a random set of thresholds from [0,1]n is equal to the number of active vertices in reached by DDS G from the initial state α. Influence of specific is defined as the expected value of :
| (3) |
For any IM w.r.t. the considered model consists in finding
As it was shown in [4], the function (3) is submodular and hence the simple greedy algorithm from [36] can be used to solve this variant of IM, and it approximates the desired optimal solution with constant multiplicative factor equal to 1–e−1 in the following sense: if is the solution found by the greedy algorithm then the following inequality holds: (It should be noted that this is valid only under the assumption that the exact values of function (3) are given by some oracle).
On the other hand, according to [4], the function is in general not submodular and cannot be approximated with any constant multiplicative factor under the assumption that . Therefore, it is reasonable to apply various heuristics and metaheuristics for solving IM under DLTM.
Apparently, the first algorithm for solving IM and TSS specifically under DLTM is the greedy heuristic proposed in [37]. Let us give its brief description. This algorithm is quite similar to the greedy algorithm from [4]: that is, it starts with an empty set and at each subsequent step one vertex is added to the current target set for which the influence measure of this vertex is maximal. To define this measure, the concept of residual threshold of an arbitrary vertex is used, which we give below.
Consider an arbitrary state of DDS G given by a vector . Let us associate the following value with α and an arbitrary vertex :
| (4) |
The function in (4) is defined as follows:
Definition 4. We will call the value given by the formula (4) a residual threshold of vertex v in state α.
Obviously, if , then v is active in state α, otherwise it is inactive in this state. Let us now describe the greedy heuristic which is a small modification of the algorithm proposed in [37]. Given the above, we will view an arbitrary target set as both a concrete set of vertices in the network G and as a Boolean vector of length n, where n is the number of vertices in G: the ones in such a vector mark vertices in target set under consideration.
Let us construct target sets of Hamming weight k, , using a strategy similar to the one employed in [4]. The values of k will correspond to the steps of the algorithm. At each step, the Hamming weight of the current target set is increased by 1 compared to the previous step. Denote the target set at step k by Tk. For k = 0, we have . To transition from Tk to , we need to select some inactive vertex and construct . To choose such v, we compute the influence measure of vertices from .
In more detail, for each vertex , we do the following:
Construct a probing target set
Denote by the state of DDS G corresponding to ; considering as the initial state of DDS G, run activation process and obtain the fixed point ;
Select the set of vertices that were inactive in state but became active in state , and the set of vertices that remained inactive in state ;
-
For each vertex let us calculate the following value:
(5) Note that (5) is the value by which the residual threshold of vertex u decreases as we move from to . Consider the following value:(6) It is clear that (6) can be viewed as a measure of the v’s influence on the activation process: intuitively the larger the value of (6) is, the more v is involved in the process.
Let us go from Tk to , where is a vertex with the maximum value (6). For each , we will consider the Boolean vector that corresponds to the TS Tk constructed in the way described above as an approximation of defined according to (2).
The pseudocode of the algorithm is presented at Algorithm 1.
Algorithm 1. Greedy algorithm for solving IM/TSS.
![]() |
Note that the algorithm from [37] differs from the one described above only in that in [37] the activation process is run for some fixed number of time moments each time (specified by one of the algorithm parameters), without necessarily reaching a fixed point.
1.4 Finding exact solution to TSS using SAT solvers
It is easy to see that for relatively small networks one can find exact solutions of IM and TSS using exact combinatorial algorithms. In the present paper, for this purpose we employ the algorithms for solving the Boolean Satisfiability problem (SAT) [38], to which we refer as SAT solvers. Note that DDS G under DLTM defines a discrete function of the kind
| (7) |
where receives as an input an arbitrary , considered as the initial state of DDS G, and outputs a fixed point reachable from α. Obviously, if the binary representation of numbers of the kind and do not depend on n or are bounded polynomially from n, then the function (7) is computable in polynomial time in n. In this case, due to the Cook-Levin theorem [39,40] there exists a procedure with the runtime polynomial in n that reduces IM and TSS under DLTM to SAT. Recall that SAT [38] is a problem in which one has to determine the satisfiability of an arbitrary Boolean formula , that is, to answer the question whether there exists a set of values of the variables in whose substitution into converts this formula to 1 (True). SAT is a classical NP-complete and NP-hard problem [41], therefore, under the assumption that this problem cannot be solved in the general case in polynomial time. However, it is often the case that the SAT instances of large dimension can be solved in reasonable time with the help of modern SAT solvers [38]. This fact has stimulated the application of SAT solvers to an extremely wide class of combinatorial problems, which we will briefly discuss in Sect 5.
In the paper [42] the SAT solvers were applied to the study of the phenomena of conforming behavior in Kauffman networks. In essence, that paper solved TSS for DDS G under DLTM for functions of the kind (1), where the values were chosen from . Unfortunately, this approach is applicable only to networks with at most a few hundred vertices and, therefore, can not be used to analyze real-world networks of practical interest. However, one feature of this approach is that it makes it possible to construct exact solutions of TSS for networks of small size, and therefore gives us an opportunity to understand how far the solutions obtained by metaheuristic algorithms for these networks are from the exact ones. This will give us additional arguments in favor of the accuracy of the metaheuristics we use.
2 Evolutionary optimization
The main novelty of this paper is the use of metaheuristic optimization algorithms, specifically evolutionary algorithms, to solve IM and TSS. We will use such algorithms to optimize pseudo-Boolean functions which are naturally connected with IM and TSS.
Recall that a pseudo-Boolean function [43] is any function of the following form:
| (8) |
In many real-world problems, there is no information about the analytic properties of functions of the form (8), and in such cases it is reasonable to apply metaheuristic algorithms to find the optimum of F over the hypercube {0,1}n. In the role of such algorithms, one can use, for example, various local search strategies [44], but in our paper we employ the algorithms of evolutionary optimization [45] because they often show good results in application to similar problems from related areas.
Let us briefly present the main notions and constructions used below. The simplest evolutionary algorithm is (1 + 1)-EA [46]. It is based on the notion of random mutation: with an arbitrary Boolean vector : associate a sequence of independent Bernoulli trials [47] of length n: with a fixed probability of success p (called mutation rate) in each trial. For each : if , then the coordinate of vector α remains unchanged; if , then the coordinate is flipped, that is, or . This process yields the vector ; is said to be the result of a random mutation of α with mutation rate p. If α is the current vector and is the result of its random mutation such that (assuming that we consider the maximization problem for (8)) then becomes the new current vector; otherwise α remains as the current vector. Random mutations are applied to the current vector until some termination criterion is met (e.g., a limit on the number of mutations is exceeded).
Despite its simple nature, (1 + 1)-EA has many important theoretical properties. For example, it is easy to see that when , the expected value of random variable (Hamming distance between α and ) is 1. This property is very important because it means that, on average, (1 + 1)-EA behaves like a simple Hill Climbing (HC) algorithm [48] and thus can adjust to the landscape of function F, efficiently climbing to local extremums. On the other hand, unlike HC, (1 + 1)-EA is able to move with non-zero probability from an arbitrary point α to a point where the function (8) has a global extremum over {0,1}n.
A number of important theoretical properties of (1 + 1)-EA were presented in the seminal paper [49]. In particular, that paper introduced the notion of complexity for this algorithm and showed that when p = 1/n in the general case, the algorithm is extremely inefficient, since its complexity interpreted as the expected value of the number of random mutations required to reach for a number of special functions is for some constant . Thus, in the sense of this measure, (1 + 1)-EA has worse complexity even compared to the simple random walk in which we randomly choose α w.r.t. uniform distribution on {0,1}n. However, this result is a typical example of the worst case scenario whereas in application to many combinatorial problems (1 + 1)-EA is surprisingly effective and often performs better than more sophisticated algorithms. One of the reasons for this is the small expected number of flipped bits in one random mutation, which makes (1 + 1)-EA on average similar to HC [50]. Another advantage of (1 + 1)-EA is that it is easy implement and has a high speed of operation.
In several papers there have been proposed alternative variants of (1 + 1)-EA, the goal of which was to reduce the upper bound on its complexity. The (1 + 1)-Fast Evolutionary Algorithm proposed in [51] is among the best approaches of such a kind. It employs a parameter β and is usually denoted as (1 + 1)-FEAβ. The main idea of that algorithm is to adjust the mutation rate during its work. In more detail, before mutating each one observes the value of a special random variable δ with the spectrum . The distribution of δ (also called “power-law distribution ” [51]) is defined in the following manner:
| (9) |
In the Eq 9 β, is the algorithm’s parameter and is the so-called normalizing constant used to normalize probabilities: . Then, a single mutation in (1 + 1)-FEAβ is performed as follows:
Generate a value of according to the distribution
Apply the standard mutation operator with to .
The upper bound on the complexity of (1 + 1)- FEAβ is which is asymptotically significantly better than that for (1 + 1)-EA. Meanwhile, the expected value of flipped bits of (1 + 1)- FEAβ is a constant independent of n only when . For example in the case we have (see [51]).
Genetic algorithms (GA) [45] are often used to solve pseudo-Boolean optimization problems, and they can also be applied to IM and TSS. In this paper, we employ the following variant of GA.
Suppose that there is a set of Boolean vectors from {0,1}n, called population. In particular, we will talk about the current population and use the notation . The vectors , are referred to as individuals. Let us set the probability distribution on Pcurr in accordance with the following rules (assuming that we aim to maximize (8) over {0,1}n):
Next, we select l individuals with the best values of function (8) from Pcurr and add these individuals into new population Pnew. Then, in accordance with the distribution Dcurr, we select g random individuals from Pcurr and apply to each of them random mutation with mutation rate p = 1/n or the mutation variant from [51]. The obtained individuals are added to the new population Pnew. Finally, we construct h pairs of randomly chosen vectors from Pcurr w.r.t. distribution Dcurr and apply to each pair some variant of the crossover operator [45] (in the experiments we used standard two-point crossover) and add the resulting vectors to Pnew. The values of l,g,h are selected so that the relation is satisfied. After this the constructed population Pnew becomes Pcurr.
2.1 A new variant of (1 + 1)-EA
In this section we describe the algorithm that seeks an extremum of an arbitrary function of the kind (8) over . We will later apply this algorithm to solving IM and TSS.
Consider arbitrary : 0<k<n and the set formed by all vectors from the hypercube {0,1}n with Hamming weight k. Assume that we have an optimization problem for some pseudo-Boolean function of the kind (8) over . Our closest goal is to construct an algorithm for solving this problem which inherits the basic properties of (1 + 1)-EA. The main idea of this algorithm is to apply the standard (1 + 1) mutation to some part of , and then use the so-called correcting transformation on the other part. Let us present the details below.
By let us denote the length of vector α. For an arbitrary we split it into two parts . Vectors and are formed by the coordinates of α which are equal to 1 and 0, respectively. We assume that and . Next, we define the transformation of α as the transformation of vectors .
We will use the following notation: and . Apply the standard random mutation operator with rate to . If no bits of were flipped during the mutation then . Now, suppose that r bits in were flipped, r>0. Let us fix the coordinates of flipped bits in . We need to apply to such a transformation that the Hamming weight of the result of mutation of α will be equal to that of α. For this purpose, we choose r random coordinates in using the standard sampling without replacement [47], and flip the bits with these coordinates in α. Denote the resulting vector by . It is clear that . We denote the operator that maps onto itself as . For convenience, we will also denote the images of vectors and in w.r.t. as and . Let us refer to the operator as k-weight random mutation.
Note that we can employ the operator to perform transitions between vectors from similar to how the standard random mutation operator in (1 + 1)-EA is employed: if α is the current vector from , and , then becomes current only if (assuming that we aim to maximize F). The pseudocode of this algorithm is presented in Algorithm 2. Further we will refer to it as (1 + 1)-WEA (from ‘Weighted Evolutionary Algorithm’). Let us establish the main properties of the algorithm.
Algorithm 2. The pseudocode for (1 + 1)-WEA.
![]() |
Proposition 1. For any two vectors there exists a transition of the kind .
Proof: Consider two arbitrary vectors and outline the part in α. Without loss of generality, assume that vector is formed by ones and . Assume that the coordinates of have the numbers , and the coordinates of have the numbers with respect to the order (of coordinates) common for both α and . Let be the vectors formed by the coordinates with numbers in vector and numbers in vector respectively. Note that if , then , thus . Therefore, in this case the transition is realized by means of the event in which during the (1 + 1)-random mutation of there are no flipped bits. This event has non-zero probability and, therefore, can be realized. Suppose that , . It means that vector is formed by t zeroes and k–t ones. But in this case, the vector must contain t ones and, accordingly, n–k–t zeros. Then it is easy to see that the transition is also possible (and consequently has non-zero probability) thanks to the use of the operator: ; . Due to the fact that we can arbitrarily choose vectors from the proposition is proved. □
As we already noted above, it is important that the vector which is the random mutation of α is as close to α as possible Hamming distance-wise, since it enables the algorithm to ‘sense’ the landscape of a plot of a considered function. Note that for an arbitrary pair of vectors : we have . Now, let us establish the following property of the described algorithm.
Proposition 2. For : it holds that .
Proof: Assume that the application of to a part of vector α resulted in the flipping of r bits. But in this case in exactly r bits will also be flipped (chosen randomly). Then it holds that . Since the expected value of the number of flipped bits in w.r.t. mutation rate is 1 (see above), we have that . Thus the proposition is proved. □
Now let us construct an upper bound on the complexity of the described algorithm. We will use the reasoning similar to that from [49]. Remind that in [49] they proposed the idea to estimate the complexity of (1 + 1)-EA via the expected value of a random variable which is the number of random mutations performed by (1 + 1)-EA until it reaches a global extremum of the considered function. We will denote this expectation for specific as . Let P be the probability of success of the event . Then due to the properties of the geometric distribution [47] we have that the considered expectation is .
In [49] an apparatus was presented which makes it possible to construct upper bounds for values of the kind keeping in mind an arbitrary vector α. As it was showed in [49] the value P reaches its minimum for α such that and in this case we have P = pn where p is a mutation rate. If, for example, , then we have P = n−n and, consequently, the following upper bound on the complexity of (1 + 1)-EA holds: . As mentioned above, for some specific functions (see [49]) it was demonstrated that bounds of the kind can be achieved for some constant . However, in our case, we know nothing about the nature of the considered functions and, consequently, we need to work under the assumption that this function is a black box and construct for (1 + 1)-WEA only the upper bounds similar to ones for the classical variant of (1 + 1)-EA. The corresponding results are presented by the following theorem.
Theorem 2. Suppose, that (1 + 1)-WEA is applied to maximization of an arbitrary pseudo-Boolean function over , 0<k<n, and let . Then for some and for all the following upper bound holds:
Proof: Fix an order on the set of coordinates of all vectors from {0,1}n. Consider the operator . Let us construct vectors and and assume, that , . From the definition of it follows that . The vector undergoes the standard (1 + 1)-EA mutation with the rate p = 1/l. Let be the vector on which the considered (arbitrary) fitness function achieves a global extremum in . Denote by the vector formed by the coordinates in with the same numbers (w.r.t. the order fixed above) as in . We want to estimate the probability that after the application of to α the following transition happens: . By using the reasoning similar to that in [49], we have that:
| (10) |
Suppose, that during the transition there were r () mutated bits in , and the result of transition was . Thus, . Now, consider the corresponding correcting transformation of . Under the assumption that the transition is we can conclude, that and must differ in r coordinates, and these are the exact coordinates that must be flipped during the application of to . Thus, we need to correctly choose r coordinates out of n–l, using the sampling without replacement. It is a well-known fact (see e.g. [47]) that if the order of chosen coordinates matters, then the corresponding number of possible alternative choices is expressed by the following falling factorial: . But in our case, the order of chosen coordinates does not matter because n–l considered coordinates are either all zeroes or all ones. Thus, the number of possible alternatives of the choice is the following binomial coefficient: . Then, the probability of the transition looks as follows:
| (11) |
Note, that when r = 0 the vector α will not change and the probability of this event is . Hereinafter, we assume that . Taking into account (10) and (11), for a fixed we have, that:
Let us make use of the fact, that . Then, when we have . Therefore,
| (12) |
From (12), using the reasoning analogous to [49], we have an expression for the expected value of the number of independent mutations that need to happen for the transition to occur:
| (13) |
To construct a more explicit estimation for let us introduce the notation and consider the value:. Then, directly from (13) the following holds:
Taking into account, that , for all we have:
| (14) |
Note, that tends to e−1 with the increase of n. Then, keeping in mind that for all r > 6 it holds that we can rewrite (14) as follows:
| (15) |
Since , then for all . Also, taking into account (13) we can derive from (15) that:
| (16) |
Let us note, that the value in the right part of (16) is increasing with the increase of r. However, since it is clear that the validity of the theorem follows from (16). □
3 IM and TSS as pseudo-Boolean optimization problems
In this section we will give formal definitions of IM and TSS in form of maximization problems for pseudo-Boolean functions of the kind (8). To solve them we will employ the concept in which the solving itself is split in two stages:
The stage at which we construct an initial solution using a Greedy Algorithm (GrA);
The stage at which we improve the solution found at stage 1 using evolutionary optimization.
In the context of IM we consider DDS G, with weight functions of the kind (1) and fixed thresholds (DLTM). For each let us focus on the following function:
| (17) |
For an arbitrary we will suppose that , where is the influence computed w.r.t. Definition 2.
Using GrA let us find some which will serve as an initial approximation of , where is defined by (2). Next, we use as a starting current point and apply k-weight mutation operator to and the following current vectors, aiming to maximize (17) until some termination criterion is met. The resulting vector is the approximate solution of IM for network G over . Let us additionally note, that to find , the greedy algorithm uses the approximate solutions of IM over the sets of the kind , .
Next, consider the TSS problem. The general scheme for solving TSS employed further is as follows. On the first stage, similar to IM we construct an approximate solution using GrA. Thus, the vector is an approximate solution for IM for the smallest k such that the number of active vertices in the state reachable from is at least R.
Once we found in we make use of the two following strategies for solving TSS. The first strategy implies launching (1 + 1)-EA, (1 + 1)-FEA [51] or GA. Each algorithm is tasked with finding a vector of minimal Hamming weight such that starting from this vector, the DDS G transitions to a fixed point in which the number of active vertices is at least R.
The second strategy consists in the following steps:
Start from the current dimension k such that and use only operator on to find some approximate solution of IM denoted by ;
Transition from to some vector using the heuristics described below;
Starting from apply operator and try to find such that the number of active vertices in fixed point reached from is at least R. In the case of success, decrease the current dimension by 1 and go to step 2. Otherwise, output the current best known vector as an approximate solution of TSS.
To transit from to on step 2 we use several heuristics. Since for each value of k we will perform the same actions, let us without the loss of generality use the notation instead of . An arbitrary TS defined by vector we will denote as .
The first and the simplest heuristic, to which we further refer as to (1 + 1)-WEA_v1 consists in removing from the set the vertex with the smallest number of outgoing arcs.
In the second heuristic denoted as (1 + 1)-WEA_v2 for each vertex we compute the so-called activation potential:
| (18) |
Then we remove from the vertex with minimal value of .
We also employed a heuristic which is close in spirit to the greedy algorithm described in Sect 1.3. Consider and let be the fixed point reached by DDS G when it starts from the initial state . For each vertex denote by the vector which specifies the target set and let be a fixed point reached by DDS G from state . Next we find the following vertex
| (19) |
In the role of we pick the vector that specifies target set . However in practice this step is quite computationally intensive due to the necessity to check all vertices in . That is why we sort all vertices from in an ascending order of value (18), calculate the difference only for the first Q vertices from this ordered set, and choose a vertex with minimal value of this difference. We denote the described heuristic as (1 + 1)-WEA_v3. In the next section, we present the results that demonstrate the behavior of the described heuristic for different values of parameter Q. The pseudocode of the proposed strategy for solving TSS by iteratively decreasing the value of k using the heuristic variant v3 is presented at Algorithm 3.
Algorithm 3. The hybrid algorithm with v3 heuristic.
![]() |
4 Computational experiments
We performed several series of experiments to assess the algorithms proposed in the previous sections and evaluate their performance compared to competition. All data related to experiments is available online at https://github.com/Alex-Andrv/evo_network_public.
4.1 Benchmarks
In the experiments we considered two classes of networks. The first one is formed by random networks of small size generated using the well-known models: the Watts-Strogatz model [52] (WS-networks) and Barabasi-Albert model [53] (BA-networks). We used these networks to construct exact solutions for TSS and evaluate the quality of solutions found by metaheuristic algorithms by comparing them against the exact one. The second class of benchmarks was formed by the fragments of real-world networks taken from the SNAP [32] repository. We used them to compare the considered evolutionary and hybrid algorithms against each other when solving IM and TSS.
4.1.1 Network graphs.
In more detail, for the experiment with small random networks to generate the latter we employed the well-known NetworkX package [54]. For the Watts-Strogatz model we generated networks with 40 vertices, where each node is joined with k = 8 neighbors in a ring and each edge has the probability of to be rewired. In the case of Barabasi-Albert model we generated graphs with 50 vertices, and used the value of parameter m (the number of edges to attach from a new node) equal to 4, with no initial graph specified (in which case NetworkX starts from a star graph on nodes).
As for the benchmarks from the SNAP repository, we used a number of networks with several thousand vertices on average. Specifically, we picked facebook_combined (which is a whole network from the ego-Facebook SNAP entry), p2p-Gnutella06_combined, p2p-Gnutella08_combined, wiki-vote and ca-GrQc networks.
The majority of network graphs from SNAP as well as the generated small graphs are undirected, thus we replaced each edge by a pair of arks going in opposite directions.
4.1.2 Arc weights and thresholds.
To generate arc weights we chose the numbers (see Sect 1.1) uniformly from a set , where K was chosen depending on the network size. For the small networks the values of were picked from the range of for the Watts-Strogatz networks and from for the Barabasi-Albert networks. In the case of SNAP networks to reflect the possibility that the influence of some agents may be significantly larger than that of other we picked from .
The vertex thresholds were generated with respect to one of the two different thresholds distributions. In the first case we used equal thresholds for all vertices , while in the second case, the value of θ for each vertex was chosen randomly from .
Typically, we encapsulate parameter values in a benchmark name. For example, in benchmark BA_50_4_uni_1-5_const_0.8, BA refers to the Barabasi-Albert model, 50_4 to the network size and the parameter for the graph generator (in this case, the number of edges to attach from a new node to existing nodes). The following block refers to the way the arc weights are chosen: uni_1-5 means that the arc weights were picked in accordance with the uniform distribution from the range The final block const_0.8 refers to the scheme used to choose thresholds and in the example it employs constant thresholds equal to 0.8 for each vertex.
4.2 Comparing to exact solution
The objective of the first experiment was to demonstrate that evolutionary algorithms and the combination of the greedy and evolutionary algorithms given enough resources can produce solutions that are close in size to the exact solution. To show it we used small random networks described in Sect 4.1, so that we were able to find the exact solutions using combinatorial algorithms, in the role of which we employed modern SAT solvers.
Note that in this context it is natural to consider TSS in the following formulation: to find the TS of smallest size that activates at least network vertices in M time steps. To reduce this problem to SAT we used the methods similar to those described in [42] and the PySAT toolkit [55]. As the value of M we used the number of vertices in networks, thus in this part of the experiments M = n. If in the process of solving the algorithm finds a set of size then the formula is constructed that encodes the existence of a TS of size which activates at least R vertices in M time steps. If this formula is unsatisfiable for TS of size then the TS of size is considered to be the exact solution.
We applied to the same networks the combination of the greedy and the evolutionary algorithm, in particular, (1 + 1)-EA. The evolutionary algorithm started either from a set corresponding to all network vertices or from an initial point corresponding to the solution found by the greedy algorithm.
The results of the experiment are presented in Table 1. It is clear from the table that in most cases the hybrid strategy manages to find the solution which is close to optimal, and often does find the optimal solution.
Table 1. Results on the comparison of evolutionary and greedy algorithms with exact solution.
| Benchmark | SAT (exact) | GrA | (1 + 1)-EA | GrA&(1 + 1)-EA |
|---|---|---|---|---|
| BA_50_4_uni_1-5_uni_0.75-1 | 15 | 18 | 16.35 | 15.6 |
| BA_50_4_uni_1-5_const_0.8 | 16 | 19 | 17.5 | 17.1 |
| WS_40_8_0.5_uni_1-2_uni_0.75-1 | 15 | 20 | 17.65 | 17.1 |
| WS_40_8_0.5_uni_1-2_const_0.8 | 16 | 21 | 16.9 | 16.25 |
In all cases the value represents the TS size. Evolutionary algorithms were allowed to perform up to 10000 iterations. The results of nondeterministic algorithms are averaged over 20 launches and are presented in cursive.
4.3 Solving IM with (1 + 1)-WEA
The goal of the next series of experiments was to see whether the proposed mutation operator in (1 + 1)-WEA really manages to improve the quality of solution in the context of IM, e.g. when we fix the size of the target set and aim to maximize the number of activated networks.
For this purpose, we used the GrA algorithm to find TSS solutions for , where and n is the number of vertices in a network. Recall that similar to above we denote by k the size of the constructed TS. Next, we applied (1 + 1)-WEA (with the heuristics to decrease TS size disabled) to see how it performs in solving IM for specific values of k. For this purpose, we used the benchmarks based on the SNAP networks described in Sect 4.1.
In the experiments, we observed that in all cases (1 + 1)-WEA was able to improve the quality of the solution found by the greedy algorithm. The Fig 2 presents the convergence plot for one of the considered benchmarks in which the improvements were better than average. It is clear from the figure that (1 + 1)-WEA is sometimes capable of explosive improvements (see orange line on the plot) that result in significant increases in the quality of the found TS. In many observed cases, however, the behavior is more akin to that presented on red and blue lines: clear improvements, but without sudden large increases in influence.
Fig 2. Convergence plots for solving IM for p2p-Gnutella06_uni_1-1000_uni_0.75-1 via (1 + 1)-WEA.

The value of k corresponds to the size of the target set, the influence (vertical axis) refers to the number of activated vertices in a fixed point, while mutations (horizontal axis) means the number of mutations (iterations) made by the evolutionary algorithm.
To illustrate the improvements that can be obtained using (1 + 1)-WEA, we generated a small network with 50 vertices, and used greedy algorithm to find a target set of size k = 19. The set found by the greedy algorithm is presented in the left part of Fig 3, while that found by the hybrid approach – in the right part of the figure. It is clear that the quality of the set constructed by the proposed approach is significantly higher.
Fig 3. Comparison of the set found when solving IM for k = 19 by the greedy algorithm (left) and the one found by (1 + 1)-WEA for a network with n = 50 vertices.
Vertices from the target set are marked red, active vertices—by blue and inactive vertices are gray. It is clear that the set found using (1 + 1)-WEA manages to activate 11 more vertices out of 50 compared to the set found by the greedy algorithm.
4.4 Evaluation of (1 + 1)-WEA for solving TSS
In the final series of experiments we considered SNAP networks and solved TSS using the following algorithms to evaluate their performance: Greedy Algorithm (GrA), (1 + 1)-EA, (1 + 1)-FEA with , Genetic algorithm (GA), (1 + 1)-WEA augmented with heuristics for moving from to , which were described in Sect 3: (1 + 1)-WEA_v1, (1 + 1)-WEA_v2, (1 + 1)-WEA_v3. All evolutionary algorithms started from the solution found by the greedy algorithm, which stopped once it found a TS that activates at least vertices. Evolutionary algorithms were allowed to perform 10000 iterations: in the case of evolutionary algorithms it was 10000 mutations and in the case of Genetic Algorithm it was 10000 populations of size 10.
From Table 2 one can see that the Genetic Algorithm while having the most resources yields the worst results. We believe that the reason for this behavior is that the crossover and mutation operators in GA need to be specifically tuned to work well with IM/TSS in mind. The analysis shows that the standard crossover operator often produces TSs the Hamming weight of which is too different from that of current best known TS. Therefore, such individuals are typically useless, but they nevertheless consume a lot of allocated resources. From our point of view, adapting crossover for IM and TSS represents a promising direction of future research.
Table 2. Results on solving TSS using different evolutionary and greedy algorithms.
| GrA | EA | FEA | GA | WEA_v1 | WEA_v2 | WEA_v3 | |
|---|---|---|---|---|---|---|---|
| Wiki-Vote | |||||||
| uni_1-1000_const_0.8 | 3450 | 3224 | 3227 | 3381 | 3193 | 3180 | 3180 |
| uni_1-1000_uni_0.75-1 | 3380 | 3170 | 3174 | 3305 | 3142 | 3136 | 3135 |
| p2p-Gnutella06 | |||||||
| uni_1-1000_const_0.8 | 582 | 490 | 489 | 536 | 400 | 393 | 397 |
| uni_1-1000_uni_0.75-1 | 437 | 376 | 376 | 406 | 319 | 309 | 300 |
| p2p-Gnutella08 | |||||||
| uni_1-1000_const_0.8 | 332 | 268 | 271 | 301 | 233 | 216 | 219 |
| uni_1-1000_uni_0.75-1 | 281 | 227 | 228 | 252 | 185 | 176 | 176 |
| facebook_combined | |||||||
| uni_1-1000_const_0.8 | 1526 | 1428 | 1426 | 1486 | 1371 | 1411 | 1411 |
| uni_1-1000_uni_0.75-1 | 1371 | 1254 | 1255 | 1318 | 1211 | 1243 | 1243 |
| ca-GrQc | |||||||
| uni_1-1000_const_0.8 | 1592 | 1536 | 1535 | 1574 | 1496 | 1500 | 1499 |
| uni_1-1000_uni_0.75-1 | 1475 | 1413 | 1412 | 1451 | 1366 | 1376 | 1374 |
In all cases the value represents the TS size. Evolutionary algorithms were allowed to perform up to 10000 iterations and started from the solution found by the greedy algorithm in all cases, thus we omit this information in column names. We also omit the (1 + 1) part for brevity. The results of nondeterministic algorithms are averaged over 20 launches and rounded. The best result for each benchmark is outlined in bold. The values are rounded to the closest integer.
We computed a Wilcoxon Signed Rank Test to study the (1 + 1)-WEA performance improvement over other methods. It revealed a statistically significant difference in TS size found by different methods, in particular, the results of the proposed WEA algorithm (in any of three variations) significantly differ from others (p<0.05), with a large effect size (r = 0.886) [56].
Fig 4 presents convergence plots for a single run of the algorithms for four out of five networks. From both Fig 4 and Table 2 it is clear, that (1 + 1)-EA and (1 + 1)-FEA show comparable results, which are substantially better compared to that of GA. In almost all cases they make it possible to improve the solution found by the Greedy algorithm.
Fig 4. Convergence plots for solving TSS for SNAP networks.
Finally, three versions of (1 + 1)-WEA show the best results overall. An interesting fact is that in the considered test series typically different versions of (1 + 1)-WEA perform better than the others, and this behavior perseveres for different arcs’ weights and vertices’ thresholds. It is natural to assume that they tend to align well with graph structure. Thus, another interesting venue for future research consists in learning to pick the best heuristic based on the effectively computable network properties, such as centrality, etc.
4.5 Sensitivity analysis
In this subsection we present the results on the sensitivity analysis of the behavior of the (1 + 1)-WEA_v3 algorithm described in Sect 3 under different values of parameter Q. In this series of experiments we used the same networks as in the previous subsection.
Remind, that the v3 heuristic described in Sect 3 attempts to remove from the current target set of size k a single vertex, and then launches (1 + 1)-WEA over {0,1}k−1 to maximize the influence. If it manages to find then this becomes the current best known target set.
When attempting to transition from TS of size k to TS of size k–1 using (19), we want to compute the value for as few TS candidates as possible, thus we pic the first Q of them when sorted by the value of (18).
We tested the values , the corresponding results are presented in Table 3. It is clear from them that the value of Q has quite little impact on the quality of the obtained solutions in most cases.
Table 3. Results on solving TSS using (1 + 1)-WEA_v3 with different values of Q.
| Wiki-Vote | |||||
| uni_1-1000_const_0.8 | 3175 | 3180 | 3181 | 3181 | 3179 |
| uni_1-1000_uni_0.75-1 | 3138 | 3135 | 3131 | 3138 | 3139 |
| p2p-Gnutella06 | |||||
| uni_1-1000_const_0.8 | 399 | 397 | 396 | 396 | 400 |
| uni_1-1000_uni_0.75-1 | 309 | 300 | 315 | 303 | 310 |
| p2p-Gnutella08 | |||||
| uni_1-1000_const_0.8 | 223 | 219 | 218 | 217 | 221 |
| uni_1-1000_uni_0.75-1 | 179 | 176 | 175 | 177 | 178 |
| facebook_combined | |||||
| uni_1-1000_const_0.8 | 1407 | 1411 | 1411 | 1407 | 1415 |
| uni_1-1000_uni_0.75-1 | 1235 | 1243 | 1235 | 1246 | 1242 |
| ca-GrQc | |||||
| uni_1-1000_const_0.8 | 1497 | 1499 | 1500 | 1498 | 1497 |
| uni_1-1000_uni_0.75-1 | 1375 | 1374 | 1376 | 1377 | 1375 |
In all cases the value represents the TS size. The results are averaged over 20 launches and rounded. The best result for each benchmark and value of Q is outlined in bold.
We used Wilcoxon Signed Rank test to see whether any of the values of Q shows statistically significant performance improvement over the others, but on the considered benchmarks the amount of data was not sufficient enough to reject a null hypothesis about the equality of medians between every pair of values.
On Fig 5 we show how the (1 + 1)-WEA_v3 algorithm improves the target set during several iterations. Here, we used the same network as in Fig 3.
Fig 5. Illustration of how (1 + 1)-WEA_v3 improves the target set when solving TSS.
The goal is to activate all vertices in the network. The network with 50 vertices is the same as in Fig 3. The first target set (left) is found by the greedy algorithm. The next three target sets are found consecutively by our algorithm. Vertices from the target set are marked red, active vertices – by blue.
4.6 Analysis of practical efficiency
Let us say a few words about the practical complexity of the employed algorithms. It is clear that a single mutation is performed by (1 + 1)-EA and (1 + 1)-WEA in time O(n), where n is the size of the mutated vector. Theoretical upper bounds on the number of mutations until the considered algorithms reach a global extrema can be found in Sect 2. In practice, the typical approach is to limit the number of mutations by some constant. In our experiments, we used the limits from 1000 to 10000 mutations. Usually, within the allocated mutation budget, the algorithms manage to substantially improve the initial value of the target function found by the greedy algorithm.
As for the memory, the described algorithms do not have any specific requirements: it is clear that the total amount of memory used by the (1 + 1)-EA and (1 + 1)-WEA algorithms from start to the moment they find a global extrema does not exceed O(n). Indeed, they basically store in the memory the best point, corresponding to the Best Known Value of a considered function in that point, and the current point. Also note, that the functions we optimized in the present paper are relatively cheap, compared to e.g. functions from [57,58]. Therefore, for these functions we can see that the (1 + 1)-evolutionary strategy works significantly better compared to more complex genetic algorithms operating with populations of size greater than one. Thus, for the considered functions the more effective strategy is the one that processes more points per the unit of time.
5 Related work
As it was mentioned above, to the best of our knowledge, the paper [33] by Stewart Kauffman was the first work in which networks were used to define dynamical systems with discrete time and finite number of different states. Also, in that work it was proposed to associate Boolean functions with network vertices, defining functions via the truth tables. The information diffusion model considered in [33] is exactly the same as the DLTM in the variant described in our paper. It was used there to describe possible interactions in biological systems known as Gene Regulatory Networks.
Later, the term “Kauffman Networks” was used in a number of works, e.g. [59,60], that studied different critical phenomena in these networks, such as distributions of the length of cycles produced by these networks, statistical characteristics for some kinds of clusters (called modules in [60]), as well as the analysis of similarities with the percolation theory (in this point the authors of [59,60] refer to [61]). Papers [62,63] studied the cyclic structure of state transition graphs for DDS induced by Kauffman Networks. In particular, in [62] to find cycles of small length they use Binary Decision Diagramns (BDD) – the well-known structute actively used in symbolic verification [64].
The idea to define the weight function of network vertices using thresholds was apparently first proposed by Mark Granovetter in [26]. In that paper, a number of examples that demonstrate the correspondence between the threshold model of activation dynamics and the so-called conforming behavior which is often observed in real life collectives, was provided. We would like to mention several further works which developed the ideas from [26], in particular [65–67], etc.
The paper [4] for more than 20 years remains the central work in the context of maximizing influence in networks and continues to stimulate novel results in this area. It introduced all the main notions related to influence maximization. At the same time, the authors of [4] note that they were largely inspired by the problems considered in [68]. The main models under which IM is usually considered, in particular, Linear Threshold Model and Independent Cascade Model, were also proposed in [4]. The paper also showed that for the nondeterministic variants of these models there exists a polynomial algorithm that approximates the solution of IM with approximation factor 1–e−1. However, for the deterministic variants of both models it gave strong arguments on their inapproximability.
IM is one of the most actively researched problems in network science in the last 20 years: see [4–21]-[25,68],etc. Let us briefly comment only on the results of the most cited of the mentioned works. Thus, [6] is a direct development of [4]: in particular, they studied in detail the so-called decreasing cascade model, the basic properties of which were formulated in [4] in form of hypotheses. The authors of the paper [69] state that the approach they proposed based on the meticulous analysis of the submodularity property outperforms the basic greedy algorithm from [4] in hundreds of times on some networks. In [7] there have been studied the questions related to influence diffusion in social networks by aggregating the groups of users by the topics of their interests, the corresponding approach is called Topical Affinity Propagation. In [8] they proposed the strategy of Two-phase Influence Maximization which makes it possible for the non-deterministic Independent Cascade model to significantly speed up the standard greedy algorithm without falsifying the approximability guarantees of the obtained solutions. The paper [9] studies the problem of activating the network by means of top-K influential vertices. As the authors mention in text, their algorithm is aimed at solving the problem of activating mobile social networks with a large number of participants. Several works, such as [11,12], proposed improvements to the greedy algorithm and some additional centrality-based heuristics, which make it possible to increase the dimension of effectively activated networks to millions of vertices. In [15] the authors studied nontrivial questions related to the structural complexity of IM, in particular, they showed the -hardness of exact influence computation in the context of the Linear Threshold model. Some peculiarities of the structure of real-world information networks that pass through themselves colossal volumes of information (millions of blogs and news articles) were studied in [17]. A time-constrained variant of IM was apparently first considered in [19]. The Influence Maximization for a network which changes with time was analyzed in [20] In [21] it was showed that taking into account the information about location may play an important role when maximizing the influence in real-world marketing strategies. The robustness of IM solutions under uncertainty of some network parameters (specifically, probabilities assigned to edges) was studied in [24].
Surprisingly, the papers dealing with IM under DLTM are significantly fewer in number compared to that for LTM, despite the fact that the ideas how one can solve IM under DLTM using heuristic and metaheuristic algorithms look natural. The greedy algorithm for DLTM which we use to construct an initial solution of IM and TSS was described in [37]. Several subsequent works used some generalizations of this algorithm: [70,71], etc. The paper [70] considered the targeted and budgeted variant of IM, in which the profits and costs of activating vertices were added, and proposed the modification of the greedy algorithm. In [71] there was considered a time-constrained variant of IM under DLTM. The intention to demonstrate that it is possible to solve IM and TSS under DLTM by combining greedy heuristics and evolutionary optimization became our main motivation for writing the present paper.
The TSS problem was first implicitly formulated in [4]. The formulation that we use in our paper was employed in [72] and [35]. The structural complexity of IM and TSS was first evaluated in [4]. As we already mentioned, there are approximation algorithms with constant factor for IM under nondeterministic LTM (when influence is defined as an expectation (3)), meanwhile, under DLTM the solution of IM cannot be efficiently approximated if . It was noted in [35], that the intractability of approximation of TSS under DLTM follows from this fact. A number of further results related to the complexity of TSS was also presented in [35]. In particular, it was shown that some non-monotonic variants of TSS are -hard.
As we mentioned above, both IM and TSS under DLTM can be solved exactly using modern combinatorial algorithms. However, the dimension of the problems with which such algorithms are able to successfully deal with is quite small: typically in the dozens or at most hundreds of vertices. In [42,73] the algorithms for solving the Boolean satisfiability problem (SAT) were used to solve TSS. Note, that in [42,74–76] the algorithms based on SAT or on Binary Decision Diagrams (BDD) were also used to find attractors (cycles) of small length in State Transition Graphs of Kauffman networks and networks with more elaborate weight functions modeling dynamic processes in gene networks. It is also worth mentioning, that the standard network activation model considered in the majority of related works (including the present paper) implies the monotonic scenario, in which DDS has only attractors of length 1 (i.e. fixed points). This scenario corresponds to the so-called conforming behavior in terms of [26]. In paper [42] it was showed that replacing functions of the kind (1) by functions of the kind , corresponding to anti-conforming behavior, results in DDS with STG the structure of which is significantly different from the structure of STG for the conforming scenario.
We would like to specifically note the series of works in which the Discrete Dynamic Systems of the same kind as the ones considered in our paper were studied from the point of view of cellular automata. In particular, in [77] they showed that the general reachability problem for models close to DLTM is PSPACE complete. The works [78,79] studied a number of problems related to reachability and specifically showed the NP-completeness of the general Predecessor Existence problem in the context of DDS similar to Synchronous Boolean Networks.
The networks we use in the paper can be broadly divided into two main classes. First, these are network constructed using the random graph models, such as the Watts-Strogatz model [52] and Barabasi-Albert model [53]. Second, there are fragments of real-world networks available in the SNAP repository [32].
The evolutionary and genetic algorithms are among the most popular tools for metaheuristic optimization [45]. The (1 + 1)-EA algorithm can be viewed as the simplest implementation of the idea of evolutionary computations. Nevertheless, this algorithm has a lot of nontrivial mathematical properties studied in e.g. [49,51,80–83], etc. In particular, as we mentioned above, in [49] there was proposed a natural approach to the study of the behavior of (1 + 1)-EA on a wide class of pseudo-Boolean functions. For a number of artificially constructed functions in [49] it was shown that (1 + 1)-EA can be both either very effective (e.g. for ONEMAX function) or vice versa. For an arbitrary pseudo-Boolean function without any analytical specification (1 + 1)-EA is highly ineffective in the context of the bounds from [49]. Note, that similar results were presented in paper [84]. In [51] an approach was proposed in the context of which (1 + 1)-EA changes its mutation rate according to some probabilistic distribution. A similar idea was used earlier in [80]. The (1 + 1)-EA variant from [51] has a significantly better upper bound on its complexity compared to the original (1 + 1)-EA. In papers [83] and [85] variants of (1 + 1)-EA were described with the complexity upper bound significantly smaller than that of the classic algorithm.
The variant of (1 + 1)-EA which is based on the operator was first described in the conference paper [86]. In that paper this algorithm was used to find the so-called ρ-backdoors described in [87] which in essence represent a probabilistic generalization of backdoors for SAT introduced in [88]. In the present paper we performed the extended analysis of the properties of the algorithm from [86] and significantly improved the upper bound of its complexity for pseudo-Boolean functions without any analytical specification. One of the contributions of the present paper lies in adapting (1 + 1)-WEA for solving IM and TSS.
To the best of our knowledge, the conference paper [89] has the first examples of the application of evolutionary algorithms to solving IM under DLTM. Our two conference papers [90,91] present the preliminary results on solving TSS using different evolutionary algorithms, including that from [51,86] and [85].
6 Discussion, conclusion and future work
In the present paper we considered two well-known combinatorial problems for Boolean networks: Influence Maximization and Target Set selection. We studied these problems in the context of the Deterministic Linear Threshold Model, while the majority of papers on the topic consider nondeterministic models, first formally presented in [4]. This choice is motivated by a number of factors. First, DLTM agrees well with many practical examples, the nature of which was described by Mark Granovetter in [26]. Second, our goal was to employ for solving IM and TSS the metaheuristic algorithms used in pseudo-Boolean optimization and specifically evolutionary algorithms. It should be noted that the described computational schemes may be applied to non-deteministic LTM from [4], but in this case it makes sense to implement them in parallel environments, since computing the fitness function under LTM is much more expensive compared to DLTM. Indeed, for large networks the value (3) can only be estimated using the Monte Carlo method [92], and the accuracy of this estimation improves with the increase of the size of the random sample. From our point of view, the study of how the network size influences the accuracy of computing influence in the sense of (3) is an important problem which we plan to consider in the nearest future. In particular, we believe that the results presented in this paper look promising in that the corresponding algorithms for solving IM and TSS under LTM can be built upon the parallel versions of the algorithms proposed in this paper.
The use of neural networks for solving IM and TSS looks promising. This direction fits into a general area where neural networks are used to tackle NP-hard problems. It grew in popularity in recent years thanks to the ability of neural networks to perform the role of universal approximators. The examples of such an approach can be found in papers [93–97], etc. Some of the works, in particular [98–100], and some others, show the possible applications of neural networks for solving IM and TSS. However, the cited papers typically either consider an Independent Cascade Model or the Linear Threshold Model in their standard non-deterministic variants from [4]. Our attempts to apply the tools presented in these articles to DLTM did not yield good results, since we observed the effects mentioned in [94], e.g., various “bottlenecks”, in particular, the lack of robustness of the neural network models when fed with data only implicitly related to the problems they were designed to solve. As a result, to construct the target sets similar in their effectiveness to that found by our hybrid approach, one would need to employ a very resource-intensive training stage. Nevertheless, we believe that an approach to solving IM via Neural Networks is very promising and we plan to address it in our future works. In particular, we believe that Binarized Neural Networks [101] suit well to solving IM and TSS in the context of DLTM and plan to conduct corresponding research in the near future. In particular, we plan to study the possibilities in combining the proposed evolutionary algorithms with binarized neural networks, specifically to use the latter to improve the initial solutions to IM and TSS found by greedy heuristics.
Acknowledgments
We are grateful to anonymous reviewers for their constructive feedback that made it possible to significantly improve the presentation of our results. We thank Konstantin Chukharev and Mikhail Smirnov for fruitful discussions that helped a lot at the earlier stages of the paper preparation.
Data Availability
The data related to experiments is available in a public repository https://github.com/Alex-Andrv/evo_network_public.
Funding Statement
This research is financially supported by the Ministry of Economic Development of the Russian Federation (IGK 000000C313925P4C0002), agreement No 139-15-2025-010. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Newman MEJ. The Structure and Function of Complex Networks. SIAM Review. 2003;45(2):167–256. [Google Scholar]
- 2.Dorogovtsev SN, Goltsev AV, Mendes JFF. Critical phenomena in complex networks. Rev Mod Phys. 2008;80:1275–335. [Google Scholar]
- 3.Newman MEJ. Networks: An Introduction. Oxford University Press; 2010.
- 4.Kempe D, Kleinberg JM, Tardos É. Maximizing the spread of influence through a social network. In: KDD. ACM; 2003. p. 137–46.
- 5.Kempe D, Kleinberg JM, Tardos É. Influential Nodes in a Diffusion Model for Social Networks. In: ICALP. vol. 3580 of Lecture Notes in Computer Science. Springer; 2005. p. 1127–38.
- 6.Bharathi S, Kempe D, Salek M. Competitive Influence Maximization in Social Networks. In: WINE. vol. 4858 of Lecture Notes in Computer Science. Springer; 2007. p. 306–11.
- 7.Tang J, Sun J, Wang C, Yang Z. Social influence analysis in large-scale networks. In: KDD. ACM; 2009. p. 807–16.
- 8.Tang Y, Xiao X, Shi Y. Influence maximization: near-optimal time complexity meets practical efficiency. In: SIGMOD Conference. ACM; 2014. p. 75–86.
- 9.Wang Y, Cong G, Song G, Xie K. Community-based greedy algorithm for mining top-K influential nodes in mobile social networks. In: KDD. ACM; 2010. p. 1039–48.
- 10.Borodin A, Filmus Y, Oren J. Threshold Models for Competitive Influence in Social Networks. In: WINE. vol. 6484 of Lecture Notes in Computer Science. Springer; 2010. p. 539–50.
- 11.Chen W, Wang Y, Yang S. Efficient influence maximization in social networks. In: KDD. ACM; 2009. p. 199–208.
- 12.Chen W, Wang C, Wang Y. Scalable influence maximization for prevalent viral marketing in large-scale social networks. In: KDD. ACM; 2010. p. 1029–38.
- 13.Das A, Kempe D. Submodular meets Spectral: Greedy Algorithms for Subset Selection, Sparse Approximation and Dictionary Selection. In: ICML. Omnipress; 2011. p. 1057–64.
- 14.Jiang Q, Song G, Cong G, Wang Y, Si W, Xie K. Simulated Annealing Based Influence Maximization in Social Networks. In: AAAI. AAAI Press; 2011, p. 127–32.
- 15.Chen W, Yuan Y, Zhang L. Scalable Influence Maximization in Social Networks under the Linear Threshold Model. In: ICDM. IEEE Computer Society; 2010. p. 88–97.
- 16.Sun J, Tang J. A Survey of Models and Algorithms for Social Influence Analysis. In: Social Network Data Analytics. Springer; 2011. p. 177–214.
- 17.Gomez-Rodriguez M, Leskovec J, Krause A. Inferring Networks of Diffusion and Influence. ACM Trans Knowl Discov Data. 2012;5(4):1–37. doi: 10.1145/2086737.2086741 [DOI] [Google Scholar]
- 18.He X, Song G, Chen W, Jiang Q. Influence Blocking Maximization in Social Networks under the Competitive Linear Threshold Model. In: SDM. SIAM/ Omnipress; 2012. p. 463–74.
- 19.Liu B, Cong G, Xu D, Zeng Y. Time Constrained Influence Maximization in Social Networks. In: ICDM. IEEE Computer Society; 2012. p. 439–48.
- 20.Zhuang H, Sun Y, Tang J, Zhang J, Sun X. Influence Maximization in Dynamic Social Networks. In: ICDM. IEEE Computer Society; 2013. p. 1313–8.
- 21.Li G, Chen S, Feng J, Tan K, Li W. Efficient location-aware influence maximization. In: SIGMOD Conference. ACM; 2014. p. 87–98.
- 22.Ohsaka N, Akiba T, Yoshida Y, Kawarabayashi K. Fast and Accurate Influence Maximization on Large Networks with Pruned Monte-Carlo Simulations. In: AAAI. AAAI Press; 2014. p. 138–44.
- 23.Huiyuan Zhang MTT, Subhankar Mishra. Recent Advances in Information Diffusion and Influence Maximization of Complex Social Networks. In: Opportunistic Mobile Social Networks. CRC Press; 2014. p. 37–68.
- 24.Chen W, Lin T, Tan Z, Zhao M, Zhou X. Robust Influence Maximization. In: KDD. ACM; 2016. p. 795–804.
- 25.Li Y, Fan J, Wang Y, Tan K-L. Influence Maximization on Social Graphs: A Survey. IEEE Trans Knowl Data Eng. 2018;30(10):1852–72. doi: 10.1109/tkde.2018.2807843 [DOI] [Google Scholar]
- 26.Granovetter M. Threshold Models of Collective Behavior. Am J Soc. 1978;83(6):1420–43. doi: 10.1086/226707 [DOI] [Google Scholar]
- 27.Akiba T, Sano S, Yanase T, Ohta T, Koyama M. Optuna: A Next-generation Hyperparameter Optimization Framework. In: KDD. ACM; 2019. p. 2623–31.
- 28.Real E, Liang C, So DR, Le QV. AutoML-Zero: Evolving Machine Learning Algorithms From Scratch. In: ICML. vol. 119 of Proceedings of Machine Learning Research. PMLR; 2020. p. 8007–19.
- 29.He X, Zhao K, Chu X. AutoML: A survey of the state-of-the-art. Knowledge-Based Systems. 2021;212:106622. doi: 10.1016/j.knosys.2020.106622 [DOI] [Google Scholar]
- 30.Narodytska N, Kasiviswanathan SP, Ryzhyk L, Sagiv M, Walsh T. Verifying Properties of Binarized Deep Neural Networks. In: AAAI. AAAI Press; 2018. p. 6615–24.
- 31.Ignatiev A, Narodytska N, Marques-Silva J. Abduction-Based Explanations for Machine Learning Models. In: AAAI. AAAI Press; 2019. p. 1511–9.
- 32.Leskovec J, Krevl A. SNAP Datasets: Stanford Large Network Dataset Collection. 2014. http://snap.stanford.edu/data
- 33.Kauffman SA. Metabolic stability and epigenesis in randomly constructed genetic nets. J Theor Biol. 1969;22(3):437–67. doi: 10.1016/0022-5193(69)90015-0 [DOI] [PubMed] [Google Scholar]
- 34.MacWilliams FJ, Sloane NJA. The theory of error-correcting codes. 2nd edition. North-Holland Publishing Company; 1978.
- 35.Ben-Zwi O, Hermelin D, Lokshtanov D, Newman I. An exact almost optimal algorithm for target set selection in social networks. In: EC. ACM; 2009. p. 355–62.
- 36.Nemhauser GL, Wolsey LA, Fisher ML. An analysis of approximations for maximizing submodular set functions–I. Math Program. 1978;14(1):265–94. [Google Scholar]
- 37.Swaminathan A. An Algorithm for Influence Maximization and Target Set Selection for the Deterministic Linear Threshold Model [Master’s thesis]. USA: Virginia Polytechnic Institute and State University; 2014.
- 38.Biere A, Heule M, van Maaren H, Walsh T, editors. Handbook of Satisfiability - Second Edition. vol. 336 of Frontiers in Artificial Intelligence and Applications. IOS Press; 2021.
- 39.Cook SA. The Complexity of Theorem-Proving Procedures. In: STOC. ACM; 1971. p. 151–8.
- 40.Levin LA. Universal Sequential Search Problems. Problems Inform Transmission. 1973;9:265–6. [Google Scholar]
- 41.Garey MR, Johnson DS. Computers and Intractability: A Guide to the Theory of NP-Completeness. New York (NY): W. H. Freeman & Co.; 1979.
- 42.Kochemazov S, Semenov A. Using synchronous Boolean networks to model several phenomena of collective behavior. PLoS One. 2014;9(12):e115156. doi: 10.1371/journal.pone.0115156 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Boros E, Hammer PL. Pseudo-Boolean optimization. Discrete Applied Mathematics. 2002;123(1–3):155–225. doi: 10.1016/s0166-218x(01)00341-9 [DOI] [Google Scholar]
- 44.Burke EK, Kendall G, editor. Search Methodologies. 2nd edition. Springer. 2016.
- 45.Luke S. Essentials of metaheuristics. 2nd edition. Lulu. 2013.
- 46.Mühlenbein H. How Genetic Algorithms Really Work: Mutation and Hillclimbing. In: PPSN; 1992. p. 15–26.
- 47.Feller W. An introduction to probability theory and its applications. 3rd edition. John Wiley & Sons, Inc.; 1968.
- 48.Russell SJ, Norvig P. Artificial Intelligence – A Modern Approach. 3rd edition. Pearson Education; 2010.
- 49.Droste S, Jansen T, Wegener I. On the Analysis of the (1 + 1) Evolutionary Algorithm. Theor Comput Sci. 2002;276(1–2):51–81. [DOI] [PubMed] [Google Scholar]
- 50.Wegener I. Theoretical Aspects of Evolutionary Algorithms. In: ICALP; 2001. p. 64–78.
- 51.Doerr B, Le HP, Makhmara R, Nguyen TD. Fast Genetic Algorithms. In: GECCO; 2017. p. 777–84.
- 52.Watts DJ, Strogatz SH. Collective dynamics of “small-world” networks. Nature. 1998;393(6684):440–2. doi: 10.1038/30918 [DOI] [PubMed] [Google Scholar]
- 53.Barabasi A, Albert R. Emergence of scaling in random networks. Science. 1999;286(5439):509–12. doi: 10.1126/science.286.5439.509 [DOI] [PubMed] [Google Scholar]
- 54.Hagberg AA, Schult DA, Swart PJ. Exploring Network Structure, Dynamics, and Function using NetworkX. In: Varoquaux G, Vaught T, Millman J, editors. Proceedings of the 7th Python in Science Conference. Pasadena (CA); 2008, p. 11–15.
- 55.Ignatiev A, Morgado A, Marques-Silva J. PySAT: A Python Toolkit for Prototyping with SAT Oracles. In: SAT. LNCS; 2018. p. 428–37.
- 56.Wilcoxon F. Individual Comparisons by Ranking Methods. Biometrics Bulletin. 1945;1(6):80. doi: 10.2307/3001968 [DOI] [Google Scholar]
- 57.Semenov A, Chivilikhin D, Pavlenko A, Otpuschennikov I, Ulyantsev V, Ignatiev A. Evaluating the Hardness of SAT Instances Using Evolutionary Optimization Algorithms. In: CP; 2021. p. 47:1–18.
- 58.Semenov A, Zaikin O, Kochemazov S. Finding Effective SAT Partitionings Via Black-Box Optimization. In: Black Box Optimization, Machine Learning, and No-Free Lunch Theorems. vol. 170 of Springer Optimization and Its Applications. Springer; 2021. p. 319–55.
- 59.Bastolla U, Parisi G. A numerical study of the critical line of Kauffman networks. J Theor Biol. 1997;187(1):117–33. doi: 10.1006/jtbi.1997.0423 [DOI] [PubMed] [Google Scholar]
- 60.Bastolla U, Parisi G. The modular structure of Kauffman networks. Physica D: Nonlinear Phenomena. 1998;115(3–4):219–33. doi: 10.1016/s0167-2789(97)00242-x [DOI] [Google Scholar]
- 61.Derrida B, Stauffer D. Phase Transitions in Two-Dimensional Kauffman Cellular Automata. Europhys Lett. 1986;2(10):739–45. doi: 10.1209/0295-5075/2/10/001 [DOI] [Google Scholar]
- 62.Dubrova E, Teslenko M, Martinelli A. Kauffman networks: analysis and applications. In: ICCAD. IEEE Computer Society; 2005. p. 479–84.
- 63.Hawick KA, James HA, Scogings C. Structural Circuits and Attractors in Kauffman Networks. In: ACAL. vol. 4828 of Lecture Notes in Computer Science. Springer; 2007. p. 189–200.
- 64.Bryant. Graph-Based Algorithms for Boolean Function Manipulation. IEEE Trans Comput. 1986;C–35(8):677–91. doi: 10.1109/tc.1986.1676819 [DOI] [Google Scholar]
- 65.Granovetter M, Soong R. Threshold models of diffusion and collective behavior. J Math Soc. 1983;9(3):165–79. doi: 10.1080/0022250x.1983.9989941 [DOI] [Google Scholar]
- 66.Granovetter M, Soong R. Threshold models of interpersonal effects in consumer demand. J Eco Beh Organ. 1986;7(1):83–99. doi: 10.1016/0167-2681(86)90023-5 [DOI] [Google Scholar]
- 67.Chwe MS. Structure and Strategy in Collective Action. Am J Soc. 1999;105(1):128–56. doi: 10.1086/210269 [DOI] [Google Scholar]
- 68.Domingos PM, Richardson M. Mining the network value of customers. In: KDD. ACM; 2001. p. 57–66.
- 69.Leskovec J, Krause A, Guestrin C, Faloutsos C, VanBriesen JM, Glance NS. Cost-effective outbreak detection in networks. In: KDD. ACM; 2007. p. 420–9.
- 70.Gursoy F, Gunnec D. Influence maximization in social networks under Deterministic Linear Threshold Model. Knowledge-Based Systems. 2018;161:111–23. doi: 10.1016/j.knosys.2018.07.040 [DOI] [Google Scholar]
- 71.Kochemazov S, Semenov AA. Computational Study of Time Constrained Influence Maximization Problem under Deterministic Linear Threshold Model for Networks with Nonuniform Thresholds. In: MIPRO. IEEE; 2019. p. 1121–5.
- 72.Chen N. On the approximability of influence in social networks. In: SODA. SIAM; 2008. p. 1029–37.
- 73.Kochemazov S, Zaikin O, Semenov AA. Improving the effectiveness of SAT approach in application to analysis of several discrete models of collective behavior. In: MIPRO. IEEE; 2017. p. 1172–7.
- 74.Dubrova E, Teslenko M. A SAT-based algorithm for finding attractors in synchronous Boolean networks. IEEE/ACM Trans Comput Biol Bioinform. 2011;8(5):1393–9. doi: 10.1109/TCBB.2010.20 [DOI] [PubMed] [Google Scholar]
- 75.Evdokimov A, Kochemazov S, Semenov A. Application of symbolic computations to the study of discrete models of some gene networks. VichislTechnol. 2011;16(1):30–47. [Google Scholar]
- 76.Evdokimov AA, Kochemazov SE, Otpushchennikov IV, Semenov AA. Study of discrete automaton models of gene networks of nonregular structure using symbolic calculations. J Appl Ind Math. 2014;8(3):307–16. doi: 10.1134/s1990478914030028 [DOI] [Google Scholar]
- 77.Barrett CL, Hunt HBI, Marathe MV, Ravi SS, Rosenkrantz DJ, Stearns RE. Complexity of reachability problems for finite discrete dynamical systems. J Comput Syst Sci. 2006;72(8):1317–45. [Google Scholar]
- 78.Barrett CL, Hunt HBI, Marathe MV, Ravi SS, Rosenkrantz DJ, Stearns RE. Predecessor existence problems for finite discrete dynamical systems. Theor Comput Sci. 2007;386(1–2):3–37. [Google Scholar]
- 79.Marathe MV, Ravi SS, Rosenkrantz DJ, Stearns RE. Computational Aspects of Fault Location and Resilience Problems for Interdependent Infrastructure Networks. In: COMPLEX NETWORKS (1). vol. 812 of Studies in Computational Intelligence. Springer; 2018. p. 879–90.
- 80.Witt C. Runtime analysis of the (mu + 1) EA on simple Pseudo-Boolean functions. Evol Comput. 2006;14(1):65–86. doi: 10.1162/evco.2006.14.1.65 [DOI] [PubMed] [Google Scholar]
- 81.Doerr B, Johannsen D, Winzen C. Multiplicative drift analysis. In: GECCO. ACM; 2010. p. 1449–56.
- 82.Doerr B. Analyzing randomized search heuristics via stochastic domination. Theor Comput Sci. 2019;773:115–37. [Google Scholar]
- 83.Semenov AA. Merging Variables: One Technique of Search in Pseudo-Boolean Optimization. In: MOTOR (2). vol. 1090 of Communications in Computer and Information Science. Springer; 2019. p. 86–102.
- 84.He J, Yao X. Drift analysis and average time complexity of evolutionary algorithms. Artif Intell. 2001;127(1):57–85. [Google Scholar]
- 85.Semenov AA, Otpuschennikov IV, Antonov K. On Some Variants of the Merging Variables Based (1 + 1)-Evolutionary Algorithm with Application to MaxSAT Problem. In: MOTOR. vol. 12755 of LNCS; 2021. p. 111–24.
- 86.Pavlenko A, Semenov AA. Using Island Model in Asynchronous Evolutionary Strategy to Search for Backdoors for SAT. In: CEC. IEEE; 2024. p. 1–7.
- 87.Semenov AA, Pavlenko A, Chivilikhin D, Kochemazov S. On Probabilistic Generalization of Backdoors in Boolean Satisfiability. In: AAAI. AAAI Press; 2022. p. 10353–61.
- 88.Williams R, Gomes CP, Selman B. Backdoors To Typical Case Complexity. In: IJCAI. Morgan Kaufmann; 2003. p. 1173–8.
- 89.Kochemazov S. Comparative Study of Combinatorial Algorithms for Solving the Influence Maximization Problem in Networks under a Deterministic Linear Threshold Model. Procedia Computer Science. 2018;136:190–9. doi: 10.1016/j.procs.2018.08.252 [DOI] [Google Scholar]
- 90.Smirnov M, Kochemazov S, Semenov AA. The Study of the Target Set Selection Problem under Deterministic Linear Threshold Model Using Evolutionary Algorithms. In: MIPRO. IEEE; 2023. p. 1039–44.
- 91.Andreev A, Chukharev K, Kochemazov S, Semenov AA. Solving Influence Maximization Problem Under Deterministic Linear Threshold Model Using Metaheuristic optimization. In: MIPRO. IEEE; 2024. p. 19–24.
- 92.Metropolis N, Ulam S. The Monte Carlo Method. J Amer statistical assoc. 1949;44(247):335–41. [DOI] [PubMed] [Google Scholar]
- 93.Ding J, Zhang C, Shen L, Li S, Wang B, Xu Y, et al. Accelerating Primal Solution Findings for Mixed Integer Programs Based on Solution Prediction. In: AAAI. AAAI Press; 2020. p. 1452–9.
- 94.Veličković P, Blundell C. Neural algorithmic reasoning. Patterns (N Y). 2021;2(7):100273. doi: 10.1016/j.patter.2021.100273 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Ferber AM, Song J, Dilkina B, Yue Y. Learning Pseudo-Backdoors for Mixed Integer Programs. In: CPAIOR. vol. 1329 2 of LNCS. Springer; 2022. p. 91–102.
- 96.Cappart Q, Chételat D, Khalil EB, Lodi A, Morris C, Velickovic P. Combinatorial Optimization and Reasoning with Graph Neural Networks. J Mach Learn Res. 2023;24:130:1–61. [Google Scholar]
- 97.Cai J, Huang T, Dilkina B. Learning Backdoors for Mixed Integer Linear Programs with Contrastive Learning. In: ECAI. vol. 392 of Frontiers in Artificial Intelligence and Applications. IOS Press; 2024. p. 2418–25.
- 98.Manchanda S, Mittal A, Dhawan A, Medya S, Ranu S, Singh AK. GCOMB: Learning Budget-constrained Combinatorial Algorithms over Billion-sized Graphs. In: NeurIPS; 2020.
- 99.Meirom EA, Maron H, Mannor S, Chechik G. Controlling Graph Dynamics with Reinforcement Learning and Graph Neural Networks. In: ICML. vol. 139 of Proceedings of Machine Learning Research. PMLR; 2021. p. 7565–77.
- 100.Ling C, Jiang J, Wang J, Thai MT, Xue R, Song J, et al. Deep Graph Representation Learning and Optimization for Influence Maximization. In: ICML. vol. 202 of Proceedings of Machine Learning Research. PMLR; 2023. p. 21350–61.
- 101.Hubara I, Courbariaux M, Soudry D, El-Yaniv R, Bengio Y. Binarized Neural Networks. In: NIPS; 2016. p. 4107–15.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data related to experiments is available in a public repository https://github.com/Alex-Andrv/evo_network_public.






