Wavelength rerouting in optical networks, or the Venetian Routing problem

Wavelength rerouting in optical networks, or the Venetian Routing problem

Journal of Algorithms 45 (2002) 93–125 www.academicpress.com Wavelength rerouting in optical networks, or the Venetian Routing problem ✩ Alberto Capr...

308KB Sizes 0 Downloads 40 Views

Journal of Algorithms 45 (2002) 93–125 www.academicpress.com

Wavelength rerouting in optical networks, or the Venetian Routing problem ✩ Alberto Caprara,a,1 Giuseppe F. Italiano,b,2 G. Mohan,c Alessandro Panconesi,d,∗,3 and Aravind Srinivasan e,4 a DEIS, Università di Bologna, Viale Risorgimento 2, 40136 Bologna, Italy b Dipartimento di Informatica, Sistemi e Produzione, Università di Roma Tor Vergata, Roma, Italy c Dept. of Electrical and Computer Engineering, National University of Singapore, Singapore d DSI, Università La Sapienza, via Salaria 113, piano terzo, 00198, Roma, Italy e Bell Laboratories, Lucent Technologies, 600-700 Mountain Avenue,

Murray Hill, NJ 07974-0636, USA Received 19 March 2001

Abstract Wavelength rerouting has been suggested as a viable and cost-effective method to improve the blocking performance of wavelength-routed wavelength-division multiplexing (WDM) networks. This method leads to the following combinatorial optimization problem,

✩ A preliminary version of this work appears in the Proc. Third International Workshop on Approximation Algorithms for Combinatorial Optimization Problems, 2000. * Corresponding author. E-mail addresses: [email protected] (A. Caprara), [email protected] (G.F. Italiano), [email protected] (G. Mohan), [email protected] (A. Panconesi), [email protected] (A. Srinivasan). 1 Work partially supported by CNR and MURST, Italy. 2 Work partially supported by the IST Programme of the EU under contract n. IST-1999-14.186 (ALCOM-FT), by the Italian Ministry of University and Scientific Research (Project “Algorithms for Large Data Sets: Science and Engineering”), and by CNR, the Italian National Research Council, under contract n. 00.00346.CT26. 3 This work was done when this author was at Università Ca’ Foscari di Venezia and Università di Bologna. 4 Part of this work was done while this author was at the School of Computing of the National University of Singapore, Singapore 119260, and while visiting the Università Ca’ Foscari di Venezia, Italy.

0196-6774/02/$ – see front matter  2002 Elsevier Science (USA). All rights reserved. PII: S 0 1 9 6 - 6 7 7 4 ( 0 2 ) 0 0 2 1 4 - 6

94

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

dubbed Venetian Routing. Given a directed multigraph G along with two vertices s and t and a collection of pairwise arc-disjoint paths, we wish to find an st-path which arcintersects the smallest possible number of the given paths. In this paper we prove the computational hardness of this problem even in various special cases, and present several approximation algorithms for its solution. In particular we show a non-trivial connection between Venetian Routing and Label Cover.  2002 Elsevier Science (USA). All rights reserved. Keywords: Optical networks; Wavelength rerouting; Wavelength-division multiplexing; Shortest paths; Approximation algorithms; Label Cover

1. Introduction Recent excavations in the once-vibrant merchant city of Venice have unearthed documents which show that as early as the Renaissance, complicated optimization problems were studied for state planning purposes. One of the most interesting problems to come to light is the following. As is well known, Venice is built on an archipelago of several hundred islands in the midst of a lagoon. Even today (miraculously) there are no roads and one has to either walk or go by boat. While today it is possible to walk between any two points inside the city—in modern terminology, the underlying graph is connected—in spite of the many wonderful little bridges, this was not the case back then. The citizenry could make use of a rather sophisticated and pervasive mass transit system, which was a huge source of revenues for the City of Venice, consisting of the characteristic Venetian boats— the gondolas—shuffling back and forth along predefined routes. The Municipality adopted a peculiar ticketing scheme. Tickets would be expensive but each ticket could be used repeatedly, as many times as one wanted in a single day, for the same route. But, to take a gondola covering a different route would require a new ticket. From the point of view of a citizen going from point A to point B the problem was that of minimizing the number of routes. The historical record shows that the municipality was concerned about how to make such reckoning virtually impossible, and that at some point it switched to a system in which routes were unidirectional. Here we show why this was advantageous in order to increase state revenues and how this old problem can shed light on more recent ones. Old wine, new bottles. We will be concerned with an optimization problem called V ENETIAN ROUTING (VR), and with some of its relatives. The input to VR consists of a directed multigraph G, two vertices s and t of G, respectively called the source and the sink, and a collection of pairwise arc-disjoint paths in G. These paths will be referred to as special or laid-out paths. The goal is to find an st-path which arc-intersects the smallest possible number of laid-out paths. (Recall that G is a multigraph. If e1 and e2 are parallel arcs, it is allowed for one laid-out path Q1 to contain e1 , and for another laid-out path Q2 to contain e2 .

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

95

But Q1 and Q2 cannot both contain e1 .) Note that the solution to the problem can be assumed to be a simple st-path without loss of generality. Besides state-planning in medieval Venice, this problem arises from wavelength rerouting in Wavelength-Division Multiplexing (WDM) networks. In WDM networks, lightpaths are established between pairs of vertices by allocating the same wavelength throughout the path [4]: Two lightpaths can use the same fiber link if they use a different wavelength. When a connection request arrives, a proper lightpath (i.e., a route and a wavelength) must be chosen. The requirement that the same wavelength must be used in all the links along the selected route is known as the wavelength continuity constraint. This constraint often leads to poor blocking performance: A request may be rejected even when a route is available, if the same wavelength is not available all along the route. To improve the blocking performance, one proposed mechanism is wavelength rerouting: Whenever a new connection arrives, wavelength rerouting may move a few existing lightpaths to different wavelengths in order to make room for the new connection. Lee and Li [15] proposed a rerouting scheme called “parallel move-to-vacant wavelength rerouting (MTV-WR).” Given a connection request from s to t, rather than just blocking the request if there is currently no available lightpath from s to t, this scheme does the following. Let A(W ) denote the set of current lightpaths that use wavelength W and that may be migrated to some other wavelength. That is, for each Q ∈ A(W ), there is some W  = W such that no lightpath with wavelength W  uses any arc of Q. Separately for each W , we try to assign the new connection request to wavelength W . To this end, we solve VR on the sub-network obtained by deleting all wavelength-W lightpaths not lying in A(W ); the set of laid-out paths is taken to be A(W ). (Since all lightpaths with the same wavelength must be arc-disjoint, the “arc-disjointness” condition of VR is satisfied.) If there is no feasible solution to any of these VR instances, the connection request is rejected. Otherwise, let P be an st-path representing the best solution obtained over all W . By migrating each lightpath arc-intersected by P to an available wavelength, we obtain an st-path with minimal disruption of existing routes. In the case of directed multi-ring network topologies, such an approach achieves a reduction of about 20% in the blocking probability [16]. Thus, VR naturally arises in such rerouting schemes, helping in dynamic settings where future requests cannot be predicted and where one wishes to accommodate new requests reasonably fast without significantly disrupting the established lightpaths. VR is also naturally viewed as finding a minimum number of laid-out paths needed to connect s to t. Since these paths have to be disrupted in order to honor the incoming request, this measure is very significant. Lee and Li gave a polynomial-time algorithm for VR on undirected networks [15]. This algorithm was later improved in [17]. These are exact algorithms. While the complexity of the undirected case is settled, to the best of our knowledge, no result (including the possible hardness of exactly solving VR) was known for general directed networks.

96

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

A natural generalization of VR is to associate a priority with each laid-out path, where we wish to find an st-path P for which the sum of the priorities of the laid-out paths which arc-intersect P is minimized; we call this Priority VR or PVR. An important special case of PVR is Weighted VR or WVR, where the priority of a laid-out path is its length. This is useful in our wavelength rerouting context, since the effort in migrating a lightpath to a new wavelength, is roughly proportional to the length of the path. Finally, one could look at VR from the complementary angle: Finding a maximum number of laid-out paths that can be removed without disconnecting t from s. Naturally, this variant will be referred to as the G ONDOLIER AVOIDANCE problem (GA). 1.1. Our results In this paper we study the worst-case complexity of VR and of its variants, and present several hardness results and approximation algorithms. In particular, we show an interesting connection between VR and the S YMMETRIC L ABEL C OVER (SLC) problem [1,6,13] which allows us to obtain new upper bounds for the latter. In what follows, let denote the maximum length (number of arcs) of any laidout path, and n, m be the number of vertices and arcs in G, respectively. Let QP  c stand for quasi-polynomial time, i.e., c>0 D TIME[2(log n) ]. For lower bounds, we establish the following: (1.1) VR is APX-hard even when = 3 (actual lower bound is 183/176). If  2, the problem is polynomially solvable exactly. (1.2) When is unbounded things get worse, for we exhibit an approximation preserving (L-) reduction from SLC to VR which implies that any lower bound for SLC also holds for VR. In particular, using the hardness of approximation result for SLC mentioned in [6,13], this shows that 1/2−ε m ), for any fixed ε > 0, is impossible approximating VR within O(2log unless N P ⊆ QP. (1.3) We exhibit an approximation preserving (L-) reduction from S ET C OVER to VR with unbounded . This implies two things. First, if N P = P, a weaker hypothesis than N P ⊆ QP, there exists a constant c > 0 such that VR cannot be approximated within c log n if is unbounded. Second, GA is at least as hard to approximate as I NDEPENDENT S ET. We also give several approximation algorithms. All algorithms reduce the problem to shortest path computations by using suitable metrics and (randomized) preprocessing. We remark that almost all of our algorithms actually work for a more general problem than VR defined as follows. The input consists of a multigraph G, a source s, and a sink t, and a collection of disjoint sets of arcs S1 , . . . , Sk , called laid-out sets. The goal is to find an st-path which arcintersects the smallest possible number of laid-out sets Si . In other words, we have

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

97

removed the constraint that the laid-out sets be paths. We refer to this problem as G ENERALIZED VR or GVR for short. Clearly, all previously described lowerbounds apply directly to GVR. Let SSSP(n, m) denote the time complexity of single-source single-sink shortest paths on an n-node, m-arc directed multigraph with non-negative lengths on the arcs. Given an optimization problem, let opt denote the optimal solution value for a given instance of the problem. We now summarize our results concerning upper bounds. (2.1) For VR, we give a linear-time algorithm for the case  2 and an /2 approximation algorithm for the general case. The same approach yields: (i) an /2 -approximation algorithm for PVR running in O(SSSP(n, m)) time, and (ii) an -approximation algorithm for GVR running in O(m) time. √ (2.2) We give an O( m/opt)-approximation algorithm for GVR. In view √ of the reduction described in point (1.2) above, this will yield an O( q/opt)approximation for SLC, where q and opt denote the input size and optimal solution value respectively of an SLC instance. This is the first non-trivial approximation for SLC, to our knowledge. √ ε (2.3) We show that for any fixed ε > 0, there is a 2O(m log m) -time O( m1−ε )approximation algorithm for VR. In particular, this shows a “separation” of VR (and hence SLC) from C HROMATIC N UMBER, as the same approximation cannot be achieved for the latter under the hypothesis that o(1) SAT cannot be solved in time 2n . We also show that, if N P = P, there N UMBER to VR. can be no L-reduction from C HROMATIC  algo(2.4) Improving on (2.2), we give an O( m/(opt · nc/opt ))-approximation √ rithm for GVR, for any constant c > 0; this also yields an O( m/ log n)approximation. The algorithm, which is randomized, is based on a technique dubbed “sparsification” which might be useful in other contexts. The algorithm can be derandomized. √ (2.5) We give an O(min{opt, m})-approximation for WVR running in O(m log m) time. Moreover, we show that all approximation results for VR can be extended to PVR if one replaces m and n in the approximation ratio by mw, where w is the maximum weight of a laid out path, which can be assumed to be polynomial in m. Finally, we perform an experimental analysis of our algorithms. In the analysis we are not concerned with their quality of approximation in the worst case. Rather we measure them with respect to what in practical situations is a more relevant performance parameter, namely the blocking probability (the average number of requests that can be honored instead of being rejected). Our tests show that our algorithms can lead to an improved performance at the expense of a moderate

98

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

computational effort, while at the same time the solutions they found are of good quality, i.e., paths consist of few hops. 1.2. Related work Recent work of [3], done independently of our work, has considered the following red–blue set cover problem. We are given a finite set V , and a collection E of subsets of V . V has been partitioned into red and blue elements. The objective is to choose a collection C ⊆ E that minimizes the number of red elements covered, subject to all blue elements being covered. Here, C = ∪ ft . Letting k denote {f1 , f2 , . . . , ft } covers v ∈ V if and only if v ∈ f1 ∪ f2 ∪ · · ·√ the maximum number of blue elements in any f ∈ E, a 2 k|E|-approximation algorithm for this problem is shown in [3]. It is also proved in [3] that the red–blue set cover problem is a special case of GVR, and that there is an approximationpreserving L-reduction from SLC to the red–blue set cover problem, thus getting the same hardness result for GVR as we do in (1.2) above. Furthermore, red–blue set cover is shown to be equivalent to Minimum Monotone Satisfying Assignment in [3]. It is not clear if there is a close relationship between red–blue set cover and VR; in particular, the results of [3] do not seem to imply our results for VR, or our approximation algorithms for GVR. Also, a variant of GVR where we have laid-out sets and where we wish to find a spanning tree edge-intersecting the minimum number of laid-out sets, has been studied in [14]. This paper gives a polynomial-time approximation algorithm with logarithimic performance guarantee, and show that, unless P = N P there is no polynomial-time algorithm whose approximation gurantee is constant. 1.3. Preliminaries The number of vertices and arcs of the input multigraph G will be denoted by n and m, respectively. Given any VR instance, we may assume without loss of generality that every node is reachable from s, and t can be reached from each node, implying m  n if we exclude the trivial case in which G is a path. The maximum length of a laid-out path will be denoted by , and the number of laid-out paths by p. As G is a multigraph, in principle m may be arbitrarily large with respect to n. In fact, for each pair of vertices u, v, we can assume that there is at most one arc uv not belonging to any laid-out path, whereas the other arcs of the form uv will belong to distinct laid-out paths. Hence, we can assume m  n2 + p  n(n + p). Moreover, the fact that the laid out paths are arc disjoint implies m  p; therefore the size of a VR (or WVR) instance is always Θ(m). G will always be a (simple) graph in our lower bound proofs. So our lower bounds apply to graphs, while our upper bounds apply to multigraphs as well. We will only consider optimization problems for which every feasible solution has a non-negative objective function value. Throughout the paper, optP will denote the optimal solution value to an optimization problem P. When no

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

99

confusion arises, we will simply denote it by opt. Given a feasible solution y to an instance x of P, val(x, y) denotes the value that the objective function of P takes on the feasible solution y. Consider a minimization (respectively maximization) problem. Given a parameter ρ  1, we say that an algorithm is a ρ-approximation algorithm if the value of the solution it returns is at most ρ · opt (respectively is at least opt/ρ). Similarly, given a parameter k and a function f (·), an O(f (k))approximation algorithm is an algorithm that returns a solution whose value is O(f (k) · opt) (respectively O(opt/f (k))). We say that this heuristic approximates the problem within (a factor) O(f (k)). A problem F is APX-hard if there exists some constant σ > 1 such that F cannot be σ -approximated in polynomial time unless P = N P. Occasionally we will establish hardness under different complexity assumptions such as N P ⊆ QP, N P ⊆ ZPP, etc. The following definitions follow [19]. Given two optimization problems A and B, an L-reduction from A to B consists of a pair of polynomial-time computable functions (f, g) such that, for two fixed constants α and β: (a) f maps input instances of A into input instances of B; (b) given an A-instance a, the corresponding B-instance f (a), and any feasible solution b for f (a), g(a, b) is a feasible solution for the A-instance a; (c) |optB (f (a))|  α|optA (a)| for all a; and (d) |optA (a) − val(a, g(a, b))|  β|optB (f (a)) − val(f (a), b)| for each a and for every feasible solution b for f (a). From this definition it follows that the relative errors are linearly related, i.e. |optB (f (a)) − val(f (a), b)| |optA (a) − val(a, g(a, b))|  αβ . optA (a) optB (f (a)) (Thus, the “L” in “L-reduction” stands for “linear.”) From this the following wellknown fact follows; it will be used repeatedly in the paper. Fact 1.1. Given optimization problems A and B and an L-reduction from A to B with parameters α and β, the existence of a polynomial-time (1 + ε)approximation algorithm for B implies the existence of the following types of polynomial-time approximation algorithms for A: (i) an (1 + αβε)-approximation algorithm, if both A and B are minimization problems; (ii) an (1 − αβε)−1 -approximation algorithm, if A is a maximization problem and B is a minimization problem and if αβε < 1; (iii) an (1 − αβε/(1 + ε))−1 -approximation algorithm, if both A and B are maximization problems and if αβε < 1 + ε. Notice also that L-reductions map optimal solution into optimal solutions.

100

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

2. Hardness of approximation Recall that stands for the maximum length of any laid-out path of the VRinstance. 2.1. Hardness of VR when is bounded In this subsection we show that the problem is APX-hard even for the special case in which equals 3. We will show in Section 3 that the problem is polynomially solvable for = 1, 2 and that there is a polynomial time /2 approximation algorithm. Our reduction is from the M AX 3-S ATISFIABILITY (MAX 3-SAT) problem in which one is given a collection of CNF clauses with exactly three literals per clause, and wants to find a truth assignment to the variables so as to maximize the number of clauses satisfied. To get a reasonable constant we make use of a result of Håstad [11] which shows that if we can approximate MAX 3-SAT in polynomial time to within 8/7 − ε with ε > 0 being any fixed constant, then N P = P. Notation. Given a MAX 3-SAT instance, we let px and qx denote the number of positive and negative occurrences of a variable x in it, respectively. For any constant λ  1, let MS(λ) denote the family of MAX 3-SAT instances in which px  λqx and qx  λpx for all variables x. Theorem 2.1. For any constant λ  1, there is an L-reduction from MS(λ) to VR (wherein = 3), with α=

25λ + 1 7(λ + 1)

and β = 1.

Proof. Let x1 , . . . , xn and C1 , . . . , Cm be the variables and clauses of the given MS(λ) instance. To each variable x we associate a variable gadget depicted in Fig. 1; two parallel paths going from the entry vertex ENTRY(x) to the exit vertex EXIT(x). The paths have Mx := max{px , qx } many arcs each. Consider any ordering of the variables and, for each variable, an arbitrary ordering of

Fig. 1. The gadget for a variable x with px = 5 and qx = 4.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

101

Fig. 2. The gadget associated with the clause C = (x ∨ y ∨ z¯ ), where x’s occurrence is the ith positive occurrence, y’s occurrence is the j th positive occurrence, and z’s occurrence is the kth negative occurrence.

its occurrences. The ith arc of the upper path corresponds to the ith positive occurrence of x and has label x i . The ith arc of the lower path corresponds to the ith negative occurrence of x and has label x¯ i . Since the number of positive and negative occurrences may differ, some arcs may have no label. To every clause C we associate a clause gadget like the one depicted in Fig. 2; there are three parallel arcs going from the entry point ENTRY (C) to the exit point EXIT (C). If the occurrence of x in C is its ith positive occurrence, the label of the corresponding arc in the clause gadget is x i . If the occurrence of x in C is its ith negative occurrence, the label is x¯ i . These gadgets are strung together as follows. The vertex EXIT (xi ) is identified with ENTRY (xi+1 ), for i = 1, . . . , n − 1. The vertex EXIT (xn ) is identified with ENTRY (C1 ), and EXIT (Ci ) is identified with ENTRY (Ci+1 ), for i = 1, . . . , m − 1. The source vertex s coincides with ENTRY (x1 ) while the sink vertex t coincides with EXIT (Cm ). The laid-out paths can have length either 1 or 3. For each positive occurrence of a variable x, say the ith, we form a path P (x, i) of length 3. Let C be the clause containing the ith positive occurrence of x. The first arc of P (x, i) is the arc with label x i of the gadget of the clause C; the second arc is a new arc fx,i , joining the head of the first arc to the tail of the third arc; the third arc is the arc with label x i of the gadget corresponding to the variable x. Analogously, we have a path N(x, j ) for each negative occurrence of x. The first arc comes from the clause gadget and has label x¯ j , the second is a new arc gx,j , and the third comes from the variable gadget and has label x¯ j . Arcs of the variable gadgets without label form special laid-out paths of length 1. The overall structure of the construction is therefore that of a layered directed acyclic graph with some additional backward arcs (the fx,i ’s and gx,j ’s) going from clause gadgets back to variable gadgets. The reduction can obviously be carried out in polynomial time. To prove that it is an L-reduction we establish some preliminary facts. The first observation is that, as mentioned before, we can assume without loss of generality that any solution to the VR instance is a simple path from s to t (i.e., no loops). In particular, for each variable gadget we either

102

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

take the upper path, corresponding to the positive occurrences of the variable, or the lower path. For each clause gadget we traverse exactly one arc, corresponding  to one variable occurrence. Letting M := ni=1 Mxi , we now see that any path from s to t has cost of the form M + m − k, where m is the number of clauses of the boolean formula, and k is a function of the path that will be determined next. A cost of M is incurred because, in order to traverse every variable gadget, we must use either the upper or the lower path. To traverse the clause gadgets we will pay a cost of 1 for every clause, unless we can take an arc belonging to a path we have already paid for when traversing the variable gadget. Therefore the reduction establishes a bijection between feasible solutions of the MS(λ) and VR instances specified as follows. To an assignment that satisfies clauses C1 , . . . , Ck there corresponds a path P from s to t that uses only M + m − k laid-out paths, defined as follows: For each variable gadget, take the upper part if the variable is set to True, and the lower part otherwise. For each clause, take an arc of a path you have already paid for, if it exists (i.e., if the clause is satisfied), or an arbitrary arc otherwise (i.e., if the clause is not satisfied). Conversely, a solution P to VR of cost M + m − k can be mapped in polynomial time into a satisfying assignment satisfying k clauses: A variable is set to True if P traverses the upper portion of a variable gadget, and to False otherwise. Thus, the two optimal solution values, optVR for the VR instance defined and optSAT for the original MS(λ) instance, are related as follows: optVR = M + m − optSAT , and the reduction is an L-reduction with β = 1. To find the value of α, consider any variable x. Since px  λqx and qx  λpx by the definition of MS(λ), we can verify that Mx  (λ/(λ + 1))(px + qx ). So, M=

n 

Mxi 

i=1

n  λ 3λ · · m. (px + qx ) = λ+1 λ+1 i=1

Then, recalling that optSAT  7m/8, we obtain the bound  optVR = M + m − optSAT   

 25λ + 1 optSAT , 7(λ + 1)

i.e., α = (25λ + 1)/(7(λ + 1)).

 3λ m + m − optSAT λ+1



Corollary 2.1. The existence of a polynomial-time (183/176 − ε)-approximation algorithm for VR, even in the case = 3 and for any fixed ε > 0, implies N P = P.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

103

Proof. Suppose indeed that for some constant ε < 7/176, there is a polynomialtime (183/176 − ε)-approximation algorithm for VR, in the case = 3. Assuming this, we present a polynomial-time approximation algorithm for MAX 3-SAT. Suppose we are given a MAX 3-SAT instance. We will first run a preprocessing algorithm that repeatedly: (i) sets certain variables x (appropriately to True or False), and (ii) deletes all clauses that contain the variable x. Let δ be a positive constant that will be specified below; let px and qx respectively denote the number of positive and negative occurrences of a variable x in the current (shrunk) formula. The preprocessing is as follows. If there is a variable x with px > (7 + δ)qx , then we set x to True and delete all clauses that contain x; if there is a variable x with qx > (7 + δ)px , we set x to False and delete all clauses that contain x. We repeat this process until we have px  (7 + δ)qx and qx  (7 + δ)px for all remaining variables x in the remaining formula F  . We then run the L-reduction of Theorem 2.1 on F  with λ = 7 + δ, and apply the claimed (183/176 − ε)-approximation algorithm on the resulting VR instance to get an approximately optimal solution for the remaining formula F  . Let us analyze the above algorithm. Given an original formula F , we produce a shrunk formula F  ; let F  denote the formula induced by the clauses deleted in the preprocessing. Let opt, opt , and opt denote the optimal objective function values for the MAX 3-SAT instances F , F  , and F  , respectively. Since the set of clauses in F is the disjoint union of those of F  and F  , we have opt  opt + opt . It is easy to see that our preprocessing algorithm satisfies at least an a1 := (7 + δ)/(8 + δ) fraction of the clauses of F  . Also, Theorem 2.1 and Fact 1.1(ii) show that we satisfy at least an   25(7 + δ) + 1 ·ε a2 := 1 − 7(7 + δ + 1) fraction of opt , for the formula F  . Thus, we satisfy a total of at least a1 opt + a2 opt  min{a1 , a2 } · (opt + opt )  min{a1 , a2 } · opt

(1)

clauses. Setting δ = (7 − 176ε)/25ε to equate a1 and a2 , we get that min{a1 , a2 } = (8/7 − δ  )−1 , where δ  = δ  (ε) is some positive constant. Thus, (1) shows that we have an (8/7 − δ  )-approximation algorithm for MAX 3-SAT, which would imply that N P = P [11]. ✷ 2.2. A stronger negative result when is unbounded In this subsection we exhibit an L-reduction from the S YMMETRIC L ABEL C OVER (SLC) problem to VR. The reduction implies that, unless N P ⊆ QP, 1/2−ε m VR is hard to approximate within O(2log ) for any fixed ε.

104

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

The problem SLC is the following. The input of the problem is a bipartite graph H = (L, R, E) and two sets of labels: A, to be used on the left-hand side L only, and B, to be used on the right-hand side R only. Each edge e has a list Pe of admissible pairs of the form (a, b) where a ∈ A and b ∈ B. A feasible solution is a label assignment Au ⊆ A and Bv ⊆ B for every vertex u ∈ L and v ∈ R such that, for every edge e = uv, there is at least one admissible pair (a, b) ∈ Pe with a ∈ Au andb ∈ Bv . The objective is to find  a feasible label assignment which minimizes u∈L |Au | + v∈R |Bv |. Let q := e∈E |Pe | and note that the size of an SLC instance is equal to Θ(q). 1/2−ε q )-approximation algorithm The existence of a polynomial-time O(2log for SLC, for any ε > 0, implies that every problem in N P can be solved by a quasi-polynomial-time algorithm, i.e., N P ⊆ QP [6,13]. Here, we exhibit an Lreduction from SLC to VR with α = β = 1. Theorem 2.2. There is an L-reduction from SLC to VR with α = β = 1. Proof. For notational convenience, let A = {a1, a2 , . . . , a|A| } and B = {b1 , b2 , . . . , b|B| }. To each edge e of the SLC instance H = (L, R, E), we associate an edge gadget shown in Fig. 3. The gadget consists of many parallel paths, each formed by 5 arcs, from the entry point ENTRY (e) to the exit point EXIT (e). There is one path for every pair in the list Pe . On the path corresponding to the pair (a, b), the second arc on the path is labeled with a, while the fourth is labeled with b (the other arcs in the paths have no label). These edge gadgets are strung together as follows. Let e1 , . . . , em be the edges of H . For i = 1, . . . , m − 1, the exit point EXIT (ei ) is identified with the entry point ENTRY (ei+1 ). The source vertex s coincides with ENTRY (e1 ) while the sink vertex t coincides with EXIT (em ). Let the resulting graph be denoted by G. The VR instance will be obtained by adding arcs to G, as explained below. Notice that G is a planar, layered, acyclic graph that can be easily laid out on a piece of paper by drawing the edge gadgets from left to right. We use this to define a total ordering between

Fig. 3. An edge gadget for an edge e whose admissible pairs are (a1 , b1 ), (a1 , b2 ), (a2 , b3 ), and (a3 , b3 ).

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

105

the arcs of G, as follows. An edge a is smaller than an edge b if a belongs to a layer to the left of that of b or, in case they belong to the same layer, if a is above b. The laid-out paths are constructed as follows. For each vertex u ∈ L and label a ∈ A, there is a laid-out path Pua . We define the set S of arcs of G that will be contained in Pua as follows. Consider all edges incident with u which have admissible pairs with a as a first coordinate. For each such edge, let S contain the arcs labeled a in the corresponding gadget. Let w1 , w2 , . . . , wk be these edges ordered according to the total ordering described above. We introduce “backward” edges fi joining the head of wi to the tail of wi−1 , for 2  i  k. This defines the laid-out path Pua = wk fk wk−1 . . . f2 w1 . Analogously, define “backward” paths of the form Pvb = xk gk xk−1 . . . g2 x1 , where the xi ’s are the arcs with label b ∈ B belonging to gadgets of edges incident with v ∈ R. This reduction establishes a bijection between feasible solutions which preserves the costs. For the proof we need a couple of observations. The first is that it suffices to restrict oneself to feasible solutions of SLC which are minimal. That is, the solution can be constructed by processing each edge once, as follows. When an edge e = uv is processed a label a is added to the partial solution Au so far, and a label b is added to the partial solution Bv computed so far, only if the current set Bu × Bv does not contain an admissible pair for e. A second observation is that without loss of generality we can consider only solutions to the VR instance of the following form. To generate the path, we start walking from the source towards the sink. When we enter each arc gadget we choose one of the parallel paths reaching the exit of the gadget. We never use a “backward” arc fi or gj . We stop when we reach the sink. We call these the canonical solutions. The reduction establishes a cost-preserving reduction between minimal label covers and canonical solutions to VR. Furthermore, given one, it is possible to obtain the other in polynomial time. This is because a minimal solution is nothing else than a way to prescribe, for each arc e, which label pair to choose. By construction, this set of choices can be mimicked exactly in the VR instance with a canonical solution, and vice versa. ✷ Noting that, in the L-reduction of Theorem 2.2, the number of arcs of the  VR instance defined is Θ( e∈E |Pe |), we get the following corollary. 1/2−ε m

Corollary 2.2. The existence of a polynomial-time O(2log algorithm for VR, for any fixed ε > 0, implies N P ⊆ QP .

)-approximation

2.3. The hardness of approximating Gondolier Avoidance Recall that GA, the complementary problem to VR, asks for the maximum number of laid-out paths which can be removed without disconnecting t from s.

106

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

We start by showing a very simple L-reduction from S ET C OVER (SC) to VR. Recall that an instance of SC is given by a ground set E = {e1 , . . . , ef } and a family S = {I1 , . . . , Iq } of subsets of E. The objective is to select a minimumcardinality collection of subsets in S such that each element in the ground set is contained in at least one subset of the collection. Theorem 2.3. There is an L-reduction from SC to VR with α = β = 1. Proof. Given an SC instance, define the following VR instance. For i = 1, . . . , f , let Si = {Ij1 , . . . , Ijs } be the family of sets in S containing element ei . To each element ei we associate a gadget formed by |Si | parallel arcs going from ENTRY (ei ) to EXIT (ei ), the hth of these arcs being associated with subset Ijh . The gadgets are connected by identifying EXIT (ei ) with ENTRY (ei+1 ) for i = 1, . . . , f − 1. Moreover, the source vertex s coincides with ENTRY (e1 ) while the sink vertex t coincides with EXIT (ef ). There is a laid-out path associated with each subset Ij , defined as follows. Consider subset Ij = {ei1 , ei2 , . . . , eir−1 , eir }, with r = |Ij |, assuming i1 < i2 < · · · < ir−1 < ir . The laid-out path associated with Ij has the form Pj = ar fr ar−1 . . . f2 a1 , where, for k = 1, . . . , r, ak is the arc associated with Ij in the gadget corresponding to element eik , while fk (2  k  r) is an appropriate “backward” arc from the head of ak to the tail of ak−1 . In other words, Pj contains the arcs associated with subset Ij in the gadgets of elements ei ∈ Ij , and these arcs appear in the path according to decreasing index of elements. As in the L-reductions presented before, it is easy to see that any simple path from s to t intersecting k laid-out paths corresponds to an SC solution of value k and vice versa. ✷ The V ERTEX C OVER (VC) problem is the special case of SC where each element of the ground set is contained in exactly two subsets. The problem therefore can be represented by an undirected graph G where vertices correspond to subsets and edges denote non-empty intersections between pairs of subsets. If we complement the objective function of VC, we have the well-known I NDEPENDENT S ET (IS) problem. Hence, Theorem 2.3 yields also an L-reduction from IS to GA with α = β = 1. Using the inapproximability results for IS from [10], and noting that the number of laid-out paths in the GA instances defined by Theorem 2.3 is equal to the number of vertices in the IS instance, yields the following corollary. Corollary 2.3. Let p denote the number of laid-out paths of a GA instance. The existence of a polynomial-time O(p1−ε )-approximation algorithm for GA, for any ε > 0, implies N P ⊆ ZPP. The existence of a polynomial-time O(p1/2−ε )approximation algorithm for GA, for any fixed ε > 0, implies N P = P.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

107

A more careful evaluation of the reduction above also shows the following. Corollary 2.4. Let m denote √ the number of arcs of a GA instance. The existence of a polynomial-time O( m1−ε )-approximation algorithm for GA, √ for any ε > 0, implies N P ⊆ ZPP. The existence of a polynomial-time O( m1/2−ε )approximation algorithm for GA, for any fixed ε > 0, implies N P = P. Proof. If G = (V , E) is the original IS instance, then f = |E| and q = |V |. For the GA instance defined by the reduction, m = Θ(f ), whereas p = q √ as noted above. Since we have |E| = O(|V |2 ), we have m = O(p2 ), i.e., p = Ω( m) and the claim follows from Corollary 2.3. ✷

3. Approximation algorithms Having shown that even some simple cases of the problem are computationally difficult, we turn to presenting approximation algorithms. Our algorithm are variations of the following idea. We divide the laid-out paths into “long” and “short.” To find an st-path we take all “long” paths, for there cannot be too many of these. Then we put a weight of 1 on every arc belonging to each “short” path, a weight of 0 on all other arcs and run a shortest path algorithm (henceforth SPA). This weighting scheme overestimates the cost of using “short” paths and hence they will be used parsimoniously. In many cases, we will try all the possible values of opt. Note that, for VR, we know that opt  n − 1, as the solution may be assumed to be a simple path, and opt  p, as in the worst case all the laid out paths are intersected by the optimal solution. (In short, we can assume w.l.o.g. that the value of opt be known.) The term µ := min{n − 1, p} will often appear in the time complexity of the algorithms we present. We will use a weight function defined as follows. Given a non-negative integer x, for each arc a of the graph, let  1, if a belongs to some laid-out path of length  x, zx (a) := 0, otherwise. The next fact is obvious but we record it for future reference. Lemma 3.1. Let P be the SPA solution for weights zx , and let opt be the optimum of VR. Then P intersects at most opt · x forbidden paths of length at most x. 3.1. Constant factor approximation for fixed Recall that the maximum length of a laid-out path is denoted by . We now present an /2 -approximation, which is especially useful in the common case

108

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

where is “small” (in practice, lightpaths will typically only consist of a few arcs). The first algorithm we present, called A0 , initially assigns weight 1 to each arc that lies in some laid-out path, and weight 0 to all other arcs. Then, for each laidout path of length at least 2 and consecutive arcs (u, v), (v, t) in the path, A0 adds a new arc (u, t) with weight 1. Finally, A0 applies SPA to the graph obtained. At the end, possible new arcs are replaced by the corresponding pair of “real” consecutive arcs. Also, since all arc-weights lie in {0, 1}, SPA only takes O(m) time here. The following fact follows from Lemma 3.1, since the new “shortcut” arcs (such as (u, t)) that we add above essentially “reduce” to /2 . Fact 3.1. Algorithm A0 returns a solution of value at most opt · /2 in O(m) time. The following corollaries are immediate consequences of the above fact. Corollary 3.1. Algorithm A0 is an O(m)-time /2 -approximation algorithm for VR. Corollary 3.2. Algorithm A0 is an O(m)-time exact algorithm for VR when  2. Basically, we get an /2 –approximation by “short-cutting” sub-paths of length 2. This follows from the following fact. Suppose e1 , e2 , . . . is the sequence of edges in some laid-out path Q, and let P be an optimal st-path. Then, our shortcutting works fine if P uses ei and then ei+1 , for some i. Also, since P can be assumed to be acyclic, we can assume that P does not use ei+1 , and then ei . This is the reason why our short-cutting works. However, since P can use ei+a and then ei , for a  2, we cannot in general short-cut sub-paths of length more than 2. It is also easy to see that the above approach yields an O(SSSP(n, m))-time

/2 -approximation algorithm for PVR. For GVR, we use the same algorithm, except that we do not add any “shortcut arcs” such as (u, t) above. Lemma 3.1 shows that this is an O(m)-time -approximation algorithm for GVR. 3.2. An approximation algorithm for unbounded We next show an approximation algorithm for√unbounded , called A1 . The approximation guarantee of this algorithm is O( m/opt); we then show how this factor of opt in the denominator of the approximation bound, can be used to show that VR (and hence SLC) are different from the C HROMATIC N UMBER problem, in a certain quantitative sense related to approximability.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

109

√ Let x := m/opt. As opt is unknown, Algorithm A1 tries all the O(µ) possible values opt, considering the associated x value. Given the graph G and any non-negative integer x, let f (G, x) := (# of laid-out paths of length > x). For each value of x, A1 assigns weight zx (a) to each arc a of the graph, and applies SPA to the corresponding instance. On output, A1 gives the best solution found by the SPA calls for the different values of x. Lemma 3.2. For any non-negative integer x: (a) f (G, x)  m/(x + 1); (b) the solution obtained by applying SPA after having assigned weight zx (a) to each arc a of the graph has value at most opt · x + f (G, x). Proof. Since all laid-out paths are arc-disjoint, the number of paths of length more than x is at most m/(x + 1), yielding (a). Moreover, for each x, we know by Lemma 3.1 that the number of paths of length  x intersected by the SPA solution with weights zx is at most opt · x. So the total number of paths intersected is at most opt · x + f (G, x), showing (b). ✷ Theorem 3.1. (a) Algorithm A1 runs in O(µm) time, and returns a solution of √ √ value O( opt · m ). (b) A solution of value O( opt · m ), with a slightly worse constant than in the bound of (a), can also be computed in time O(m log µ). Proof. By Lemma 3.2 (a) and (b), the total number of paths intersected is at most √ m/(x + 1) + opt · x. This quantity is√minimized when x is approximately

m/opt , and the minimum value is O( opt · m ). Since we run SPA for each possible value of opt, the running time follows, for part (a). Part (b) follows by guessing opt = 1, 2, 4, . . . . ✷ √ Corollary 3.3. Algorithm A1 is an O( m/opt )-approximation algorithm for VR. √ The above approach yields an O( m/opt )-approximation algorithm for GVR as well. Notation (q-exhaustive search). For a given VR instance, suppose we know that opt is at most some value q. Then, the following simple algorithm to find an optimal solution, called q-exhaustive search, will be useful in a few contexts from now on: Enumerate all subsets X of the set of laid-out paths such that |X|  q, and check if the laid-out paths in X alone are sufficient to connect s to t. The time complexity of this approach is   q   q     p m O m· . O m· i i i=0

i=0

In particular, if q is a constant, then we have a polynomial-time algorithm.

110

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

Some separation results. We now present two “separation” results that indicate that VR is likely to be in an approximation class different from that to which C HROMATIC N UMBER (CN) belongs. First, we show that, unless P = N P, there can be no L-reduction from CN to VR. Second, we use Theorem 3.1 to show that VR (and hence also SLC, in view of Theorem 2.2) has certain subexponentialtime algorithms which are unlikely to exist for CN (and also for other problems such as I NDEPENDENT S ET). Proposition 3.1. The existence of an L-reduction from CN to VR implies N P = P. Proof. Let us consider a subclass S of CN-instances that have this property: a graph G in S is always 4-colorable and maybe 3-colorable, but it is N Pcomplete to tell the two cases apart. (Such classes of instances arise naturally from N P-hardness proofs. An example is those graphs obtained via the polynomialtime reduction from 3-SAT to CN [9,20]. Another example is given by the construction of Holyer showing that edge-coloring cubic graphs is N P-hard [9, 12].) Let us assume there is an L-reduction from CN to VR. We shall use it to solve CN on S in polynomial time. The L-reduction obviously works for S too. For the VR instances obtained via this L-reduction on CN instances in S, we have optVR  α · optCN  4α for some constant α that depends on the L-reduction. So, to find an optimal solution for these VR instances, it suffices to do a polynomial-time (4α)exhaustive search. Once the VR optimum is obtained, it is mapped back to the optimum of the CN instance via the L-reduction (L-reductions map optimal solutions into optimal solutions). ✷ We now switch to the second result of this subsection. Let N denote the size of an instance of SAT, and s denote the number of vertices in a CN instance. Note that the size S of a CN instance is at most O(s 2 ). Given any fixed ε > 0 and any oracle that approximates CN to within O(s 1−ε ), work of [8] presents a randomized polynomial-time algorithm that uses this oracle to correctly solve SAT instances with a probability of at least 2/3. Thus, if a positive constant δ δ is such that any randomized algorithm for SAT must have running time Ω(2N ),   Sδ then there √ is a constant δ > 0 such that for any constant ε > 0, there is no O(2 )time O( S 1−ε )-approximation algorithm for CN. We now show that the situation is different for VR. √ ε Theorem 3.2. For any fixed ε > 0, there is a 2O(m log m) -time O( m1−ε )approximation algorithm for VR. Proof. As before, we assume we know opt, in the sense that we are allowed to try the procedure below for all the O(µ) possible values of opt. Suppose a positive

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

111

constant ε is given. Theorem 3.1 shows that in√polynomial time, we can find a √ VR solution of value O( opt · m) = O(opt · m/opt). Thus, if opt√ mε , we √ can approximate the problem to within a factor O( m/opt)  O( m1−ε ) in polynomial time. Accordingly, we assume from now on that opt  mε . In this case, we can do an mε -exhaustive search to find an optimal solution, with running  ε m O(mε ) = 2O(mε log m) . ✷ time O(m · m i=0 i ), which is m · m 3.3. Randomized sparsification: improvements when opt is small √ We next show how to improve on the O( opt · m) solution value for VR in case opt is “small,” i.e., O(log n). In particular, for any constant C > 0, we present a randomized polynomial-time Algorithm A2 that with high probability  computes a solution of value at most O( opt · m/nC/opt ). For instance, if opt = √ √ log n, A2 computes a solution of value O( opt · m/2C log n ). The difference between A1 and A2 is a randomized “sparsification” technique. Note that, due to Corollary 3.3, the following discussion is interesting only if opt  log n. Suppose that for a given constant C > 0, we aim to find a solution of value  O( opt · m/nC/opt ). We will assume throughout that n is sufficiently large as a function of C, i.e., n  f (C) for some appropriate function f (·). Indeed, if n < f (C), then opt  log n is bounded by a constant, and we can find an optimal solution by a polynomial-time (log n)-exhaustive search. As in Algorithm A1 , Algorithm A2 has to try all the possible O(µ) values of opt, computing a solution for each, and returning on output the best one found. In the following, we will √ consider the iteration with the correct opt value. Let λ := nC/opt and x :=  m/(opt · λ), where C is the given constant. If opt  2C, Algorithm A2 finds an optimal solution by a (2C)-exhaustive search. Hence, in the following we describe the behavior of A2 for opt > 2C. By the definition of x and the fact that m  n, we have

m opt · m   n1/4 . (2) λx λ The deletion of a laid-out path corresponds to the removal of all the associated arcs from G. For each possible value of opt, Algorithm A2 uses the following “sparsification” procedure: Each laid-out path is independently deleted with probability (1 − 1/λ). Then, SPA is applied on the resulting graph, after having assigned weight zx (a) to each arc a that was not removed. The sparsification procedure and the associated shortest path computation are repeated O(nC ) times, and the best solution found is output. Theorem 3.3. Given any constant C >0, Algorithm A2 runs in polynomial time and returns a VR solution of value O( opt · m/nC/opt ) with high probability.

112

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

Proof. Let P be an optimal VR solution. We have that Pr[all laid-out paths intersected by P survive] = (1/λ)opt = n−C .

(3)

Moreover, let Z be the number of laid-out paths of length more than x that remain undeleted. By Lemma 3.2(a) and linearity of expectation, E[Z]  m/(λx). Using a simple Chernoff bound [18], we get   2m 1/4 Pr Z  (4)  (e/4)m/(λx)  (e/4)n λx by (2), where e denotes the base of the natural logarithm as usual. As mentioned earlier in this subsection, we may assume that n is large enough: 1/4 We will assume that n−C − (e/4)n  1/(2nC ). Thus, with probability at least 1/4 n−C − (e/4)n  1/(2nC ), we have that (i) all laid-out paths intersected by P survive, and (ii) Z  2m/(λx). Thus, sparsification ensures that f (G, x) is at most 2m/(λx) after path deletion. By Lemma  3.2, running SPA yields a solution of value at most opt · x + 2m/λx = O( opt · m/nC/opt ). The above success probability of 1/(2nC ) can be boosted to, say, 0.9 by repeating the procedure O(nC ) times. The time complexity follows immediately from the algorithm description. ✷ √ Corollary 3.4. Algorithm A2 (with C = 1, say) is a randomized O( m/ log n )approximation algorithm for VR. (In other words, Algorithm √ A2 always runs in polynomial time, and delivers a solution that is at most O( m/ log n ) times the optimal solution with high probability.) Proof.  By Theorem 3.3, A2 with C = 1, computes a solution of value opt · O( m/(opt · n1/opt )). Now, elementary calculus shows that opt · n1/opt is 1/opt )) is at most minimized √ when opt = Θ(log n). Thus, opt · O( m/(opt · n opt · O( m/ log n ). ✷ In the next subsection we show that Algorithm A2 can be derandomized. 3.4. Derandomizing Algorithm A2 Our aim is to prove the following theorem. Theorem 3.4. For any given constant C > 0, there is a deterministic polynomialtime approximation algorithm for VR that returns a solution of value O( opt · m/nC/opt ). The proof will go through several lemmas and observations. The basic starting point is that randomized algorithms are often robust, i.e., they remain correct even

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

113

if some small changes are made to the underlying distribution. This opens up a well-known approach to derandomization: We first construct a small sample space that approximates a given (larger) sample space sufficiently well. We may then conduct an exhaustive search in the small sample space to find a sample point that is sufficiently good for a randomized algorithm in consideration. More precisely, we will need the following useful result of [7]. Suppose Y1 , Y2 , . . . , YN ∈ {0, 1, . . . , a − 1} are independent random variables with arbitrary individual distributions and joint distribution D. We define a finite set 1 , . . . , Y N ) S ⊆ {0, 1, . . . , a − 1}N to be a (d, ε)-approximation for D if, for (Y sampled uniformly at random from S, we have for all I ⊆ {1, 2, . . ., N} with |I |  d and for all γ1 , γ2 , . . . , γ|I | ∈ {0, 1, . . . , a − 1} that       

 Pr i = γi − (5) Y Pr[Y = γ ] i i   ε.  i∈I

i∈I

It is shown in [7] that such a set S with cardinality poly(2d , log N, 1/ε) can be constructed explicitly. We will also require a useful lemma of [2]. The following result is Lemma A.4 of [2]. Lemma 3.3 [2]. Let t  2 be an even integer. Suppose Y1 , . . . , Yn are independent random variables taking values in [0, 1]. Let Y = Y1 + · · · + Yn and µ = E[Y ]. Then  t /2 √   t /2

5 t 1/(6t ) E (Y − µ)  2e πt · tµ + t 2 . 2e We present a useful corollary. Corollary 3.5. Suppose Y1 , Y2 , . . . , YN ∈ {0, 1} are independent random variables, and that Y = Y1 + · · · + YN with µ = E[Y ]. Let S ⊆ {0, 1}N be any (d, ε)1 , . . . , Y N ) be chosen approximation for the joint distribution D of the Yi . Let (Y = Y 1 + · · · + Y N . Then for any even uniformly at random from S, and define Y positive integer t  d and any b > 0, √    2e1/(6t ) πt(5/(2e))t /2 · (tµ + t 2 )t /2 + ψ   Pr Y − µ  b  , bt where ψ = ε(Nµ/t)t et /µ if µ  1, and ψ = ε(N/t)t et otherwise. Proof. By the standard tth moment inequality,    Pr  Y − µ  b  − µ)t ] E[(Y  bt

(6)

114

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

t

i=0 (−1)

=

t =

i=0 (−1)

t −i µt −i t −i µt −i

 



I ⊆{1,...,N}: |I |=i bt

E

I ⊆{1,...,N}: |I |=i bt

Pr

j ∈I



j Y



 = 1)

j ∈I (Yj

 .

(7)

 j = 1)] Consider any set I in (7). Since |I |  t  d, we haveby (5) that Pr[ j ∈I (Y is within an additive distance of at most ε from Pr[ j ∈I (Yj = 1)]. Thus, t 

     − µ t  E (Y − µ)t + µt −i E Y i=0



ε.

(8)

I ⊆{1,...,N}: |I |=i

Let us bound H (N, µ, t) :=

t 



µt −i

ε.

I ⊆{1,...,N}: |I |=i

i=0

We have   N H (N, µ, t) = εµ µ i i=0 t    N −1 i t t tµ /N  εµ (N/t) i i=0

N  εµt (N/t)t 1 + tµ−1 /N t

t 

−i

 ε(Nµ/t)t et /µ . If µ < 1, we see that H (N, µ, t)  H (N, 1, t)  ε(N/t)t et . Lemma 3.3, (6), and (8) now help complete the proof. ✷ Define t to be the smallest even integer that is at least as high as 9C, and let d = log n ; we assume n is large enough so that t  d. (Recall from the beginning of this subsection that we may assume that n is large enough.) Also define ε = m−9C−4 . Recall that in Algorithm A2 , each laid-out path is independently deleted with probability (1 − 1/λ). Let Yi be the indicator random variable for the ith laid-out path (under an arbitrary but fixed ordering of the laidout paths) not being deleted, and let D denote the joint distribution of all the Yi . (Thus, D is the product distribution of the Yi , where Pr[Yi = 1] = 1/λ for each i.) Define S to be any (d, ε)-approximation for D; as seen above, the work of [7] provides an explicit construction of such an S of cardinality poly(n, m). We will 2 , . . . , Y p ) uniformly at random 1 , Y now show that it in fact suffices to choose (Y  from S, and delete the ith path if and only if Yi = 0; thus, since S is polynomialsized, we can derandomize the algorithm by exhaustive search over S. The values of x and λ are the same as in Section 3.3.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

115

Recall that Algorithm A2 is of relevance only in the case where opt  log n. Let P be an optimal VR solution. Suppose, as mentioned above, that we delete i = 0. Then, since opt  d, bound (5) shows that in the ith path if and only if Y place of (3), we now have Pr[all laid-out paths intersected by P survive]  (1/λ)opt − ε = n−C − m−9C−4.

(9)

Suppose the laid-out paths of length   more than x are those numbered i1 , i2 , . . . , iN .  =  Define Z = j Yij and Z j Yij . Note that Z is the number of laid-out paths of length more than x that remain undeleted (in our current randomized algorithm which samples from S). Analogous to (4), we wish to upper  2m/(λx)]. Let µ = E[Z]. Let ψ = ε(N · µ/t)t et /µ if µ  1, and bound Pr[Z   2m/(λx)] is at most ψ = ε(N/t)t et otherwise. Thus, Pr[Z    Z − µ  m/(λx) Pr  √ 2e1/(6t ) πt(5/(2e))t /2 · (tµ + t 2 )t /2 ψ  + , (10) t (m/(λx)) (m/(λx))t by Corollary 3.5. We now bound the right-hand side of (10). Recall from Section 3.3 that µ  m/(λx), and that n1/4  m/(λx). Also, t ∈ [9C, 9C + 2] and ε = m−9C−4 . Thus, the first summand in the right-hand side of (10) is at most

−t /2



−t

 O m/(λx)  O n−t /8 O µt /2 · m/(λx)

(11)  O n−9C/8 , where the constant hidden in the O(·) notation depends upon C. As for the second summand in the r.h.s. of (10), we have

t

t

t ψ  ε N · max{µ, 1}/t et  ε Nm/(λxt) et  ε m2 /(λxt) et , since N  m. Thus,

−t  ε(me/t)t = m−9C−4 (me/t)t . ψ · m/(λx)

(12)

Since t ∈ [9C, 9C + 2] and m  n, we see that for large enough n, the bounds   2m/(λx)]  of (11) and (12) add up to at most n−C /4. So, by (10), Pr[Z −C −C n /4. Thus, by (9), we have that with probability at least 3n /4 − m−9C−4  3n−C /4 − n−9C−4 , the events (i) all laid-out paths intersected by P survive, and   2m/(λx), hold simultaneously. In particular, these events (i) and (ii) hold (ii) Z simultaneously with positive probability (if n is large enough), which is what we required for the derandomization. This completes the proof of Theorem 3.4. The same approach and results hold for GVR, by deleting the laid-out sets instead of the laid-out paths.

116

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

3.5. Approximation for the weighted cases Recall that in WVR we have a weight √ for each laid-out path, which equals its length. We now show a simple O( m)-approximation for WVR. A simple algorithm returns a solution of value at most opt2 for WVR. This algorithm, which √ that opt  m. If opt  m, we call A3 , guesses the value of opt as 1, 2, 4, . . . ; note√ we simply include√ all laid-out paths, leading to an O( m)-approximation. Suppose opt < m. We can restrict attention to laid-out paths of length at most opt. Accordingly, the algorithm: (i) removes all the laid-out paths of length > opt; (ii) assigns weight 1 to all remaining arcs that lie in laid-out paths, and weight 0 to arcs that do not lie in any laid-out path; and (iii) applies SPA. Clearly, the path that we get has cost at most opt2 . √ Theorem 3.5. Algorithm A3 is an O(m log m)-time O(min{opt, m})-approximation algorithm for WVR. In order to reduce PVR to VR we proceed in two steps. First, we make use of a general scaling technique due to Crescenzi et al. [5] which shows that without loss of generality one can consider just polynomially bounded weights for the laid-out paths. More precisely, for any given fixed ε > 0, we can do the scaling so as to lose a factor of at most (1 + ε) in the approximation. Here is the sketch of one approach. First, using bisection search, we can assume (as before) that we know the value of opt. Then, all laid-out paths of weight more than opt can be deleted. Also, we can remove the “laid-out” status of any laid-out path of weight less than opt/n2 (i.e., reset its weight to 0), solve the problem, and finally add back the weights of these low-weight paths; this will add at most opt/n to the value of the final solution. Now, by rescaling, we can assume that all the weights lie in the range [1, n2 ]; we can next perturb each weight to the next higher multiple of, say, 1/n2 , to negligibly alter the problem. Thus, we may assume that all weights are integers in a polynomial range. The second step is a very simple reduction from PVR with polynomial weights to VR, defined as follows. Let P = a1 , a2 , . . . , ak , where the ai = ui−1 ui (1  i  k) are arcs, be a laid-out path of weight y. First we replace each arc i xyi where x0i = ui−1 ai = ui−1 ui by a path of length y, Ai = x0i x1i , x1i x2i , . . . , xy−1 and xyi = ui . Then, we form y new laid-out paths by stringing together all arcs xji −1 xji : Pj = xj1−1 xj1 xj2−1 xj2 . . . xjk−1 xjk , 1  j  y. It is easy to see that any s–t path intersecting P in the original graph corresponds to a path intersecting P1 , . . . , Py in the new graph, and vice versa. Letting w denote the maximum weight of a laid-out path in the original graph, we have that the new graph has O(mw) vertices and arcs. Moreover, the maximum length of a laid out path, , is unchanged. This shows the following proposition.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

117

Proposition 3.2. Any polynomial-time O(f ( , m, n))-approximation algorithm for VR yields a polynomial-time O(f ( , mw, mw))-approximation algorithm for PVR with maximum weight w. 3.6. Limits to approximation We briefly illustrate some limitations to approximating VR by some specific natural techniques. First, in terms of approximation algorithms via linear programming (LP) relaxations, there is a natural integer programming formulation for VR, and a corresponding LP relaxation. We have an integer variable yQ ∈ {0, 1} for each laid-out path Q, and a “capacity” xe ∈ {0, 1} for each arc e. Given an st-path P , xe is supposed to be 1 if and only if e lies in P , and yQ is 1 if and only if P arc-intersects Q. We clearly have the constraints that for each laid-out path Q and all e ∈ Q, xe  yQ . The integer programming formulation is to install these capacities  and send a unit flow from s to t respecting these capacities, so as to minimize Q yQ . The LP relaxation is to let the variables xe , yQ lie in [0, 1]. Unfortunately, one can show instances where the optimal solution is as high as Θ( ) times the LP’s optimal solution. We sketch the proof of this gap for odd ; a minor modification works for the case of even . Consider the digraph with vertex set {s, t, u0 , u1 , . . . , u }. The arcs are of three types: • a single laid-out path Q = (u0 , u1 , . . . , u ), • arcs (s, u0 ), (s, u2 ), . . . , (s, u −1 ), and • arcs (u1 , t), (u3 , t), . . . , (u , t). It is easy to see that any st-path must edge-intersect the laid-out path Q. However, the above LP relaxation has a feasible solution of value 2/( + 1): Set yQ = 2/( + 1), and send one unit of flow on each of the paths (s, u0 , u1 , t), (s, u2 , u3 , t), . . . , (s, u −1 , u , t). Furthermore, even some natural pre-processing (e.g., as in our /2 -approximation for VR), still leaves a gap of Θ( ). Second, SPA-based algorithms cannot in general give solutions much better than those here. This intuition can be formalized by constructing an infinite set of instances such that, for any “reasonable” √ weight function f on the arcs, SPA returns a solution of value Θ( ) = Θ( m), whereas the optimal solution has value 1. (For WVR, these values are Θ( 2 ) and , respectively.) By “reasonable” we mean that f can be any function of the length of the laid-out path and of the position of the arc (first arc of the path, second arc of the path, etc.). We consider a class of instances defined by a parameter k  2. Besides vertices s and t, G contains k + 1 vertex-disjoint laid-out paths Q, P 1 , . . . , P k of length = 2k − 1. Path Q visits vertices q1 , q2 , . . . , q2k , whereas path P i ,

118

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

i . In addition to the arcs in the laidi = 1, . . . , k, visits vertices p1i , p2i , . . . , p2k out paths, G contains arcs (s, q2k−1 ), (q2k , q2k−3 ), (q2k−2 , q2k−5), . . . , (q6 , q3 ), i−1 i , p2k−1 ), (q4 , q1 ), (q2 , t); as well as arcs (s, p11 ), (p21 , p32 ), (p42 , p53 ), . . . , (p2k−2 k (p2k , t). Suppose one uses any “reasonable” weight function to the arcs of G, according to the definition above. Then, the optimal VR and WVR solutions are associated with path (s, q2k−1 ), (q2k−1, q2k ), (q2k , q2k−3 ), (q2k−3 , q2k−2 ), . . . , (q4 , q1 ), (q1 , q2 ), (q2 , t), and have cost 1 and 2k − 1, respectively. This is also an optimal solution for SPA, but a solution of the same value is given by the path i−1 i i i i , p2k−1 ), (p2k−1 , p2k ), (p2k , t), whose VR (s, p11 ), (p11 , p21 ), (p21 , p32 ), . . . , (p2k−2 and WVR values are respectively k and (2k − 1)k.

4. Experimental results In this section we discuss the outcome of a simulation in which the effectiveness of the Algorithms A0 and A1 is tested. Starting with an initial random graph we generate a sequence of requests in the following way. We use a queue called the event queue, storing events of two possible kinds: request and release. Associated with every event there is an event time denoting the time at which the event is triggered (processed). At any instant of time the event queue is sorted by event time so that the next event to be processed always appears at the front end of the queue. Each request event has information about the source node, the destination node and the event time, while a release event contains information about the source node, the destination node, the route and wavelength of the lightpath, and the event time. Initially the event queue has one request event for each node, with the event time equal to 0. Thereafter events are processed according to the following algorithm: 1. Remove the event from the front end of event queue. Call it e. 2. if e is a request then (a) set current time equal to the event time of e; (b) Let se and de be the source and destination nodes of e, respectively; (c) Generate a new request event e with source se as follows. Generate an inter-arrival time te using an exponential distribution with mean 1/r (r a parameter) and set the event time of e equal to current time plus te . Generate the destination node de , choosing it uniformly at random among all vertices except se . Insert e in the event queue in the sorted order based on the event time; (d) Process e (see below);

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

119

(e) if request e is honored then generate a new release event e with source se as follows: (a) Generate a holding time te using an exponential distribution with mean 1/r (r a parameter); (b) Choose a destination d uniformly at random among all nodes except se ; (c) Set the event time of e equal to the current time plus te . Insert e in the event queue in the sorted order based on the event time. 3. else (e is a release event) release the lightpath associated with e by updating the status information of the edges of the path. 4. repeat the above steps m times (m = 105 , a parameter). We now specify how a request is processed. When a request arrives, it is first checked whether a free source-destination path is available. In case there is no such path, Algorithm A0 (respectively, A1 ) is used to look for a path intersecting a (hopefully) small number of already laid-out paths. A path is allowed to be intersected only if it can be migrated to another wavelength. Prior to establishing a new path, the intersecting paths are migrated. If neither a free nor an available laid-out path is found, the request is not honored and rejected. We measure the following relevant performance parameters of Algorithms A0 and A1 : • The blocking probability, that is the fraction of requests which fail to be honored, plotted against the arrival rate of the requests. This is perhaps the most important performance parameter; • The average number of intersected paths, plotted against the request arrival rate; • The average number of hops, i.e., the average path length of the requests that are honored, plotted against the arrival rate; • The average connection request processing time, measured in CPU ticks, again plotted against the arrival rate of the requests. A WDM network with W wavelengths is represented as a layered graph each corresponding to a wavelength, called a wavelength graph. All the laid-out paths on a wavelength graph are arc-disjoint. As said, the performance of the algorithms is evaluated on randomly generated networks. A randomly generated network is characterized by N and p, where N is the number of nodes and p is the edge probability, i.e., for any two given nodes u and v, a “coin” is tossed and if the outcome is head, an event which has probability p of happening, the edge is included in the network. A small value of p typically generates a sparsely connected network, while if p is large the network is likely to be densely connected. In our experiments N = 32 and p = 0.1, 0.2. For each

120

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

Fig. 4. Blocking probability vs. arrival rate per node for a 50-node random graph with p = 0.10.

Fig. 5. Average number of intersected paths vs. arrival rate per node for a 50-node random graph with p = 0.10.

experiment we have generated 50 graphs, for each value of p, and for each graph we have generated in the order of m = 105 connection requests. Figs. 4–7 plot the performance of the algorithms for random graphs with N = 32 and p = 0.10. The blocking probability by different algorithms is plotted in Fig. 4 as a function of arrival rate per node. The arrival rate is measured in Erlangs. We consider three algorithms: no-intersection, A0 , and A1 . The no-intersection algorithm does not allow a new path to intersect with any of the laid-out paths.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

121

Fig. 6. Average number of hops per intersected path vs. arrival rate per node for a 50-node random graph with p = 0.10.

Fig. 7. Average connection request processing time in CPU ticks for a 50-node random graph with p = 0.10.

It can observed that the performance gain by A0 and A1 is significant. The figure shows that more than 20% of the connections rejected by no-intersection algorithm have been successfully routed by A0 and A1 . As the load increases, the above percentage of improvement slowly decreases. This is because, under heavy load, the number of laid-out paths that qualify for migration is small.

122

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

Fig. 5 plots the average number of intersecting paths by a new path by A0 and A1 for different arrival rates per node. The plots show that A0 performs better than A1 . This is because, A1 fails to identify that two edges belong to the same path. However, for A0 it is possible by way of crossover edges (shortcut arcs). It also shows that arrival rate does not have much effect on the number of intersecting paths. Fig. 6 plots the average number of hops used by an intersecting path by A0 and A1 for different arrival rates per node. The curves show that A1 selects shorter hop paths than A0 . The reason is that A1 performs an exhaustive search. The plot also shows that as the arrival rate increases the hop count decreases. This is because, under heavy load conditions, longer paths do not qualify for intersection, as their migration might not be possible. The actual running time to find a path is measured in terms of CPU ticks (1 CPU tick = 10 milliseconds). The time required by A0 and A1 for different arrival rates per node are plotted in Fig. 7. Algorithm A1 requires more time than A0 , since it considers several possible solutions before choosing the best one. Figs. 8–11 plot the performance of the algorithms for a random graph G2 with N = 50 and p = 0.20. The connectivity of G2 is denser than that of G1 . The plots demonstrate the usefulness of the Algorithms A0 and A1 . Due to dense connectivity, the performance on G2 in terms of blocking probability is better than G1 . For the same reason, the average number of intersected paths is slightly more and the average number of hops per intersected path is slightly less for G2

Fig. 8. Blocking probability vs. arrival rate per node for a 50-node random graph with p = 0.20.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

123

Fig. 9. Average number of intersected paths vs. arrival rate per node for a 50-node random graph with p = 0.20.

Fig. 10. Average number of hops per intersected path vs. arrival rate per node for a 50-node random graph with p = 0.20.

124

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

Fig. 11. Average connection request processing time in CPU ticks for a 50-node random graph with p = 0.20.

than that for G1 . Also, the average connection processing time for G2 is less than that for G1 . Thus we see that with a modest amount of computational effort, our algorithms can lead to better resource utilization in WDM networks, via wavelength rerouting.

Acknowledgments We thank Sanjeev Arora, Sanjeev Khanna, Goran Konjevod, Madhav Marathe, Riccardo Silvestri, and Luca Trevisan for helpful discussions. Part of this work was done at the University Ca’ Foscari, Venice.

References [1] S. Arora, L. Babai, J. Stern, Z. Sweedyk, The hardness of approximate optima in lattices, codes, and systems of linear equations, J. Comput. System Sci. 54 (1997) 317–331. [2] M. Bellare, J. Rompel, Randomness-efficient oblivious sampling, in: Proc. IEEE Symposium on Foundations of Computer Science, 1994, pp. 276–287. [3] R.D. Carr, S. Doddi, G. Konjevod, M. Marathe, On the red–blue set cover problem, in: Proc. ACM–SIAM Symposium on Discrete Algorithms, 2000, pp. 345–353.

A. Caprara et al. / Journal of Algorithms 45 (2002) 93–125

125

[4] I. Chlamtac, A. Ganz, G. Karmi, Lightpath communications: An approach to high bandwidth optical WANs, IEEE Trans. Commun. 40 (1992) 1171–1182. [5] P. Crescenzi, R. Silvestri, L. Trevisan, To weight or not to weight: Where is the question?, in: Proc. 4th Israel Symposium on Theory of Computing and Systems, 1996, pp. 68–77. [6] Y. Dodis, S. Khanna, Designing networks with bounded pairwise distance, in: Proc. ACM Symposium on Theory of Computing, 1999, pp. 750–759. [7] G. Even, O. Goldreich, M. Luby, N. Nisan, B. Veliˇckovi´c, Efficient approximations for product distributions, Random Structures Algorithms 13 (1998) 1–16. [8] U. Feige, J. Kilian, Zero knowledge and the chromatic number, J. Comput. System Sci. 57 (1998) 187–199. [9] M.R. Garey, D.S. Johnson, Computers and Intractability, Freeman, San Francisco, 1979. [10] J. Håstad, Clique is hard to approximate within n1−ε , Acta Math., to appear. [11] J. Håstad, Some optimal inapproximability results, in: Proc. ACM Symposium on Theory of Computing, 1997, pp. 1–10. [12] I. Holyer, The NP-completeness of edge colouring, SIAM J. Comput. 10 (1981) 718–720. [13] S. Khanna, M. Sudan, L. Trevisan, Constraint satisfaction: The approximability of minimization problems, ECCC tech. report TR96-064, available at http://www.eccc.uni-trier.de/eccc-local/ Lists/TR-1996.html. Also see in: Proc. IEEE Conference on Computational Complexity, 1997, pp. 282–296. [14] S.O. Krumke, H.-C. Wirth, On the minimum label spanning tree problem, Inform. Process. Lett. 66 (1998) 81–85. [15] K.C. Lee, V.O.K. Li, A wavelength rerouting algorithm in wide-area all-optical networks, IEEE/OSA J. Lightwave Technol. 14 (1996) 1218–1229. [16] G. Mohan, C.S.R. Murthy, Efficient algorithms for wavelength rerouting in WDM multi-fiber unidirectional ring networks, Comput. Commun. 22 (1999) 232–243. [17] G. Mohan, C.S.R. Murthy, A time optimal wavelength rerouting for dynamic traffic in WDM networks, IEEE/OSA J. Lightwave Technol. 17 (1999) 406–417. [18] R. Motwani, P. Raghavan, Randomized Algorithms, Cambridge University Press, 1995. [19] C. Papadimitriou, M. Yannakakis, Optimization, approximation, and complexity classes, J. Comput. System Sci. 43 (3) 425–440. [20] L.J. Stockmeyer, Planar 3-colorability is NP-complete, SIGACT News 5 (1973) 19–25.