Fully polynomial-time approximation scheme

From HandWiki
Revision as of 23:31, 6 February 2024 by Jslovo (talk | contribs) (update)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

A fully polynomial-time approximation scheme (FPTAS) is an algorithm for finding approximate solutions to function problems, especially optimization problems. An FPTAS takes as input an instance of the problem and a parameter ε > 0. It returns as output a value which is at least [math]\displaystyle{ 1-\varepsilon }[/math] times the correct value, and at most [math]\displaystyle{ 1 + \varepsilon }[/math] times the correct value. In the context of optimization problems, the correct value is understood to be the value of the optimal solution, and it is often implied that an FPTAS should produce a valid solution (and not just the value of the solution). Returning a value and finding a solution with that value are equivalent assuming that the problem possesses self reducibility.

Importantly, the run-time of an FPTAS is polynomial in the problem size and in 1/ε. This is in contrast to a general polynomial-time approximation scheme (PTAS). The run-time of a general PTAS is polynomial in the problem size for each specific ε, but might be exponential in 1/ε.[1]

The term FPTAS may also be used to refer to the class of problems that have an FPTAS. FPTAS is a subset of PTAS, and unless P = NP, it is a strict subset.[2]

Relation to other complexity classes

All problems in FPTAS are fixed-parameter tractable with respect to the standard parameterization.[3]

Any strongly NP-hard optimization problem with a polynomially bounded objective function cannot have an FPTAS unless P=NP.[4] However, the converse fails: e.g. if P does not equal NP, knapsack with two constraints is not strongly NP-hard, but has no FPTAS even when the optimal objective is polynomially bounded.[5]

Converting a dynamic program to an FPTAS

Woeginger[6] presented a general scheme for converting a certain class of dynamic programs to an FPTAS.

Input

The scheme handles optimization problems in which the input is defined as follows:

  • The input is made of n vectors, x1,...,xn.
  • Each input vector is made of some [math]\displaystyle{ a }[/math] non-negative integers, where [math]\displaystyle{ a }[/math] may depend on the input.
  • All components of the input vectors are encoded in binary. So the size of the problem is O(n+log(X)), where X is the sum of all components in all vectors.

Extremely-simple dynamic program

It is assumed that the problem has a dynamic-programming (DP) algorithm using states. Each state is a vector made of some [math]\displaystyle{ b }[/math] non-negative integers, where [math]\displaystyle{ b }[/math] is independent of the input. The DP works in n steps. At each step i, it processes the input xi, and constructs a set of states Si. Each state encodes a partial solution to the problem, using inputs x1,...,xi. The components of the DP are:

  • A set S0 of initial states.
  • A set F of transition functions. Each function f in F maps a pair (state,input) to a new state.
  • An objective function g, mapping a state to its value.

The algorithm of the DP is:

  • Let S0 := the set of initial states.
  • For k = 1 to n do:
    • Let Sk := {f(s,xk) | f in F, s in Sk−1}
  • Output min/max {g(s) | s in Sn}.

The run-time of the DP is linear in the number of possible states. In general, this number can be exponential in the size of the input problem: it can be in O(n Vb), where V is the largest integer than can appear in a state. If V is in O(X), then the run-time is in O(n Xb), which is only pseudo-polynomial time, since it is exponential in the problem size which is in O(log X).

The way to make it polynomial is to trim the state-space: instead of keeping all possible states in each step, keep only a subset of the states; remove states that are "sufficiently close" to other states. Under certain conditions, this trimming can be done in a way that does not change the objective value by too much.

To formalize this, we assume that the problem at hand has a non-negative integer vector d = (d1,...,db), called the degree vector of the problem. For every real number r>1, we say that two state-vectors s1,s2 are (d,r)-close if, for each coordinate j in 1,...,b: [math]\displaystyle{ r^{-d_j} \cdot s_{1,j} \leq s_{2,j} \leq r^{d_j} \cdot s_{1,j} }[/math] (in particular, if dj=0 for some j, then [math]\displaystyle{ s_{1,j} = s_{2,j} }[/math]).

A problem is called extremely-benevolent if it satisfies the following three conditions:

  1. Proximity is preserved by the transition functions: For any r>1, for any transition function f in F, for any input-vector x, and for any two state-vectors s1,s2, the following holds: if s1 is (d,r)-close to s2, then f(s1,x) is (d,r)-close to f(s2,x).
    • A sufficient condition for this can be checked as follows. For every function f(s,x) in F, and for every coordinate j in 1,...,b, denote by fj(s,x) the j-th coordinate of f. This fj can be seen as an integer function in b+a variables. Suppose that every such fj is a polynomial with non-negative coefficients. Convert it to a polynomial of a single variable z, by substituting s=(zd1,...,zdb) and x=(1,...,1). If the degree of the resulting polynomial in z is at most dj, then condition 1 is satisfied.
  2. Proximity is preserved by the value function: There exists an integer G ≥ 0 (which is a function of the value function g and the degree vector d), such that for any r>1, and for any two state-vectors s1,s2, the following holds: if s1 is (d,r)-close to s2, then: g(s1) ≤ rG · g(s2) (in minimization problems); g(s1) ≥ r(-G) · g(s2) (in maximization problems).
    • A sufficient condition for this is that the function g is a polynomial function (of b variables) with non-negative coefficients.
  3. Technical conditions:
    • All transition functions f in F and the value function g can be evaluated in polytime.
    • The number |F| of transition functions is polynomial in n and log(X).
    • The set S0 of initial states can be computed in time polynomial in n and log(X).
    • Let Vj be the set of all values that can appear in coordinate j in a state. Then, the ln of every value in Vj is at most a polynomial P1(n,log(X)).
    • If dj=0, the cardinality of Vj is at most a polynomial P2(n,log(X)).

For every extremely-benevolent problem, the dynamic program can be converted into an FPTAS. Define:

  • [math]\displaystyle{ \epsilon }[/math] := the required approximation ratio.
  • [math]\displaystyle{ r := 1 + \frac{\epsilon}{2 G n} }[/math], where G is the constant from condition 2. Note that [math]\displaystyle{ \frac{1}{\ln{r}} \leq 1 + \frac{2Gn}{\epsilon} }[/math].
  • [math]\displaystyle{ L := \left\lceil \frac{P_1(n,\log(X))}{\ln(r)} \right\rceil }[/math], where P1 is the polynomial from condition 3 (an upper bound on the ln of every value that can appear in a state vector). Note that [math]\displaystyle{ L\leq \left\lceil \left(1 + \frac{2 G n}{\epsilon}\right) P_1(n, \log{X}) \right\rceil }[/math], so it is polynomial in the size of the input and in [math]\displaystyle{ 1/\epsilon }[/math]. Also, [math]\displaystyle{ r^L = e^{\ln{r}}\cdot L \geq e^{P_1(n,\log{x})} }[/math], so by definition of P1, every integer that can appear in a state-vector is in the range [0,rL].
  • Partition the range [0,rL] into L+1 r-intervals: [math]\displaystyle{ I_0 = [0]; I_1 = [1,r); I_2 = [r,r^2); \ldots ; I_L = [r^{L-1}, r^L] }[/math].
  • Partition the state space into r-boxes: each coordinate k with degree dk ≥ 1 is partitioned into the L+1 intervals above; each coordinate with dk = 0 is partitioned into P2(n,log(X)) singleton intervals - an interval for each possible value of coordinate k (where P2 is the polynomial from condition 3 above).
    • Note that every possible state is contained in exactly one r-box; if two states are in the same r-box, then they are (d,r)-close.
  • [math]\displaystyle{ R := (L + 1 + P_2(n,\log{X}))^b }[/math].
    • Note that the number of r-boxes is at most R. Since b is a fixed constant, this R is polynomial in the size of the input and in [math]\displaystyle{ 1/\epsilon }[/math].

The FPTAS runs similarly to the DP, but in each step, it trims the state set into a smaller set Tk, that contains exactly one state in each r-box. The algorithm of the FPTAS is:

  • Let T0 := S0 = the set of initial states.
  • For k = 1 to n do:
    • Let Uk := {f(s,xk) | f in F, s in Tk−1}
    • Let Tk := a trimmed copy of Uk: for each r-box that contains one or more states of Uk, keep exactly one state in Tk.
  • Output min/max {g(s) | s in Tn}.

The run-time of the FPTAS is polynomial in the total number of possible states in each Ti, which is at most the total number of r-boxes, which is at most R, which is polynomial in n, log(X), and [math]\displaystyle{ 1/\epsilon }[/math].

Note that, for each state su in Uk, its subset Tk contains at least one state st that is (d,r)-close to su. Also, each Uk is a subset of the Sk in the original (untrimmed) DP. The main lemma for proving the correctness of the FPTAS is:[6]:{{{1}}}

For every step k in 0,...,n, for every state ss in Sk, there is a state st in Tk that is (d,rk)-close to ss.

The proof is by induction on k. For k=0 we have Tk=Sk; every state is (d,1)-close to itself. Suppose the lemma holds for k-1. For every state ss in Sk, let ss- be one of its predecessors in Sk-1, so that f(ss,x)=ss. By the induction assumption, there is a state st- in Tk-1, that is (d,rk-1)-close to ss. Since proximity is preserved by transitions (Condition 1 above), f(st,x) is (d,rk-1)-close to f(ss,x)=ss. This f(st,x) is in Uk. After the trimming, there is a state st in Tk that is (d,r)-close to f(st-,x). This st is (d,rk)-close to ss.

Consider now the state s* in Sn, which corresponds to the optimal solution (that is, g(s*)=OPT). By the lemma above, there is a state t* in Tn, which is (d,rn)-close to s*. Since proximity is preserved by the value function, g(t*) ≥ r(-Gn) · g(s*) for a maximization problem. By definition of r, [math]\displaystyle{ r^{-Gn}\geq (1-\epsilon) }[/math]. So [math]\displaystyle{ g(t^*)\geq (1-\epsilon)\cdot OPT }[/math]. A similar argument works for a minimization problem.

Examples

Here are some examples of extremely-benevolent problems, that have an FPTAS by the above theorem.[6]

1. Multiway number partitioning (equivalently, Identical-machines scheduling) with the goal of minimizing the largest sum is extremely-benevolent. Here, we have a = 1 (the inputs are integers) and b = the number of bins (which is considered fixed). Each state is a vector of b integers representing the sums of the b bins. There are b functions: each function j represents inserting the next input into bin j. The function g(s) picks the largest element of s. S0 = {(0,...,0)}. The conditions for extreme-benevolence are satisfied with degree-vector d=(1,...,1) and G=1. The result extends to Uniform-machines scheduling and Unrelated-machines scheduling whenever the number of machines is fixed (this is required because R - the number of r-boxes - is exponential in b). Denoted Pm||[math]\displaystyle{ \max C_j }[/math] or Qm||[math]\displaystyle{ \max C_j }[/math] or Rm||[math]\displaystyle{ \max C_j }[/math].

  • Note: consider the special case b=2, where the goal is to minimize the square of the difference between the two part sums. The same DP can be used, but this time with value function g(s) = (s1-s2)2. Now, condition 2 is violated: the states (s1,s1) and (s1,s2) may be (d,r)-close, but g(s1,s1) = 0 while g(s1,s2) > 0. so the above theorem cannot be applied. Indeed, the problem does not have an FPTAS unless P=NP, since an FPTAS could be used to decide in polytime whether the optimal value is 0.

2. Sum of cubed job completion time on any fixed number of identical or uniform machines - the latter denoted by Qm||[math]\displaystyle{ \sum C_j^3 }[/math] - is ex-benevolent with a=1, b=3, d=(1,1,3). It can be extended to any fixed power of the completion time.

3. Sum of weighted completion time on any fixed number of identical or uniform machines - the latter denoted by Qm||[math]\displaystyle{ \sum w_j C_j }[/math].

4. Sum of completion time on any fixed number of identical or uniform machines, with time-dependent processing times: Qm|time-dep|[math]\displaystyle{ \sum C_j }[/math]. This holds even for weighted sum of completion time.

5. Weighted earliness-tardiness about a common due-date on any fixed number of machines: m||[math]\displaystyle{ \sum w_j|C_j| }[/math].

Simple dynamic program

Simple dynamic programs add to the above formulation the following components:

  • A set H of filtering functions, of the same cardinality as F. Each function hi in H maps a pair (state,input) to a Boolean value. The value should be "true" if and only if activating the transition fi on this pair would lead to a valid state.
  • A dominance relation, which is a partial order on states (no indifferences, not all pairs are comparable), and a quasi-dominance relation which is a total preorder on states (indifferences allowed, all pairs are comparable).

The original DP is modified as follows:

  • Let S0 := the set of initial states.
  • For k = 1 to n do:
    • Let Sk := {fj(s,xk) | fj in F, s in Sk−1, hj(s,xk)=True }, where hj is the filter function corresponding to the transition function fj.
  • Output min/max {g(s) | s in Sn}.

A problem is called benevolent if it satisfies the following conditions (which extend conditions 1, 2, 3 above):

  1. Proximity is preserved by the transition functions: For any r>1, for any transition function f in F, for any input-vector x, and for any two state-vectors s1,s2, the following holds:
    • if s1 is (d,r)-close to s2, and s1 quasi-dominates s2, then either (a) f(s1,x) is (d,r)-close to f(s2,x), and f(s1,x) quasi-dominates f(s2,x), or (b) f(s1,x) dominates f(s2,x).
    • if s1 dominates s2, then f(s1,x) dominates f(s2,x).
  2. Proximity is preserved by the value function: There exists an integer G ≥ 0 (a function of the value function g and the degree vector d), such that for any r>1, and for any two state-vectors s1,s2, the following holds:
    • if s1 is (d,r)-close to s2, and s1 quasi-dominates s2, then: g(s1) ≤ rG · g(s2) (in minimization problems); g(s1) ≥ r(-G) · g(s2) (in maximization problems).
    • if s1 dominates s2, then g(s1) ≤ g(s2) (in minimization problems); g(s1) ≥ g(s2) (in maximization problems).
  3. Technical conditions (in addition to the above):
    • The quasi-dominance relation can be decided in polynomial time.
  4. Conditions on the filter functions: For any r>1, for any filter function h in H, for any input-vector x, and for any two state-vectors s1,s2, the following holds:
    • if s1 is (d,r)-close to s2, and s1 quasi-dominates s2, then h(s1,x) ≥ h(s2,x).
    • if s1 dominates s2, then h(s1,x) ≥ h(s2,x).

For every benevolent problem, the dynamic program can be converted into an FPTAS similarly to the one above, with two changes (boldfaced):

  • Let T0 := S0 = the set of initial states.
  • For k = 1 to n do:
    • Let Uk := {fj(s,xk) | fj in F, s in Tk−1, hj(s,xk)=True }, where hj is the filter function corresponding to the transition function fj.
    • Let Tk := a trimmed copy of Uk: for each r-box that contains one or more states of Uk, choose a single element that quasi-dominates all other elements in Uk, and insert it into Tk.
  • Output min/max {g(s) | s in Tn}.

Examples

Here are some examples of benevolent problems, that have an FPTAS by the above theorem.[6]

1. The 0-1 knapsack problem is benevolent. Here, we have a=2: each input is a 2-vector (weight, value). There is a DP with b=2: each state encodes (current weight, current value). There are two transition functions: f1 corresponds to adding the next input item, and f2 corresponds to not adding it. The corresponding filter functions are: h1 verifies that the weight with the next input item is at most the knapsack capacity; h2 always returns True. The value function g(s) returns s2. The initial state-set is {(0,0)}. The degree vector is (1,1). The dominance relation is trivial. The quasi-dominance relation compares only the weight coordinate: s quasi-dominates t iff s1t1. The implication of this is that, if state t has a higher weight than state s, then the transition functions are allowed to not preserve the proximity between t and s (it is possible, for example, that s has a successor and t does not have a corresponding successor). A similar algorithm was presented earlier by Ibarra and Kim.[7] The run-time of this FPTAS can be improved to [math]\displaystyle{ O(n \log{1/\epsilon} + 1/\epsilon^4) }[/math] operations on integers.[8] The exponent was later improved to 2.5.[9]

  • Note: consider In the 2-weighted knapsack problem, where each item has two weights and a value, and the goal is to maximize the value such that the sum of squares of the total weights is at most the knapsack capacity: [math]\displaystyle{ \left(\sum_{k\in K} w_{1,k}\right)^2 + \left(\sum_{k\in K} w_{2,k}\right)^2 \leq W }[/math]. We could solve it using a similar DP, where each state is (current weight 1, current weight 2, value). The quasi-dominance relation should be modified to: s quasi-dominates t iff (s12 + s22) ≤ (t12 + t22). But it violates Condition 1 above: quasi-dominance is not preserved by transition functions [for example, the state (2,2,..) quasi-dominates (1,3,..); but after adding the input (2,0,..) to both states, the result (4,2,..) does not quasi-dominate (3,3,..)]. So the theorem cannot be used. Indeed, this problem does not have an FPTAS unless P=NP. The same is true for the two-dimensional knapsack problem. The same is true for the multiple subset sum problem: the quasi-dominance relation should be: s quasi-dominates t iff max(s1,s2) ≤ max(t1,t2), but it is not preserved by transitions, by the same example as above.

2. Minimizing the weighted number of tardy jobs, or maximizing the weighted number of early jobs, on a single machine; denoted 1||[math]\displaystyle{ \sum w_j U_j }[/math].

3. Batch scheduling for minimizing the weighted number of tardy jobs: 1|batch|[math]\displaystyle{ \sum w_j U_j }[/math].

4. Makespan of deteriorating jobs on a single machine: 1|deteriorate|[math]\displaystyle{ \max C_j }[/math].

5. Total late work on a single machine: 1||[math]\displaystyle{ \sum V_j }[/math].

6. Total weighted late work on a single machine: 1||[math]\displaystyle{ \sum w_j V_j }[/math].

Non-examples

Despite the generality of the above result, there are cases in which it cannot be used.

1. In the total tardiness problem 1||[math]\displaystyle{ \sum T_j }[/math], the dynamic programming formulation of Lawler[10] requires to update all states in the old state space some B times, where B is of the order of X (the maximum input size). The same is true for a DP for economic lot-sizing.[11] In these cases, the number of transition functions in F is B, which is exponential in the log(X), so the second technical condition is violated. The state-trimming technique is not useful, but another technique - input-rounding - has been used to design an FPTAS.[12][13]

2. In the variance minimization problem 1||[math]\displaystyle{ CTV }[/math], the objective function is [math]\displaystyle{ g(s) = s_5 - (s_4-s_3)^2/n }[/math], which violates Condition 2, so the theorem cannot be used. But different techniques have been used to design an FPTAS.[14][15]

FPTAS for approximating real numbers

A different kind of problems in which FPTAS may be useful is finding rational numbers that approximate some real numbers. For example, consider the infinite series [math]\displaystyle{ \sum_{i=1}^\infty \frac{1}{i^3} }[/math]. The sum is an irrational number. To approximate it by a rational number, we can compute the sum of the first k elements, for some finite k. One can show that the error in approximation is about [math]\displaystyle{ \frac{1}{2 k^2} }[/math]. Therefore, to get an error of ε, we need about [math]\displaystyle{ \sqrt{\frac{1}{2 \epsilon}} }[/math] elements, so this is an FPTAS. Note that this particular sum can be represented by another sum in which only O(log(ε)) elements are needed, so the sum can actually be approximated in polynomial time in the encoding length of ε.[16](p35)

Some other problems that have an FPTAS

  • The knapsack problem,[17][18] as well as some of its variants:
    • 0-1 knapsack problem.[19]
    • Unbounded knapsack problem.[20]
    • Multi-dimensional knapsack problem with Delta-modular constraints.[21]
    • Multi-objective 0-1 knapsack problem.[22]
    • Parametric knapsack problem.[23]
    • Symmetric quadratic knapsack problem.[24]
  • Count-subset-sum (#SubsetSum) - finding the number of distinct subsets with a sum of at most C.[25]
  • Restricted shortest path: finding a minimum-cost path between two nodes in a graph, subject to a delay constraint.[26]
  • Shortest paths and non-linear objectives.[27]
  • Counting edge-covers.[28]
  • Vector subset search problem where the dimension is fixed.[29]

See also

  • The "benevolent dynamic programs", that admit an FPTAS, also admit an evolutionary algorithm.[30]

References

  1. G. Ausiello, P. Crescenzi, G. Gambosi, V. Kann, A. Marchetti-Spaccamela, and M. Protasi. Complexity and Approximation: Combinatorial optimization problems and their approximability properties, Springer-Verlag, 1999.
  2. Jansen, Thomas (1998), "Introduction to the Theory of Complexity and Approximation Algorithms", in Mayr, Ernst W.; Prömel, Hans Jürgen; Steger, Angelika, Lectures on Proof Verification and Approximation Algorithms, Springer, pp. 5–28, doi:10.1007/BFb0053011, ISBN 9783540642015 . See discussion following Definition 1.30 on p. 20.
  3. Cai, Liming; Chen, Jianer (June 1997). "On Fixed-Parameter Tractability and Approximability of NP Optimization Problems" (in en). Journal of Computer and System Sciences 54 (3): 465–474. doi:10.1006/jcss.1997.1490. 
  4. Vazirani, Vijay V. (2003). Approximation Algorithms. Berlin: Springer. Corollary 8.6. ISBN 3-540-65367-8. 
  5. H. Kellerer; U. Pferschy; D. Pisinger (2004). Knapsack Problems. Springer. Theorem 9.4.1. 
  6. 6.0 6.1 6.2 6.3 Woeginger, Gerhard J. (2000-02-01). "When Does a Dynamic Programming Formulation Guarantee the Existence of a Fully Polynomial Time Approximation Scheme (FPTAS)?". INFORMS Journal on Computing 12 (1): 57–74. doi:10.1287/ijoc.12.1.57.11901. ISSN 1091-9856. https://pubsonline.informs.org/doi/abs/10.1287/ijoc.12.1.57.11901. 
  7. Ibarra, Oscar H.; Kim, Chul E. (1975-10-01). "Fast Approximation Algorithms for the Knapsack and Sum of Subset Problems". Journal of the ACM 22 (4): 463–468. doi:10.1145/321906.321909. ISSN 0004-5411. https://doi.org/10.1145/321906.321909. 
  8. Lawler, Eugene L. (1979-11-01). "Fast Approximation Algorithms for Knapsack Problems". Mathematics of Operations Research 4 (4): 339–356. doi:10.1287/moor.4.4.339. ISSN 0364-765X. https://pubsonline.informs.org/doi/abs/10.1287/moor.4.4.339. 
  9. Rhee, Donguk (2015). Faster fully polynomial approximation schemes for Knapsack problems (Thesis thesis). Massachusetts Institute of Technology. hdl:1721.1/98564.
  10. Lawler, Eugene L. (1977-01-01), Hammer, P. L.; Johnson, E. L.; Korte, B. H. et al., eds., "A "Pseudopolynomial" Algorithm for Sequencing Jobs to Minimize Total Tardiness**Research supported by National Science Foundation Grant GJ-43227X" (in en), Annals of Discrete Mathematics, Studies in Integer Programming (Elsevier) 1: pp. 331–342, doi:10.1016/S0167-5060(08)70742-8, https://www.sciencedirect.com/science/article/pii/S0167506008707428, retrieved 2021-12-17 
  11. Florian, M.; Lenstra, J. K.; Rinnooy Kan, A. H. G. (1980-07-01). "Deterministic Production Planning: Algorithms and Complexity". Management Science 26 (7): 669–679. doi:10.1287/mnsc.26.7.669. ISSN 0025-1909. https://pubsonline.informs.org/doi/abs/10.1287/mnsc.26.7.669. 
  12. Lawler, E. L. (1982-12-01). "A fully polynomial approximation scheme for the total tardiness problem" (in en). Operations Research Letters 1 (6): 207–208. doi:10.1016/0167-6377(82)90022-0. ISSN 0167-6377. https://dx.doi.org/10.1016/0167-6377%2882%2990022-0. 
  13. van Hoesel, C. P. M.; Wagelmans, A. P. M. (2001). "Fully Polynomial Approximation Schemes for Single-Item Capacitated Economic Lot-Sizing Problems". Mathematics of Operations Research 26 (2): 339–357. doi:10.1287/moor.26.2.339.10552. https://repub.eur.nl/pub/1406/. 
  14. Cai, X. (1995-09-21). "Minimization of agreeably weighted variance in single machine systems" (in en). European Journal of Operational Research 85 (3): 576–592. doi:10.1016/0377-2217(93)E0367-7. ISSN 0377-2217. https://dx.doi.org/10.1016/0377-2217%2893%29E0367-7. 
  15. Woeginger, Gerhard J. (1999-05-01). "An Approximation Scheme for Minimizing Agreeably Weighted Variance on a Single Machine". INFORMS Journal on Computing 11 (2): 211–216. doi:10.1287/ijoc.11.2.211. ISSN 1091-9856. https://pubsonline.informs.org/doi/abs/10.1287/ijoc.11.2.211. 
  16. Template:Cite Geometric Algorithms and Combinatorial Optimization
  17. Vazirani, Vijay (2001). Approximation algorithms. Berlin: Springer. pp. 69–70. ISBN 3540653678. OCLC 47097680. https://archive.org/details/approximationalg00vazi_328. 
  18. Kellerer, Hans; Pferschy, Ulrich (2004-03-01). "Improved Dynamic Programming in Connection with an FPTAS for the Knapsack Problem" (in en). Journal of Combinatorial Optimization 8 (1): 5–11. doi:10.1023/B:JOCO.0000021934.29833.6b. ISSN 1573-2886. https://doi.org/10.1023/B:JOCO.0000021934.29833.6b. 
  19. Jin, Ce (2019). An Improved FPTAS for 0-1 Knapsack. Leibniz International Proceedings in Informatics (LIPIcs). 132. pp. 76:1–76:14. doi:10.4230/LIPIcs.ICALP.2019.76. ISBN 9783959771092. 
  20. Jansen, Klaus; Kraft, Stefan E. J. (2018-02-01). "A faster FPTAS for the Unbounded Knapsack Problem" (in en). European Journal of Combinatorics. Combinatorial Algorithms, Dedicated to the Memory of Mirka Miller 68: 148–174. doi:10.1016/j.ejc.2017.07.016. ISSN 0195-6698. https://www.sciencedirect.com/science/article/pii/S0195669817301245. 
  21. Gribanov, D. V. (2021-05-10). "An FPTAS for the $$\var Delta $$-Modular Multidimensional Knapsack Problem". Mathematical Optimization Theory and Operations Research. Lecture Notes in Computer Science. 12755. pp. 79–95. doi:10.1007/978-3-030-77876-7_6. ISBN 978-3-030-77875-0. 
  22. Bazgan, Cristina; Hugot, Hadrien; Vanderpooten, Daniel (2009-10-01). "Implementing an efficient fptas for the 0–1 multi-objective knapsack problem" (in en). European Journal of Operational Research 198 (1): 47–56. doi:10.1016/j.ejor.2008.07.047. ISSN 0377-2217. https://www.sciencedirect.com/science/article/pii/S0377221708006905. 
  23. Holzhauser, Michael; Krumke, Sven O. (2017-10-01). "An FPTAS for the parametric knapsack problem" (in en). Information Processing Letters 126: 43–47. doi:10.1016/j.ipl.2017.06.006. ISSN 0020-0190. https://www.sciencedirect.com/science/article/pii/S0020019017301072. 
  24. Xu, Zhou (2012-04-16). "A strongly polynomial FPTAS for the symmetric quadratic knapsack problem" (in en). European Journal of Operational Research 218 (2): 377–381. doi:10.1016/j.ejor.2011.10.049. ISSN 0377-2217. https://www.sciencedirect.com/science/article/pii/S0377221711010022. 
  25. Gopalan, Parikshit; Klivans, Adam; Meka, Raghu; Štefankovic, Daniel; Vempala, Santosh; Vigoda, Eric (2011-10-01). "An FPTAS for #Knapsack and Related Counting Problems". 2011 IEEE 52nd Annual Symposium on Foundations of Computer Science. pp. 817–826. doi:10.1109/FOCS.2011.32. ISBN 978-0-7695-4571-4. https://ieeexplore.ieee.org/abstract/document/6108252?. 
  26. Ergun, Funda; Sinha, Rakesh; Zhang, Lisa (2002-09-15). "An improved FPTAS for Restricted Shortest Path" (in en). Information Processing Letters 83 (5): 287–291. doi:10.1016/S0020-0190(02)00205-3. ISSN 0020-0190. https://www.sciencedirect.com/science/article/pii/S0020019002002053. 
  27. Tsaggouris, George; Zaroliagis, Christos (2009-06-01). "Multiobjective Optimization: Improved FPTAS for Shortest Paths and Non-Linear Objectives with Applications" (in en). Theory of Computing Systems 45 (1): 162–186. doi:10.1007/s00224-007-9096-4. ISSN 1433-0490. https://doi.org/10.1007/s00224-007-9096-4. 
  28. Lin, Chengyu; Liu, Jingcheng; Lu, Pinyan (2013-12-18), "A Simple FPTAS for Counting Edge Covers", Proceedings of the 2014 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), Proceedings (Society for Industrial and Applied Mathematics): pp. 341–348, doi:10.1137/1.9781611973402.25, ISBN 978-1-61197-338-9, https://epubs.siam.org/doi/abs/10.1137/1.9781611973402.25, retrieved 2021-12-13 
  29. Kel’manov, A. V.; Romanchenko, S. M. (2014-07-01). "An FPTAS for a vector subset search problem" (in en). Journal of Applied and Industrial Mathematics 8 (3): 329–336. doi:10.1134/S1990478914030041. ISSN 1990-4797. https://doi.org/10.1134/S1990478914030041. 
  30. Doerr, Benjamin; Eremeev, Anton; Neumann, Frank; Theile, Madeleine; Thyssen, Christian (2011-10-07). "Evolutionary algorithms and dynamic programming" (in en). Theoretical Computer Science 412 (43): 6020–6035. doi:10.1016/j.tcs.2011.07.024. ISSN 0304-3975. 

External links