Linear-fractional programming explained

In mathematical optimization, linear-fractional programming (LFP) is a generalization of linear programming (LP). Whereas the objective function in a linear program is a linear function, the objective function in a linear-fractional program is a ratio of two linear functions. A linear program can be regarded as a special case of a linear-fractional program in which the denominator is the constant function 1.

Formally, a linear-fractional program is defined as the problem of maximizing (or minimizing) a ratio of affine functions over a polyhedron,

\begin{align} maximize&

cTx+\alpha
dTx+\beta

\\ subjectto&Ax\leqb, \end{align}

where

x\inRn

represents the vector of variables to be determined,

c,d\inRn

and

b\inRm

are vectors of (known) coefficients,

A\inRm

is a (known) matrix of coefficients and

\alpha,\beta\inR

are constants. The constraints have to restrict the feasible region to

\{x|dTx+\beta>0\}

, i.e. the region on which the denominator is positive.[1] [2] Alternatively, the denominator of the objective function has to be strictly negative in the entire feasible region.

Motivation by comparison to linear programming

Both linear programming and linear-fractional programming represent optimization problems using linear equations and linear inequalities, which for each problem-instance define a feasible set. Fractional linear programs have a richer set of objective functions. Informally, linear programming computes a policy delivering the best outcome, such as maximum profit or lowest cost. In contrast, a linear-fractional programming is used to achieve the highest ratio of outcome to cost, the ratio representing the highest efficiency. For example, in the context of LP we maximize the objective function profit = income - cost and might obtain maximum profit of $100 (= $1100 of income - $1000 of cost). Thus, in LP we have an efficiency of $100/$1000 = 0.1. Using LFP we might obtain an efficiency of $10/$50 = 0.2 with a profit of only $10, but only requiring $50 of investment.

Transformation to a linear program

Any linear-fractional program can be transformed into a linear program, assuming that the feasible region is non-empty and bounded, using the Charnes-Cooper transformation. The main idea is to introduce a new non-negative variable

t

to the program which will be used to rescale the constants involved in the program (

\alpha,\beta,b

). This allows us to require that the denominator of the objective function (

dTx+\beta

) equals 1. (To understand the transformation, it is instructive to consider the simpler special case with

\alpha=\beta=0

.)

Formally, the linear program obtained via the Charnes-Cooper transformation uses the transformed variables

y\inRn

and

t\ge0

:

\begin{align} maximize&cTy+\alphat\\ subjectto&Ay\leqbt\\ &dTy+\betat=1\\ &t\geq0. \end{align}

A solution

x

to the original linear-fractional program can be translated to a solution of the transformed linear program via the equalities

y=

1
dTx+\beta

xandt=

1
dTx+\beta

.

Conversely, a solution for

y

and

t

of the transformed linear program can be translated to a solution of the original linear-fractional program via
x=1
t

y.

Duality

Let the dual variables associated with the constraints

Ay-bt\leq0

and

dTy+\betat-1=0

be denoted by

u

and

λ

, respectively. Then the dual of the LFP above is [3] [4]

\begin{align} minimize&λ\\ subjectto&ATu+λd=c\\ &-bTu+λ\beta\geq\alpha\\ &u\in

m,
R
+

λ\inR, \end{align}

which is an LP and which coincides with the dual of the equivalent linear program resulting from the Charnes-Cooper transformation.

Properties and algorithms

The objective function in a linear-fractional problem is both quasiconcave and quasiconvex (hence quasilinear) with a monotone property, pseudoconvexity, which is a stronger property than quasiconvexity. A linear-fractional objective function is both pseudoconvex and pseudoconcave, hence pseudolinear. Since an LFP can be transformed to an LP, it can be solved using any LP solution method, such as the simplex algorithm (of George B. Dantzig),[5] [6] [7] the criss-cross algorithm,[8] or interior-point methods.

Notes

  1. Charnes . A. . Cooper . W. W. . William W. Cooper . 1962 . Programming with Linear Fractional Functionals . Naval Research Logistics Quarterly . 9 . 3–4 . 181–186 . 10.1002/nav.3800090303 . 152370.
  2. Book: Boyd . Stephen P. . Convex Optimization . Vandenberghe . Lieven . Cambridge University Press . 2004 . 978-0-521-83378-3 . 151 . October 15, 2011.
  3. Schaible . Siegfried . Parameter-free Convex Equivalent and Dual Programs. Zeitschrift für Operations Research . 18 . 1974 . 5 . 187–196. 10.1007/BF02026600. 351464. 28885670 .
  4. Fractional programming I: Duality . Schaible . Siegfried . Management Science . 22 . 8 . 858–867 . 1976. 2630017. 421679. 10.1287/mnsc.22.8.858.
  5. Chapter five: Book: Craven, B. D.. Fractional programming. Sigma Series in Applied Mathematics. 4. Heldermann Verlag. Berlin. 1988. 145. 978-3-88538-404-5 . 949209.
  6. Kruk . Serge. Wolkowicz. Henry. Pseudolinear programming . SIAM Review. 41 . 1999 . 4 . 795–805 . 1723002. 2653207. 10.1137/S0036144598335259. 10.1.1.53.7355.
  7. Mathis. Frank H.. Mathis. Lenora Jane. A nonlinear programming algorithm for hospital management . SIAM Review. 37 . 1995 . 2 . 230–234. 1343214. 2132826. 10.1137/1037046. 120626738 .
  8. The finite criss-cross method for hyperbolic programming. European Journal of Operational Research. 114. 1. 198–214. 1999 . 10.1016/S0377-2217(98)00049-6. Tibor. Illés. Ákos. Szirmai. Tamás. Terlaky. 0953.90055. Postscript preprint. 10.1.1.36.7090.

Sources

Further reading