Namespaces
Variants
Actions

Difference between revisions of "Minimization of the labour of calculation"

From Encyclopedia of Mathematics
Jump to: navigation, search
(Importing text file)
 
m (tex encoded by computer)
Line 1: Line 1:
 +
<!--
 +
m0640001.png
 +
$#A+1 = 128 n = 0
 +
$#C+1 = 128 : ~/encyclopedia/old_files/data/M064/M.0604000 Minimization of the labour of calculation,
 +
Automatically converted into TeX, above some diagnostics.
 +
Please remove this comment and the {{TEX|auto}} line below,
 +
if TeX found to be correct.
 +
-->
 +
 +
{{TEX|auto}}
 +
{{TEX|done}}
 +
 
''minimization of the computational work''
 
''minimization of the computational work''
  
A branch of modern computational mathematics, devoted to the design and investigation of methods which allow one to find an approximate solution, with preassigned accuracy <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640001.png" />, of a problem <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640002.png" /> from a class <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640003.png" /> with the least computational cost (least volume of calculation). This branch of computational mathematics can be related to the more general branch concerned with the problem of optimizing methods (see, for example, [[#References|[1]]], [[#References|[2]]]) in which alongside the problem of minimizing the computational work one considers the dual problem of finding, among the approximation methods requiring roughly the same amount of calculation, a method with maximal accuracy (least error). The latter problem is characteristic, for example, of problems in numerical integration in which the number of points of integration, which measures the computational work, is fixed.
+
A branch of modern computational mathematics, devoted to the design and investigation of methods which allow one to find an approximate solution, with preassigned accuracy $  \epsilon > 0 $,  
 +
of a problem $  P $
 +
from a class $  \{ P \} $
 +
with the least computational cost (least volume of calculation). This branch of computational mathematics can be related to the more general branch concerned with the problem of optimizing methods (see, for example, [[#References|[1]]], [[#References|[2]]]) in which alongside the problem of minimizing the computational work one considers the dual problem of finding, among the approximation methods requiring roughly the same amount of calculation, a method with maximal accuracy (least error). The latter problem is characteristic, for example, of problems in numerical integration in which the number of points of integration, which measures the computational work, is fixed.
  
Let <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640004.png" /> (usually <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640005.png" />, for <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640006.png" /> a precise solution of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640007.png" /> is sought) be the required accuracy for the solution of a problem <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640008.png" /> from a given class <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m0640009.png" /> of related problems, and let <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400010.png" /> be an admissible method for finding the solution of any problem from <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400011.png" />; the set of such methods will be denoted by <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400012.png" />. Let a number <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400013.png" /> characterize the computational cost when by using the method <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400014.png" /> a solution of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400015.png" /> is to be found with accuracy <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400016.png" />, and put
+
Let $  \epsilon \geq  0 $(
 +
usually $  \epsilon > 0 $,  
 +
for $  \epsilon = 0 $
 +
a precise solution of $  P $
 +
is sought) be the required accuracy for the solution of a problem $  P $
 +
from a given class $  \{ P \} $
 +
of related problems, and let m $
 +
be an admissible method for finding the solution of any problem from $  \{ P \} $;  
 +
the set of such methods will be denoted by $  \{ m \} $.  
 +
Let a number $  W _ {m} ( P , \epsilon ) > 0 $
 +
characterize the computational cost when by using the method m $
 +
a solution of $  P $
 +
is to be found with accuracy $  \epsilon $,  
 +
and put
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400017.png" /></td> </tr></table>
+
$$
 +
W _ {m} ( \{ P \} , \epsilon )  = \
 +
\sup _ {P \in \{ P \} }  W _ {m} ( P , \epsilon ) .
 +
$$
  
Then the minimization problem is to find a method <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400018.png" /> such that
+
Then the minimization problem is to find a method $  m _ {0} $
 +
such that
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400019.png" /></td> </tr></table>
+
$$
 +
W _ {m _ {0}  } ( \{ P \} , \epsilon )  = \
 +
\inf _ {m \in \{ m \} } \
 +
W _ {m} ( \{ P \} , \epsilon ) ,
 +
$$
  
that is, in essence one seeks an optimal method for solving not just a fixed problem <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400020.png" />, but a whole class of problems (optimization for a class). Mostly the minimization of the computational work is done in the asymptotic sense as <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400021.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400022.png" />, where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400023.png" /> is a parameter defining the  "dimension"  of the problem to be solved.
+
that is, in essence one seeks an optimal method for solving not just a fixed problem $  P $,  
 +
but a whole class of problems (optimization for a class). Mostly the minimization of the computational work is done in the asymptotic sense as $  \epsilon \rightarrow 0 $,  
 +
$  N \rightarrow \infty $,  
 +
where $  N $
 +
is a parameter defining the  "dimension"  of the problem to be solved.
  
A method <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400024.png" /> is called order optimal if the computational cost in it is at most finitely many times a quantity <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400025.png" /> which is a lower estimate of the computational cost in any possible method; a method <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400026.png" /> is called logarithmically optimal if
+
A method $  m _ {0} $
 +
is called order optimal if the computational cost in it is at most finitely many times a quantity $  \underline{W} ( \{ P \} , \epsilon ) $
 +
which is a lower estimate of the computational cost in any possible method; a method $  m _ {0} $
 +
is called logarithmically optimal if
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400027.png" /></td> </tr></table>
+
$$
 +
\mathop{\rm ln}  W _ {m _ {0}  } ( \{ P \} , \epsilon )  = \
 +
( 1 + o ( 1) )
 +
\mathop{\rm ln}  \underline{W} ( \{ P \} , \epsilon ) .
 +
$$
  
The computational work <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400028.png" /> itself is usually characterized by the number of arithmetic operations generated by the method <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400029.png" /> in attaining the accuracy <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400030.png" /> in its realization on a conventional computer. To some extent this simplification is justified, since in many methods logical operations on real computers are performed in at most finitely many arithmetic operations. For admissible methods, as a rule, their stability with respect to rounding errors is required. It is important that the asymptotic growth of the amount of machine memory used in these methods is not too large.
+
The computational work $  W _ {m} ( P , \epsilon ) $
 +
itself is usually characterized by the number of arithmetic operations generated by the method m $
 +
in attaining the accuracy $  \epsilon $
 +
in its realization on a conventional computer. To some extent this simplification is justified, since in many methods logical operations on real computers are performed in at most finitely many arithmetic operations. For admissible methods, as a rule, their stability with respect to rounding errors is required. It is important that the asymptotic growth of the amount of machine memory used in these methods is not too large.
  
To make this problem of minimizing the computational work concrete, the description of the class <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400031.png" /> must be made precise and the metric space <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400032.png" /> which is used to define the accuracy of a solution of the resulting problem must be stated.
+
To make this problem of minimizing the computational work concrete, the description of the class $  \{ P \} $
 +
must be made precise and the metric space $  H $
 +
which is used to define the accuracy of a solution of the resulting problem must be stated.
  
 
The solution of many such concrete minimization problems is not only of theoretical interest, but has also a great applied significance, allowing one to solve problems on computers with comparatively small expense of machine time. This is particularly important both for problems which require a large amount of calculations, characteristic for example of multi-dimensional problems in mathematical physics (see [[#References|[2]]]–[[#References|[8]]]), and for problems similar to the problems of calculating elementary functions and finding discrete Fourier transforms (see [[#References|[1]]], [[#References|[9]]]), which are standard and repeatedly used for the solution of other, more complex, problems. The problem of minimization of the computational work is fairly complicated and in many cases only a partial solution has been obtained.
 
The solution of many such concrete minimization problems is not only of theoretical interest, but has also a great applied significance, allowing one to solve problems on computers with comparatively small expense of machine time. This is particularly important both for problems which require a large amount of calculations, characteristic for example of multi-dimensional problems in mathematical physics (see [[#References|[2]]]–[[#References|[8]]]), and for problems similar to the problems of calculating elementary functions and finding discrete Fourier transforms (see [[#References|[1]]], [[#References|[9]]]), which are standard and repeatedly used for the solution of other, more complex, problems. The problem of minimization of the computational work is fairly complicated and in many cases only a partial solution has been obtained.
Line 25: Line 77:
 
In practice, for the solution of a problem of relatively small dimension and with no great accuracy, sometimes methods with poor asymptotic characteristics may require less machine time. Often, if the computational cost is acceptable, then the choice of the method is guided first of all by considerations of simplicity and reliability.
 
In practice, for the solution of a problem of relatively small dimension and with no great accuracy, sometimes methods with poor asymptotic characteristics may require less machine time. Often, if the computational cost is acceptable, then the choice of the method is guided first of all by considerations of simplicity and reliability.
  
Suppose that the original problem is finite dimensional and that there are methods which give a precise solution in a finite number of arithmetic operations, if these operations are performed without rounding errors. As examples one can take the problem of calculating the value of a polynomial <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400033.png" /> of degree <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400034.png" /> for given values of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400035.png" /> with <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400036.png" />, the multiplication of two square matrices of order <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400037.png" />, the solution of a given system of algebraic equations <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400038.png" /> with a square matrix of order <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400039.png" />, and the problem of finding the discrete Fourier transform (see [[#References|[1]]], [[#References|[9]]]):
+
Suppose that the original problem is finite dimensional and that there are methods which give a precise solution in a finite number of arithmetic operations, if these operations are performed without rounding errors. As examples one can take the problem of calculating the value of a polynomial $  p _ {N} ( x) $
 +
of degree $  N $
 +
for given values of $  x $
 +
with $  | x | \leq  1 $,  
 +
the multiplication of two square matrices of order $  N $,  
 +
the solution of a given system of algebraic equations $  A x = b $
 +
with a square matrix of order $  N $,  
 +
and the problem of finding the discrete Fourier transform (see [[#References|[1]]], [[#References|[9]]]):
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400040.png" /></td> <td valign="top" style="width:5%;text-align:right;">(1)</td></tr></table>
+
$$ \tag{1 }
 +
v _ {n}  = \
 +
\sum _ { k= } 0 ^ { N- }  1
 +
u _ {k} e ^ {- 2 \pi n i / N } ,\ \
 +
n = 0 \dots N - 1 ,
 +
$$
  
where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400041.png" /> is the imaginary unit, the vector <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400042.png" /> is given and a vector <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400043.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400044.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400045.png" /> is sought. No concrete restrictions are imposed on the form of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400046.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400047.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400048.png" />, the number <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400049.png" />, and the vector <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400050.png" />, and, therefore, in each of these problems the admissible class <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400051.png" /> consists of all problems of such form. In similar problems <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400052.png" /> takes the role of a parameter and special consideration is given to the behaviour of the computational cost <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400053.png" /> as <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400054.png" />. For the first of these problems Horner's method, writing <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400055.png" /> in the form
+
where $  i $
 +
is the imaginary unit, the vector $  u = ( u _ {0} \dots u _ {n-} 1 ) $
 +
is given and a vector $  v \equiv ( v _ {0} \dots v _ {n-} 1 ) $,
 +
$  N = 2  ^ {r} $,  
 +
$  r= 0 , 1 \dots $
 +
is sought. No concrete restrictions are imposed on the form of $  p _ {N} ( x) $,
 +
$  x \in [ - 1 , 1 ] $,  
 +
$  A , b $,  
 +
the number $  r $,  
 +
and the vector $  u $,  
 +
and, therefore, in each of these problems the admissible class $  \{ P \} $
 +
consists of all problems of such form. In similar problems $  N $
 +
takes the role of a parameter and special consideration is given to the behaviour of the computational cost $  W _ {m} ( P , 0 ) \equiv W _ {m} ( P  ) $
 +
as $  N \rightarrow \infty $.  
 +
For the first of these problems Horner's method, writing $  p _ {N} ( x) $
 +
in the form
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400056.png" /></td> </tr></table>
+
$$
 +
a _ {0} + x ( \dots + x ( a _ {N-} 1 + x a _ {N} ) \dots ) ,
 +
$$
  
allows one to calculate <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400057.png" /> using <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400058.png" /> multiplications and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400059.png" /> additions. It has been proved (see [[#References|[9]]]) that this method is optimal: There is no method which would require fewer additions and subtractions or fewer multiplications and divisions; the stability is acceptable if <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400060.png" /> is not large (see [[#References|[1]]]).
+
allows one to calculate $  p _ {N} ( X) $
 +
using $  N $
 +
multiplications and $  N $
 +
additions. It has been proved (see [[#References|[9]]]) that this method is optimal: There is no method which would require fewer additions and subtractions or fewer multiplications and divisions; the stability is acceptable if $  \sum _ {i} | a _ {i} | $
 +
is not large (see [[#References|[1]]]).
  
For the second and third problems there are a number of methods giving solutions as <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400061.png" /> in <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400062.png" /> arithmetic operations (see [[#References|[1]]]) and they are actually applied in practice. The least computational cost among all methods known at present is attained in a method with estimate <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400063.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400064.png" /> (see [[#References|[9]]], [[#References|[10]]]). This method is rather complicated and for a number of reasons is at present interesting only from the theoretical point of view. It is not known (1989), whether there is even a logarithmically-optimal method. There is a conjecture that there exists a logarithmically-optimal method with <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400065.png" /> arithmetic operations, where
+
For the second and third problems there are a number of methods giving solutions as $  N \rightarrow \infty $
 +
in $  W ( N) \sim N  ^ {3} $
 +
arithmetic operations (see [[#References|[1]]]) and they are actually applied in practice. The least computational cost among all methods known at present is attained in a method with estimate $  W ( N) \sim N  ^  \alpha  $,  
 +
$  \alpha \approx 2.5 $(
 +
see [[#References|[9]]], [[#References|[10]]]). This method is rather complicated and for a number of reasons is at present interesting only from the theoretical point of view. It is not known (1989), whether there is even a logarithmically-optimal method. There is a conjecture that there exists a logarithmically-optimal method with $  W ( N) $
 +
arithmetic operations, where
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400066.png" /></td> </tr></table>
+
$$
 +
\mathop{\rm ln}  W ( N)  = 2  \mathop{\rm ln}  N + o (  \mathop{\rm ln}  N ) .
 +
$$
  
For problem (1), the subject of harmonic analysis, the simplest methods require <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400067.png" /> arithmetic operations with complex numbers. In 1965 a method was suggested which allows one to find the vector in <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400068.png" /> arithmetic operations (see [[#References|[1]]], [[#References|[9]]]); this method is called the method of the fast discrete Fourier transform. This method is logarithmically optimal; it is widely used in practice. There are a large number of similar minimization problems, solved and unsolved (see [[#References|[9]]], [[#References|[10]]]); order-optimal or logarithmically-optimal estimates of the computational cost for finding the solution of similar problems may be considered as an indicator of their complexity.
+
For problem (1), the subject of harmonic analysis, the simplest methods require $  \sim N  ^ {2} $
 +
arithmetic operations with complex numbers. In 1965 a method was suggested which allows one to find the vector in $  W ( N) \sim N  \mathop{\rm ln}  N $
 +
arithmetic operations (see [[#References|[1]]], [[#References|[9]]]); this method is called the method of the fast discrete Fourier transform. This method is logarithmically optimal; it is widely used in practice. There are a large number of similar minimization problems, solved and unsolved (see [[#References|[9]]], [[#References|[10]]]); order-optimal or logarithmically-optimal estimates of the computational cost for finding the solution of similar problems may be considered as an indicator of their complexity.
  
Minimization of the computational work for the solution of grid systems of equations, which arise either in difference methods or in projection-grid methods (finite-element methods) for the approximate solution of boundary value problems for equations and systems of equations of strongly-elliptic type, has a special theoretical and applied significance, and, usually, is accomplished asymptotically as <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400069.png" />, where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400070.png" /> is the number of unknowns in the system, and as <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400071.png" />. For the solution of the simplest difference analogues of certain boundary value problems for the Poisson equation in a rectangle or a parallelopiped, certain direct methods have been applied successfully which are logarithmically optimal and require <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400072.png" /> arithmetic operations (see [[#References|[3]]], [[#References|[5]]], [[#References|[8]]], [[#References|[11]]], [[#References|[12]]]). In the case of a rectangle (see [[#References|[13]]]) an order-optimal method is known which requires <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400073.png" /> operations. Using the same iteration method one can obtain a logarithmically-optimal estimate of the type
+
Minimization of the computational work for the solution of grid systems of equations, which arise either in difference methods or in projection-grid methods (finite-element methods) for the approximate solution of boundary value problems for equations and systems of equations of strongly-elliptic type, has a special theoretical and applied significance, and, usually, is accomplished asymptotically as $  N \rightarrow \infty $,  
 +
where $  N $
 +
is the number of unknowns in the system, and as $  \epsilon \rightarrow 0 $.  
 +
For the solution of the simplest difference analogues of certain boundary value problems for the Poisson equation in a rectangle or a parallelopiped, certain direct methods have been applied successfully which are logarithmically optimal and require $  O ( N  \mathop{\rm ln}  N ) $
 +
arithmetic operations (see [[#References|[3]]], [[#References|[5]]], [[#References|[8]]], [[#References|[11]]], [[#References|[12]]]). In the case of a rectangle (see [[#References|[13]]]) an order-optimal method is known which requires $  O ( N) $
 +
operations. Using the same iteration method one can obtain a logarithmically-optimal estimate of the type
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400074.png" /></td> </tr></table>
+
$$
 +
= O ( N  \mathop{\rm ln}  N |  \mathop{\rm ln}  \epsilon | ) ,
 +
$$
  
where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400075.png" /> is the accuracy of the solution of the system in some metric, for a fairly broad class of discrete boundary value problems in a parallelopipedal grid for linear and non-linear strongly-elliptic systems in certain ideal domains (see [[#References|[2]]]–[[#References|[8]]], [[#References|[14]]]). (For example, in a plane <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400076.png" /> they may be generated from a finite number of rectangles with sides parallel to the coordinate axes; and in a three-dimensional space <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400077.png" /> they may be decomposed by a finite number of plane sections parallel to the given coordinate planes into parallelopipeds with boundaries parallel to the coordinate planes.) For more complicated domains <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400078.png" /> the use of grids, topologically equivalent to the above grid domains <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400079.png" />, and certain types of projection-grid methods often allows one to obtain systems of equations which are as effectively solvable as in the case of ideal domains (see [[#References|[3]]], [[#References|[8]]], [[#References|[15]]], [[#References|[16]]]). Here the right-hand sides of these systems can be arbitrary vectors; if it is taken into account that they arise in special form as functionals on sufficiently smooth functions, then methods have been constructed with computational cost <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400080.png" /> and even <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400081.png" />, under the condition that <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400082.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400083.png" />. Methods of the latter type, for the solution of a problem on a given grid, use a sequence of similar problems on sparser grids (see [[#References|[8]]], [[#References|[16]]]).
+
where $  \epsilon $
 +
is the accuracy of the solution of the system in some metric, for a fairly broad class of discrete boundary value problems in a parallelopipedal grid for linear and non-linear strongly-elliptic systems in certain ideal domains (see [[#References|[2]]]–[[#References|[8]]], [[#References|[14]]]). (For example, in a plane $  Q $
 +
they may be generated from a finite number of rectangles with sides parallel to the coordinate axes; and in a three-dimensional space $  Q $
 +
they may be decomposed by a finite number of plane sections parallel to the given coordinate planes into parallelopipeds with boundaries parallel to the coordinate planes.) For more complicated domains $  Q $
 +
the use of grids, topologically equivalent to the above grid domains $  Q $,  
 +
and certain types of projection-grid methods often allows one to obtain systems of equations which are as effectively solvable as in the case of ideal domains (see [[#References|[3]]], [[#References|[8]]], [[#References|[15]]], [[#References|[16]]]). Here the right-hand sides of these systems can be arbitrary vectors; if it is taken into account that they arise in special form as functionals on sufficiently smooth functions, then methods have been constructed with computational cost $  O ( N  \mathop{\rm log}  N ) $
 +
and even $  O ( N) $,  
 +
under the condition that $  \epsilon \sim N ^ {- \alpha } $,
 +
$  \alpha > 0 $.  
 +
Methods of the latter type, for the solution of a problem on a given grid, use a sequence of similar problems on sparser grids (see [[#References|[8]]], [[#References|[16]]]).
  
There are methods which allow one to find with accuracy <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400084.png" /> the minor eigen values and corresponding eigen functions for certain grid analogues of elliptic eigen value problems, at an expense of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400085.png" />, or even <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400086.png" /> (for <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400087.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400088.png" />), operations (see [[#References|[8]]], [[#References|[15]]], [[#References|[16]]]).
+
There are methods which allow one to find with accuracy $  \epsilon $
 +
the minor eigen values and corresponding eigen functions for certain grid analogues of elliptic eigen value problems, at an expense of $  O ( N  \mathop{\rm ln}  N |  \mathop{\rm ln}  \epsilon | ) $,  
 +
or even $  O ( N  \mathop{\rm ln}  N ) $(
 +
for $  \epsilon \sim N ^ {- \alpha } $,
 +
$  \alpha > 0 $),  
 +
operations (see [[#References|[8]]], [[#References|[15]]], [[#References|[16]]]).
  
Let the problem <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400089.png" /> be to solve a well-posed operator equation
+
Let the problem $  P $
 +
be to solve a well-posed operator equation
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400090.png" /></td> <td valign="top" style="width:5%;text-align:right;">(2)</td></tr></table>
+
$$ \tag{2 }
 +
L ( u) =  f ,
 +
$$
  
where the operator <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400091.png" /> acts from <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400092.png" /> to <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400093.png" />, and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400094.png" /> and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400095.png" /> are infinite-dimensional Banach spaces. Let <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400096.png" /> be the class of such problems with different <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400097.png" /> for which the solution to (2) belongs to some compact set <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400098.png" />. Usually <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m06400099.png" /> is given by one condition:
+
where the operator $  L $
 +
acts from $  H $
 +
to $  F $,  
 +
and $  H $
 +
and $  F $
 +
are infinite-dimensional Banach spaces. Let $  \{ P \} $
 +
be the class of such problems with different $  f $
 +
for which the solution to (2) belongs to some compact set $  U $.  
 +
Usually $  U $
 +
is given by one condition:
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000100.png" /></td> <td valign="top" style="width:5%;text-align:right;">(3)</td></tr></table>
+
$$ \tag{3 }
 +
\| u \| _ {H  ^  \prime  }  \leq  R ,
 +
$$
  
where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000101.png" /> is a Banach space imbedded in <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000102.png" />. If <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000103.png" /> is sought with accuracy <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000104.png" /> in the norm of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000105.png" />, then often information estimates, uniform with respect to all <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000106.png" />, are known for the least dimension <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000107.png" /> of a vector <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000108.png" />, the assignment of which allows one to obtain an element
+
where $  H  ^  \prime  $
 +
is a Banach space imbedded in $  H $.  
 +
If $  u \in U $
 +
is sought with accuracy $  \epsilon > 0 $
 +
in the norm of $  H $,  
 +
then often information estimates, uniform with respect to all $  u \in U $,  
 +
are known for the least dimension $  \underline{N} \equiv \underline{N} ( \epsilon ) $
 +
of a vector $  v _ {\underline{N}  } \equiv ( v _ {1} \dots v _ {\underline{N}  } ) $,  
 +
the assignment of which allows one to obtain an element
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000109.png" /></td> <td valign="top" style="width:5%;text-align:right;">(4)</td></tr></table>
+
$$ \tag{4 }
 +
\widehat{v}  _ {\underline{N}  }  \equiv  \sum _ { i= } 1 ^ { {\underline{N} }  } v _ {i} \psi _ {i} ( z)
 +
\in  H \  \textrm{ with }  \| u - \widehat{v}  _ {\underline{N}  } \| _ {H}  \leq  \epsilon
 +
$$
  
(see [[#References|[1]]], [[#References|[2]]], [[#References|[8]]], [[#References|[17]]], [[#References|[18]]]). These lower estimates for <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000110.png" /> give obvious lower estimates for the required computational cost in any admissible method. In a number of boundary value problems for strongly-elliptic systems variants of the projection-grid method have been constructed which lead to algebraic systems of equations
+
(see [[#References|[1]]], [[#References|[2]]], [[#References|[8]]], [[#References|[17]]], [[#References|[18]]]). These lower estimates for $  N ( \epsilon ) $
 +
give obvious lower estimates for the required computational cost in any admissible method. In a number of boundary value problems for strongly-elliptic systems variants of the projection-grid method have been constructed which lead to algebraic systems of equations
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000111.png" /></td> <td valign="top" style="width:5%;text-align:right;">(5)</td></tr></table>
+
$$ \tag{5 }
 +
L _ {N} ( \overline{u}\; _ {N} )  = \overline{f}\; _ {N}  $$
  
in <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000112.png" /> unknowns, for which, first, order-optimal iteration methods of approximate solution are known and, secondly, the corresponding functions <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000113.png" /> of the vectors <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000114.png" /> (see (4)) are <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000115.png" />-approximations in <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000116.png" /> of the solution of (2), where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000117.png" /> and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000118.png" /> is a constant not depending on <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000119.png" /> (see [[#References|[7]]], [[#References|[8]]], [[#References|[16]]]). If the effort in forming (5) is not taken into account, these methods lead to order-minimal estimates for the computational cost. For example, for a second-order elliptic equation, where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000120.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000121.png" /> (<img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000122.png" /> are the Sobolev spaces [[#References|[19]]]), if <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000123.png" /> is a domain in the plane, the estimate <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000124.png" /> for the expense of the number of arithmetic operations is obtained. The computational work in forming (5) can often also be estimated as <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000125.png" /> if information on the given functions in the corresponding spaces is used. In particular, in the example <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000126.png" /> must be considered as an element of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000127.png" />, but not of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/m/m064/m064000/m064000128.png" /> (see [[#References|[16]]]). A similar kind of estimate has been obtained for certain differential eigen value problems (see [[#References|[8]]], [[#References|[15]]]).
+
in $  N $
 +
unknowns, for which, first, order-optimal iteration methods of approximate solution are known and, secondly, the corresponding functions $  \widehat{u}  _ {N} $
 +
of the vectors $  \overline{u}\; _ {N} $(
 +
see (4)) are $  \epsilon $-
 +
approximations in $  H $
 +
of the solution of (2), where $  N \leq  k \underline{N} ( \epsilon ) $
 +
and $  k $
 +
is a constant not depending on $  \epsilon $(
 +
see [[#References|[7]]], [[#References|[8]]], [[#References|[16]]]). If the effort in forming (5) is not taken into account, these methods lead to order-minimal estimates for the computational cost. For example, for a second-order elliptic equation, where $  H = W _ {2}  ^ {1} ( \Omega ) $,  
 +
$  H  ^  \prime  = W _ {2}  ^ {2} ( \Omega ) $(
 +
$  W _ {2}  ^ {k} ( \Omega ) $
 +
are the Sobolev spaces [[#References|[19]]]), if $  \Omega $
 +
is a domain in the plane, the estimate $  \sim \epsilon  ^ {-} 2 $
 +
for the expense of the number of arithmetic operations is obtained. The computational work in forming (5) can often also be estimated as $  O ( N  \mathop{\rm ln}  N ) $
 +
if information on the given functions in the corresponding spaces is used. In particular, in the example $  f $
 +
must be considered as an element of $  F = W _ {2}  ^ {-} 1 ( \Omega ) $,  
 +
but not of $  L _ {2} ( \Omega ) $(
 +
see [[#References|[16]]]). A similar kind of estimate has been obtained for certain differential eigen value problems (see [[#References|[8]]], [[#References|[15]]]).
  
 
Questions on the minimization of the computational work for the solution of integral equations, ordinary differential equations and non-stationary partial differential equations have also been discussed (see, for example, [[#References|[1]]]–[[#References|[4]]], [[#References|[20]]]–[[#References|[22]]]).
 
Questions on the minimization of the computational work for the solution of integral equations, ordinary differential equations and non-stationary partial differential equations have also been discussed (see, for example, [[#References|[1]]]–[[#References|[4]]], [[#References|[20]]]–[[#References|[22]]]).
Line 71: Line 230:
 
====References====
 
====References====
 
<table><TR><TD valign="top">[1]</TD> <TD valign="top">  N.S. Bakhvalov,  "Numerical methods: analysis, algebra, ordinary differential equations" , MIR  (1977)  (Translated from Russian)</TD></TR><TR><TD valign="top">[2]</TD> <TD valign="top">  J.F. Traub,  H. Wozniakowski,  "A general theory of optimal algorithms" , Acad. Press  (1980)</TD></TR><TR><TD valign="top">[3]</TD> <TD valign="top">  G.I. Marchuk,  "Methods of numerical mathematics" , Springer  (1975)  (Translated from Russian)</TD></TR><TR><TD valign="top">[4]</TD> <TD valign="top">  A.A. Samarskii,  "Theorie der Differenzverfahren" , Akad. Verlagsgesell. Geest u. Portig K.-D.  (1984)  (Translated from Russian)</TD></TR><TR><TD valign="top">[5]</TD> <TD valign="top">  A.A. Samarskii,  E.S. Nikolaev,  "Numerical methods for grid equations" , '''1–2''' , Birkhäuser  (1989)  (Translated from Russian)</TD></TR><TR><TD valign="top">[6]</TD> <TD valign="top">  R. Glowinski,  "Numerical methods for nonlinear variational problems" , Springer  (1984)</TD></TR><TR><TD valign="top">[7]</TD> <TD valign="top">  V.G. Korneev,  "Schemes for the finite element method with a high order of accuracy" , Leningrad  (1977)  (In Russian)</TD></TR><TR><TD valign="top">[8]</TD> <TD valign="top">  A.G. D'yakonov,  "Minimization of computational work. Asymptotically-optimal algorithms" , Moscow  (1989)  (In Russian)</TD></TR><TR><TD valign="top">[9]</TD> <TD valign="top">  A. Borodin,  I. Munro,  "The computational complexity of algebraic and numeric problems" , Amer. Elsevier  (1975)</TD></TR><TR><TD valign="top">[10]</TD> <TD valign="top">  J.F. Traub (ed.) , ''Analytic computational complexity'' , Acad. Press  (1976)</TD></TR><TR><TD valign="top">[11]</TD> <TD valign="top">  P. Concus,  G. Golub,  "Use of fast direct methods for the efficient numerical solution of nonseparable elliptic equations"  ''SIAM J. Numer. Anal.'' , '''10'''  (1973)  pp. 1103–1120</TD></TR><TR><TD valign="top">[12]</TD> <TD valign="top">  R.J. Barker,  "Advanced techniques for the direct numerical solution of Poisson's equation" , ''Mathematical Models and Numerical Methods'' , ''Banach Center Publ.'' , '''3''' , PWN  (1978)  pp. 255–268</TD></TR><TR><TD valign="top">[13]</TD> <TD valign="top">  R.E. Bank,  D.J. Rose,  "Extrapolated fast direct algorithms for elliptic boundary value problems" , ''Algorithms and complexity (Proc. Symp. Carnegie-Mellon Univ.)'' , Acad. Press  (1976)  pp. 201–249</TD></TR><TR><TD valign="top">[14]</TD> <TD valign="top">  G.I. Marchuk,  Yu.A. Kuznetsov,  A.M. Matsokin,  "Fictitious domain and domain decomposition methods"  ''Soviet J. Numer. Anal. Math. Modelling'' , '''1''' :  1  (1986)  pp. 3–35</TD></TR><TR><TD valign="top">[15]</TD> <TD valign="top">  E.G. D'yakonov,  "Effective methods for solving eigenvalue problems with fourth-order elliptic operators"  ''Soviet J. Anal. Math. Modelling'' , '''1''' :  1  (1986)  pp. 59–82</TD></TR><TR><TD valign="top">[16]</TD> <TD valign="top">  U. Trottenberg (ed.) , ''Multigrid Methods. Proc. Köln-Porz, 1981'' , ''Lect. notes in math.'' , '''960''' , Springer  (1982)</TD></TR><TR><TD valign="top">[17]</TD> <TD valign="top"> , ''Theoretical foundations and construction of numerical algorithms of problems of mathematical physics'' , Moscow  (1979)  pp. 45–118  (In Russian)</TD></TR><TR><TD valign="top">[18]</TD> <TD valign="top">  V.M. Tikhomirov,  "Some problems in approximation theory" , Moscow  (1976)  (In Russian)</TD></TR><TR><TD valign="top">[19]</TD> <TD valign="top">  S.L. Sobolev,  "Applications of functional analysis in mathematical physics" , Amer. Math. Soc.  (1963)  (Translated from Russian)</TD></TR><TR><TD valign="top">[20]</TD> <TD valign="top">  K.V. Emel'yanov,  A.M. Il'in,  "Number of arithmetical operations necessary for the approximate solution of Fredholm integral equations of the second kind"  ''USSR Comp. Math. Math. Phys.'' , '''7''' :  4  (1967)  pp. 259–266  ''Zh. Vychisl. Mat. i Mat. Fiz.'' , '''7''' :  4  (1967)  pp. 905–910</TD></TR><TR><TD valign="top">[21]</TD> <TD valign="top">  V.V. Ivanov,  "Optimal algorithms for the numerical solution of singular integral equations" , ''Continuum Mechanics and Related Problems in Anal.'' , Moscow  (1972)  pp. 209–219  (In Russian)</TD></TR><TR><TD valign="top">[22]</TD> <TD valign="top">  Yu.R. Akopyan,  L.A. Oganesyan,  "A variational-difference method for solving two-dimensional linear parabolic equations"  ''USSR Comp. Math. Math. Phys.'' , '''17''' :  1  (1977)  pp. 101–110  ''Zh. Vychisl. Mat. Mat. Fiz.'' , '''17''' :  1  (1977)  pp. 109–118</TD></TR></table>
 
<table><TR><TD valign="top">[1]</TD> <TD valign="top">  N.S. Bakhvalov,  "Numerical methods: analysis, algebra, ordinary differential equations" , MIR  (1977)  (Translated from Russian)</TD></TR><TR><TD valign="top">[2]</TD> <TD valign="top">  J.F. Traub,  H. Wozniakowski,  "A general theory of optimal algorithms" , Acad. Press  (1980)</TD></TR><TR><TD valign="top">[3]</TD> <TD valign="top">  G.I. Marchuk,  "Methods of numerical mathematics" , Springer  (1975)  (Translated from Russian)</TD></TR><TR><TD valign="top">[4]</TD> <TD valign="top">  A.A. Samarskii,  "Theorie der Differenzverfahren" , Akad. Verlagsgesell. Geest u. Portig K.-D.  (1984)  (Translated from Russian)</TD></TR><TR><TD valign="top">[5]</TD> <TD valign="top">  A.A. Samarskii,  E.S. Nikolaev,  "Numerical methods for grid equations" , '''1–2''' , Birkhäuser  (1989)  (Translated from Russian)</TD></TR><TR><TD valign="top">[6]</TD> <TD valign="top">  R. Glowinski,  "Numerical methods for nonlinear variational problems" , Springer  (1984)</TD></TR><TR><TD valign="top">[7]</TD> <TD valign="top">  V.G. Korneev,  "Schemes for the finite element method with a high order of accuracy" , Leningrad  (1977)  (In Russian)</TD></TR><TR><TD valign="top">[8]</TD> <TD valign="top">  A.G. D'yakonov,  "Minimization of computational work. Asymptotically-optimal algorithms" , Moscow  (1989)  (In Russian)</TD></TR><TR><TD valign="top">[9]</TD> <TD valign="top">  A. Borodin,  I. Munro,  "The computational complexity of algebraic and numeric problems" , Amer. Elsevier  (1975)</TD></TR><TR><TD valign="top">[10]</TD> <TD valign="top">  J.F. Traub (ed.) , ''Analytic computational complexity'' , Acad. Press  (1976)</TD></TR><TR><TD valign="top">[11]</TD> <TD valign="top">  P. Concus,  G. Golub,  "Use of fast direct methods for the efficient numerical solution of nonseparable elliptic equations"  ''SIAM J. Numer. Anal.'' , '''10'''  (1973)  pp. 1103–1120</TD></TR><TR><TD valign="top">[12]</TD> <TD valign="top">  R.J. Barker,  "Advanced techniques for the direct numerical solution of Poisson's equation" , ''Mathematical Models and Numerical Methods'' , ''Banach Center Publ.'' , '''3''' , PWN  (1978)  pp. 255–268</TD></TR><TR><TD valign="top">[13]</TD> <TD valign="top">  R.E. Bank,  D.J. Rose,  "Extrapolated fast direct algorithms for elliptic boundary value problems" , ''Algorithms and complexity (Proc. Symp. Carnegie-Mellon Univ.)'' , Acad. Press  (1976)  pp. 201–249</TD></TR><TR><TD valign="top">[14]</TD> <TD valign="top">  G.I. Marchuk,  Yu.A. Kuznetsov,  A.M. Matsokin,  "Fictitious domain and domain decomposition methods"  ''Soviet J. Numer. Anal. Math. Modelling'' , '''1''' :  1  (1986)  pp. 3–35</TD></TR><TR><TD valign="top">[15]</TD> <TD valign="top">  E.G. D'yakonov,  "Effective methods for solving eigenvalue problems with fourth-order elliptic operators"  ''Soviet J. Anal. Math. Modelling'' , '''1''' :  1  (1986)  pp. 59–82</TD></TR><TR><TD valign="top">[16]</TD> <TD valign="top">  U. Trottenberg (ed.) , ''Multigrid Methods. Proc. Köln-Porz, 1981'' , ''Lect. notes in math.'' , '''960''' , Springer  (1982)</TD></TR><TR><TD valign="top">[17]</TD> <TD valign="top"> , ''Theoretical foundations and construction of numerical algorithms of problems of mathematical physics'' , Moscow  (1979)  pp. 45–118  (In Russian)</TD></TR><TR><TD valign="top">[18]</TD> <TD valign="top">  V.M. Tikhomirov,  "Some problems in approximation theory" , Moscow  (1976)  (In Russian)</TD></TR><TR><TD valign="top">[19]</TD> <TD valign="top">  S.L. Sobolev,  "Applications of functional analysis in mathematical physics" , Amer. Math. Soc.  (1963)  (Translated from Russian)</TD></TR><TR><TD valign="top">[20]</TD> <TD valign="top">  K.V. Emel'yanov,  A.M. Il'in,  "Number of arithmetical operations necessary for the approximate solution of Fredholm integral equations of the second kind"  ''USSR Comp. Math. Math. Phys.'' , '''7''' :  4  (1967)  pp. 259–266  ''Zh. Vychisl. Mat. i Mat. Fiz.'' , '''7''' :  4  (1967)  pp. 905–910</TD></TR><TR><TD valign="top">[21]</TD> <TD valign="top">  V.V. Ivanov,  "Optimal algorithms for the numerical solution of singular integral equations" , ''Continuum Mechanics and Related Problems in Anal.'' , Moscow  (1972)  pp. 209–219  (In Russian)</TD></TR><TR><TD valign="top">[22]</TD> <TD valign="top">  Yu.R. Akopyan,  L.A. Oganesyan,  "A variational-difference method for solving two-dimensional linear parabolic equations"  ''USSR Comp. Math. Math. Phys.'' , '''17''' :  1  (1977)  pp. 101–110  ''Zh. Vychisl. Mat. Mat. Fiz.'' , '''17''' :  1  (1977)  pp. 109–118</TD></TR></table>
 
 
  
 
====Comments====
 
====Comments====
 
  
 
====References====
 
====References====
 
<table><TR><TD valign="top">[a1]</TD> <TD valign="top">  J.F. Traub,  H. Wozniakowski,  "Information, uncertainty, complexity" , Addison-Wesley  (1983)</TD></TR></table>
 
<table><TR><TD valign="top">[a1]</TD> <TD valign="top">  J.F. Traub,  H. Wozniakowski,  "Information, uncertainty, complexity" , Addison-Wesley  (1983)</TD></TR></table>

Revision as of 08:00, 6 June 2020


minimization of the computational work

A branch of modern computational mathematics, devoted to the design and investigation of methods which allow one to find an approximate solution, with preassigned accuracy $ \epsilon > 0 $, of a problem $ P $ from a class $ \{ P \} $ with the least computational cost (least volume of calculation). This branch of computational mathematics can be related to the more general branch concerned with the problem of optimizing methods (see, for example, [1], [2]) in which alongside the problem of minimizing the computational work one considers the dual problem of finding, among the approximation methods requiring roughly the same amount of calculation, a method with maximal accuracy (least error). The latter problem is characteristic, for example, of problems in numerical integration in which the number of points of integration, which measures the computational work, is fixed.

Let $ \epsilon \geq 0 $( usually $ \epsilon > 0 $, for $ \epsilon = 0 $ a precise solution of $ P $ is sought) be the required accuracy for the solution of a problem $ P $ from a given class $ \{ P \} $ of related problems, and let $ m $ be an admissible method for finding the solution of any problem from $ \{ P \} $; the set of such methods will be denoted by $ \{ m \} $. Let a number $ W _ {m} ( P , \epsilon ) > 0 $ characterize the computational cost when by using the method $ m $ a solution of $ P $ is to be found with accuracy $ \epsilon $, and put

$$ W _ {m} ( \{ P \} , \epsilon ) = \ \sup _ {P \in \{ P \} } W _ {m} ( P , \epsilon ) . $$

Then the minimization problem is to find a method $ m _ {0} $ such that

$$ W _ {m _ {0} } ( \{ P \} , \epsilon ) = \ \inf _ {m \in \{ m \} } \ W _ {m} ( \{ P \} , \epsilon ) , $$

that is, in essence one seeks an optimal method for solving not just a fixed problem $ P $, but a whole class of problems (optimization for a class). Mostly the minimization of the computational work is done in the asymptotic sense as $ \epsilon \rightarrow 0 $, $ N \rightarrow \infty $, where $ N $ is a parameter defining the "dimension" of the problem to be solved.

A method $ m _ {0} $ is called order optimal if the computational cost in it is at most finitely many times a quantity $ \underline{W} ( \{ P \} , \epsilon ) $ which is a lower estimate of the computational cost in any possible method; a method $ m _ {0} $ is called logarithmically optimal if

$$ \mathop{\rm ln} W _ {m _ {0} } ( \{ P \} , \epsilon ) = \ ( 1 + o ( 1) ) \mathop{\rm ln} \underline{W} ( \{ P \} , \epsilon ) . $$

The computational work $ W _ {m} ( P , \epsilon ) $ itself is usually characterized by the number of arithmetic operations generated by the method $ m $ in attaining the accuracy $ \epsilon $ in its realization on a conventional computer. To some extent this simplification is justified, since in many methods logical operations on real computers are performed in at most finitely many arithmetic operations. For admissible methods, as a rule, their stability with respect to rounding errors is required. It is important that the asymptotic growth of the amount of machine memory used in these methods is not too large.

To make this problem of minimizing the computational work concrete, the description of the class $ \{ P \} $ must be made precise and the metric space $ H $ which is used to define the accuracy of a solution of the resulting problem must be stated.

The solution of many such concrete minimization problems is not only of theoretical interest, but has also a great applied significance, allowing one to solve problems on computers with comparatively small expense of machine time. This is particularly important both for problems which require a large amount of calculations, characteristic for example of multi-dimensional problems in mathematical physics (see [2][8]), and for problems similar to the problems of calculating elementary functions and finding discrete Fourier transforms (see [1], [9]), which are standard and repeatedly used for the solution of other, more complex, problems. The problem of minimization of the computational work is fairly complicated and in many cases only a partial solution has been obtained.

In practice, for the solution of a problem of relatively small dimension and with no great accuracy, sometimes methods with poor asymptotic characteristics may require less machine time. Often, if the computational cost is acceptable, then the choice of the method is guided first of all by considerations of simplicity and reliability.

Suppose that the original problem is finite dimensional and that there are methods which give a precise solution in a finite number of arithmetic operations, if these operations are performed without rounding errors. As examples one can take the problem of calculating the value of a polynomial $ p _ {N} ( x) $ of degree $ N $ for given values of $ x $ with $ | x | \leq 1 $, the multiplication of two square matrices of order $ N $, the solution of a given system of algebraic equations $ A x = b $ with a square matrix of order $ N $, and the problem of finding the discrete Fourier transform (see [1], [9]):

$$ \tag{1 } v _ {n} = \ \sum _ { k= } 0 ^ { N- } 1 u _ {k} e ^ {- 2 \pi n i / N } ,\ \ n = 0 \dots N - 1 , $$

where $ i $ is the imaginary unit, the vector $ u = ( u _ {0} \dots u _ {n-} 1 ) $ is given and a vector $ v \equiv ( v _ {0} \dots v _ {n-} 1 ) $, $ N = 2 ^ {r} $, $ r= 0 , 1 \dots $ is sought. No concrete restrictions are imposed on the form of $ p _ {N} ( x) $, $ x \in [ - 1 , 1 ] $, $ A , b $, the number $ r $, and the vector $ u $, and, therefore, in each of these problems the admissible class $ \{ P \} $ consists of all problems of such form. In similar problems $ N $ takes the role of a parameter and special consideration is given to the behaviour of the computational cost $ W _ {m} ( P , 0 ) \equiv W _ {m} ( P ) $ as $ N \rightarrow \infty $. For the first of these problems Horner's method, writing $ p _ {N} ( x) $ in the form

$$ a _ {0} + x ( \dots + x ( a _ {N-} 1 + x a _ {N} ) \dots ) , $$

allows one to calculate $ p _ {N} ( X) $ using $ N $ multiplications and $ N $ additions. It has been proved (see [9]) that this method is optimal: There is no method which would require fewer additions and subtractions or fewer multiplications and divisions; the stability is acceptable if $ \sum _ {i} | a _ {i} | $ is not large (see [1]).

For the second and third problems there are a number of methods giving solutions as $ N \rightarrow \infty $ in $ W ( N) \sim N ^ {3} $ arithmetic operations (see [1]) and they are actually applied in practice. The least computational cost among all methods known at present is attained in a method with estimate $ W ( N) \sim N ^ \alpha $, $ \alpha \approx 2.5 $( see [9], [10]). This method is rather complicated and for a number of reasons is at present interesting only from the theoretical point of view. It is not known (1989), whether there is even a logarithmically-optimal method. There is a conjecture that there exists a logarithmically-optimal method with $ W ( N) $ arithmetic operations, where

$$ \mathop{\rm ln} W ( N) = 2 \mathop{\rm ln} N + o ( \mathop{\rm ln} N ) . $$

For problem (1), the subject of harmonic analysis, the simplest methods require $ \sim N ^ {2} $ arithmetic operations with complex numbers. In 1965 a method was suggested which allows one to find the vector in $ W ( N) \sim N \mathop{\rm ln} N $ arithmetic operations (see [1], [9]); this method is called the method of the fast discrete Fourier transform. This method is logarithmically optimal; it is widely used in practice. There are a large number of similar minimization problems, solved and unsolved (see [9], [10]); order-optimal or logarithmically-optimal estimates of the computational cost for finding the solution of similar problems may be considered as an indicator of their complexity.

Minimization of the computational work for the solution of grid systems of equations, which arise either in difference methods or in projection-grid methods (finite-element methods) for the approximate solution of boundary value problems for equations and systems of equations of strongly-elliptic type, has a special theoretical and applied significance, and, usually, is accomplished asymptotically as $ N \rightarrow \infty $, where $ N $ is the number of unknowns in the system, and as $ \epsilon \rightarrow 0 $. For the solution of the simplest difference analogues of certain boundary value problems for the Poisson equation in a rectangle or a parallelopiped, certain direct methods have been applied successfully which are logarithmically optimal and require $ O ( N \mathop{\rm ln} N ) $ arithmetic operations (see [3], [5], [8], [11], [12]). In the case of a rectangle (see [13]) an order-optimal method is known which requires $ O ( N) $ operations. Using the same iteration method one can obtain a logarithmically-optimal estimate of the type

$$ W = O ( N \mathop{\rm ln} N | \mathop{\rm ln} \epsilon | ) , $$

where $ \epsilon $ is the accuracy of the solution of the system in some metric, for a fairly broad class of discrete boundary value problems in a parallelopipedal grid for linear and non-linear strongly-elliptic systems in certain ideal domains (see [2][8], [14]). (For example, in a plane $ Q $ they may be generated from a finite number of rectangles with sides parallel to the coordinate axes; and in a three-dimensional space $ Q $ they may be decomposed by a finite number of plane sections parallel to the given coordinate planes into parallelopipeds with boundaries parallel to the coordinate planes.) For more complicated domains $ Q $ the use of grids, topologically equivalent to the above grid domains $ Q $, and certain types of projection-grid methods often allows one to obtain systems of equations which are as effectively solvable as in the case of ideal domains (see [3], [8], [15], [16]). Here the right-hand sides of these systems can be arbitrary vectors; if it is taken into account that they arise in special form as functionals on sufficiently smooth functions, then methods have been constructed with computational cost $ O ( N \mathop{\rm log} N ) $ and even $ O ( N) $, under the condition that $ \epsilon \sim N ^ {- \alpha } $, $ \alpha > 0 $. Methods of the latter type, for the solution of a problem on a given grid, use a sequence of similar problems on sparser grids (see [8], [16]).

There are methods which allow one to find with accuracy $ \epsilon $ the minor eigen values and corresponding eigen functions for certain grid analogues of elliptic eigen value problems, at an expense of $ O ( N \mathop{\rm ln} N | \mathop{\rm ln} \epsilon | ) $, or even $ O ( N \mathop{\rm ln} N ) $( for $ \epsilon \sim N ^ {- \alpha } $, $ \alpha > 0 $), operations (see [8], [15], [16]).

Let the problem $ P $ be to solve a well-posed operator equation

$$ \tag{2 } L ( u) = f , $$

where the operator $ L $ acts from $ H $ to $ F $, and $ H $ and $ F $ are infinite-dimensional Banach spaces. Let $ \{ P \} $ be the class of such problems with different $ f $ for which the solution to (2) belongs to some compact set $ U $. Usually $ U $ is given by one condition:

$$ \tag{3 } \| u \| _ {H ^ \prime } \leq R , $$

where $ H ^ \prime $ is a Banach space imbedded in $ H $. If $ u \in U $ is sought with accuracy $ \epsilon > 0 $ in the norm of $ H $, then often information estimates, uniform with respect to all $ u \in U $, are known for the least dimension $ \underline{N} \equiv \underline{N} ( \epsilon ) $ of a vector $ v _ {\underline{N} } \equiv ( v _ {1} \dots v _ {\underline{N} } ) $, the assignment of which allows one to obtain an element

$$ \tag{4 } \widehat{v} _ {\underline{N} } \equiv \sum _ { i= } 1 ^ { {\underline{N} } } v _ {i} \psi _ {i} ( z) \in H \ \textrm{ with } \| u - \widehat{v} _ {\underline{N} } \| _ {H} \leq \epsilon $$

(see [1], [2], [8], [17], [18]). These lower estimates for $ N ( \epsilon ) $ give obvious lower estimates for the required computational cost in any admissible method. In a number of boundary value problems for strongly-elliptic systems variants of the projection-grid method have been constructed which lead to algebraic systems of equations

$$ \tag{5 } L _ {N} ( \overline{u}\; _ {N} ) = \overline{f}\; _ {N} $$

in $ N $ unknowns, for which, first, order-optimal iteration methods of approximate solution are known and, secondly, the corresponding functions $ \widehat{u} _ {N} $ of the vectors $ \overline{u}\; _ {N} $( see (4)) are $ \epsilon $- approximations in $ H $ of the solution of (2), where $ N \leq k \underline{N} ( \epsilon ) $ and $ k $ is a constant not depending on $ \epsilon $( see [7], [8], [16]). If the effort in forming (5) is not taken into account, these methods lead to order-minimal estimates for the computational cost. For example, for a second-order elliptic equation, where $ H = W _ {2} ^ {1} ( \Omega ) $, $ H ^ \prime = W _ {2} ^ {2} ( \Omega ) $( $ W _ {2} ^ {k} ( \Omega ) $ are the Sobolev spaces [19]), if $ \Omega $ is a domain in the plane, the estimate $ \sim \epsilon ^ {-} 2 $ for the expense of the number of arithmetic operations is obtained. The computational work in forming (5) can often also be estimated as $ O ( N \mathop{\rm ln} N ) $ if information on the given functions in the corresponding spaces is used. In particular, in the example $ f $ must be considered as an element of $ F = W _ {2} ^ {-} 1 ( \Omega ) $, but not of $ L _ {2} ( \Omega ) $( see [16]). A similar kind of estimate has been obtained for certain differential eigen value problems (see [8], [15]).

Questions on the minimization of the computational work for the solution of integral equations, ordinary differential equations and non-stationary partial differential equations have also been discussed (see, for example, [1][4], [20][22]).

References

[1] N.S. Bakhvalov, "Numerical methods: analysis, algebra, ordinary differential equations" , MIR (1977) (Translated from Russian)
[2] J.F. Traub, H. Wozniakowski, "A general theory of optimal algorithms" , Acad. Press (1980)
[3] G.I. Marchuk, "Methods of numerical mathematics" , Springer (1975) (Translated from Russian)
[4] A.A. Samarskii, "Theorie der Differenzverfahren" , Akad. Verlagsgesell. Geest u. Portig K.-D. (1984) (Translated from Russian)
[5] A.A. Samarskii, E.S. Nikolaev, "Numerical methods for grid equations" , 1–2 , Birkhäuser (1989) (Translated from Russian)
[6] R. Glowinski, "Numerical methods for nonlinear variational problems" , Springer (1984)
[7] V.G. Korneev, "Schemes for the finite element method with a high order of accuracy" , Leningrad (1977) (In Russian)
[8] A.G. D'yakonov, "Minimization of computational work. Asymptotically-optimal algorithms" , Moscow (1989) (In Russian)
[9] A. Borodin, I. Munro, "The computational complexity of algebraic and numeric problems" , Amer. Elsevier (1975)
[10] J.F. Traub (ed.) , Analytic computational complexity , Acad. Press (1976)
[11] P. Concus, G. Golub, "Use of fast direct methods for the efficient numerical solution of nonseparable elliptic equations" SIAM J. Numer. Anal. , 10 (1973) pp. 1103–1120
[12] R.J. Barker, "Advanced techniques for the direct numerical solution of Poisson's equation" , Mathematical Models and Numerical Methods , Banach Center Publ. , 3 , PWN (1978) pp. 255–268
[13] R.E. Bank, D.J. Rose, "Extrapolated fast direct algorithms for elliptic boundary value problems" , Algorithms and complexity (Proc. Symp. Carnegie-Mellon Univ.) , Acad. Press (1976) pp. 201–249
[14] G.I. Marchuk, Yu.A. Kuznetsov, A.M. Matsokin, "Fictitious domain and domain decomposition methods" Soviet J. Numer. Anal. Math. Modelling , 1 : 1 (1986) pp. 3–35
[15] E.G. D'yakonov, "Effective methods for solving eigenvalue problems with fourth-order elliptic operators" Soviet J. Anal. Math. Modelling , 1 : 1 (1986) pp. 59–82
[16] U. Trottenberg (ed.) , Multigrid Methods. Proc. Köln-Porz, 1981 , Lect. notes in math. , 960 , Springer (1982)
[17] , Theoretical foundations and construction of numerical algorithms of problems of mathematical physics , Moscow (1979) pp. 45–118 (In Russian)
[18] V.M. Tikhomirov, "Some problems in approximation theory" , Moscow (1976) (In Russian)
[19] S.L. Sobolev, "Applications of functional analysis in mathematical physics" , Amer. Math. Soc. (1963) (Translated from Russian)
[20] K.V. Emel'yanov, A.M. Il'in, "Number of arithmetical operations necessary for the approximate solution of Fredholm integral equations of the second kind" USSR Comp. Math. Math. Phys. , 7 : 4 (1967) pp. 259–266 Zh. Vychisl. Mat. i Mat. Fiz. , 7 : 4 (1967) pp. 905–910
[21] V.V. Ivanov, "Optimal algorithms for the numerical solution of singular integral equations" , Continuum Mechanics and Related Problems in Anal. , Moscow (1972) pp. 209–219 (In Russian)
[22] Yu.R. Akopyan, L.A. Oganesyan, "A variational-difference method for solving two-dimensional linear parabolic equations" USSR Comp. Math. Math. Phys. , 17 : 1 (1977) pp. 101–110 Zh. Vychisl. Mat. Mat. Fiz. , 17 : 1 (1977) pp. 109–118

Comments

References

[a1] J.F. Traub, H. Wozniakowski, "Information, uncertainty, complexity" , Addison-Wesley (1983)
How to Cite This Entry:
Minimization of the labour of calculation. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Minimization_of_the_labour_of_calculation&oldid=18453
This article was adapted from an original article by E.G. D'yakonov (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article