Óbudai Egyetem. Doktori (PhD) értekezés tézisfüzete. Automatic roundo error analysis of numerical algorithms. Attila Gáti - PDF

Óbudai Egyetem Doktori (PhD) értekezés tézisfüzete Automatic roundo error analysis of numerical algorithms by Attila Gáti Témavezet½ok: Prof. Dr. László Horváth Prof. Dr. István Krómer Alkalmazott Informatikai

Please download to get full document.

View again

of 16
All materials on our website are shared by users. If you have any questions about copyright issues, please report us to resolve them. We are always happy to assist you.


Publish on:

Views: 17 | Pages: 16

Extension: PDF | Download: 0

Óbudai Egyetem Doktori (PhD) értekezés tézisfüzete Automatic roundo error analysis of numerical algorithms by Attila Gáti Témavezet½ok: Prof. Dr. László Horváth Prof. Dr. István Krómer Alkalmazott Informatikai Doktori Iskola Budapest, 2013 Contents 1 The background of the research 2 2 The aims of the research 4 3 The methods of investigation 5 4 The new scienti c results 5 5 The applicability of the results 6 6 List of references 8 7 List of references related to the thesis 14 1 1 The background of the research The idea of automatic error analysis of algorithms and/or arithmetic expressions is as old as the scienti c computing itself and originates from Wilkinson (see Higham [27]), who also developed the theory of oating point error analysis [68], which is the basic of today s oating point arithmetic standards ([53], [50]). There are various forms of automatic error analysis with usually partial solutions to the problem (see, e.g. Higham [27],[21]). The most impressive approach is the interval analysis, although its use is limited by the technique itself (see. e.g. [47], [48], [49], [31], [32], [64], [27]). Most of the scienti c algorithms however are implemented in oating point arithmetic and the users are interested in the numerical stability or robustness of these implementations. The theoretical investigations are usually di cult, require special knowledge and they not always result in conclusions that are acceptable in practice (for the reasons of this, see, e.g. [67], [68], [11], [27], [21]). The common approach is to make a thorough testing on selected test problems and to draw conclusions upon the test results (see, e.g. [57], [58], [65]). Clearly, these conclusions depend on the lucky selection of the test problems and in any case require a huge amount of extra work. The idea of some kind of automatization arose quite early. The e ective tool of linearizing the e ects of roundo errors and using the computational graph have been widely applied since the middle of the 70 s ([40], [41], [42], [43], [44], [45], [46], [61], [35], [60]). Concerning the computational graph related error analysis Chaitin-Chatelin and Frayssé [11] gave the following summary. The stability analysis of an algorithm x = G (y) with the implementation G a lg = G (i) depends on the j computed at the various nodes of the computational graph (see 2.6). The partial derivatives show how inaccuracies in data and rounding errors at each step are propagated through the computation. The analysis can be conducted in a forward (or bottom-up) mode or a backward (or top-down) mode. There have been several e orts to automate this derivation. One can cite the following: 1. Miller and Spooner (1978); 2. the B-analysis: Larson, Pasternak, and Wisniewski (1983), Larson and Sameh (1980); 3. the functional stability analysis: Rowan (1990); 4. the automatic di erentiation (AD) techniques: Rall (1981), Griewank (1989), Iri (1991) Miller developed his approach in several papers ([37], [38], [39], [40], [41], [42], [43], [44], [45], [46]). The basic idea of Miller s method is the following. Given a numerical algorithm to analyze, a number! (d) is associated with each set d 2 (d 2 R n ) of input data. The function! : R n! R measures rounding error, i.e.,! (d) is large exactly when the algorithm applied to d produces results which are excessively sensitive to rounding errors. A numerical maximizer is applied to search for large values of! to provide information about the numerical properties of the algorithm. Finding a large value of! can be interpreted as the given numerical algorithm is su ering from a speci c kind of instability. The software performs backward error analysis. The value! (d) u (where u is the machine rounding unit) can be interpreted as the rst order approximation of the upper bound for the backward error. The computation of the error measuring number is based on the partial derivatives of the output with respect to the input and the individual rounding errors. An automatic di erentiation algorithm is used to provide the necessary derivatives. The Miller algorithm was implemented in Fortran language (actually in FOR- TRAN IV) by Webb Miller and David Spooner [44], [45] in More information on the use of the software by Miller and its theoretical background can be found in [40], [41], [42] and [44]. The software is in the ACM TOMS library with serial number 532 [45]. Miller and Wrathall published a book [46] on the construction, background and use of the software in which the potential of the software is clearly demonstrated through 14 case studies. The answers of the software are consistent in these cases with the well known formal analytical and experimental results. Miller s approach was further developed by Larson, Pasternak, and Wisniewski [35], Rowan [60], Bliss [5] and Bliss et al [6]. Upon the basis of corresponding literature, the Miller approach seems to be the most advanced although Miller s method has several setbacks. The numerical method to be analyzed must be expressed in a special, greatly simpli ed Fortranlike language. We can construct for-loops and if-tests that are based on the values of integer expressions. There is no way of conversion between real and integer types, and no mixed expressions (that contains both integer and real values) are allowed. Hence we can de ne only straight-line programs, i.e., where the ow of control does not depend on the values of oating point variables. To analyze methods with iterative loops and with branches on oating point values, the possible paths through any comparisons must be treated separately. This can be realized by constrained optimization. We con ne search for maximum to those input vectors by which the required path of control is realized. The constraints can be speci ed through a user-de ned subroutine. Higham [27] points out that the special language and its restrictions as the greatest disadvantage of the software: ... yet the software has apparently not been widely used. This is probably largely due to the inability of the software to analyse algorithms expressed in For- 3 tran, or any other standard language . Unfortunately, this can be said more or less on the other developments that followed the Miller approach. 2 The aims of the research The aim of my research work is to improve Miller s method to a level that meet today s requirements. I improved and upgraded Miller s method in two main steps. The rst step was a F77 version usable in PC environment with standard F77 compilers such as the GNU and Watcom compilers. This version was able to handle algorithms with a maximum of 3000 inputs, and a maximum of 1000 outputs, and operations of a maximum of while the original Miller program can handle a maximum of 30 inputs, 20 outputs and 300 operations. However even this new version used the simpli ed Fortran like language of Miller which is considered as a major problem by Higham and others. Using the recently available techniques such as automatic di erentiation, object oriented programming and the widespread use of MATLAB I have eliminated the above mentioned drawbacks of Miller method by creating a Matlab interface. Applying the operator overloading based implementation technique of automatic di erentiation Griewank [25] and Bischof etal [4] we have provided means of analyzing numerical methods given in the form of Matlab m-functions. In our framework, we can de ne both straight-line programs and methods with iterative loops and arbitrary branches. Since the possible control paths are handled automatically, iterative methods and methods with pivoting techniques can also be analyzed in a convenient way. Miller originally used the direct search method of Rosenbrock for nding numerical instability. To improve the e ciency of maximizing, we added two more direct search methods [7]: the well known simplex method of Nelder and Mead, and the so called multidirectional search method developed by Torczon [63]. In the thesis we present a signi cantly improved and partially reconstructed Miller method by designing and developing a new Matlab package for automatic roundo error analysis. Our software provides all the functionalities of the work by Miller and extends its applicability to such numerical algorithms that were complicated or even impossible to analyze with Miller s method before. Since the analyzed numerical algorithm can be given in the form of a Matlab m- le, our software is easy to use. 4 3 The methods of investigation The investigations and software development required the knowledge and use of the following areas the theory of programming, theory of compilers, computational graphs, the theory and practice of automatic di erentiation, backward error analysis theory of Wilkinson [67], [68] (see, also [27]), numerical analysis, nondi erentiable optimization methods object oriented programming. 4 The new scienti c results The numerical stability of computational algorithms is a very important issue. The classic error analysis techniques very rarely give computationally feasible results for the practitioner. The classical numerical testing on selected test examples often misleading depending on the selection and properties of the test problems. The purpose of the research is to provide an easy to use automatic error analysis based on compiler techniques, the computational graph techniques, the automatic di erentiation techniques, object oriented programming. It was a de nite aim that by simply using the written program of an algorithm under consideration the average or any other user can have a reliable estimate of the numerical stability without blind test problem selection. This line of research was initiated by W. Miller, who developed the most advanced program system and its theory during the 70s. Although many researcher developed similar or partly similar systems none of them achieved the high level of Miller s solution. Miller s solution however was limited in use by the computer technique of his age. In this thesis I analyzed, improved, upgraded and reimplemented his method. The results of the thesis can summarized as follows. 1. I replaced the minicompiler and its simpli ed programming language of the Miller method to object oriented Matlab. 5 2. Upon the bases of computer testing and theory I added two new optimization methods to the system that improved the performance of the software. 3. I reprogrammed and tested the system in Matlab. The new software provides all the functionalities of the work by Miller and extends its applicability to such numerical algorithms that were complicated or even impossible to analyze with Miller s method before. The analyzed numerical algorithm can be given in the form of a Matlab m- le. Hence our software is easy to use. The program consists of about lines and it is downloadable from the site http : ==phd:uni obuda:hu=images=milleranalyzer:zip together with a detailed user guide. 4. I applied the new Matlab version to investigate the numerical stability of some ABS methods (implicit LU, Huang and its variants), and three fast matrix multiplication algorithms. The obtained results indicate numerical instability of various scale and in the case of fast matrix multiplication algorithms give a de nite yes for the suspected numerical instability of these methods. 5 The applicability of the results In order to demonstrate the usability and e ciency of the developed software, we used it to examine the stability of some ABS methods [1], [2], namely the implicit LU methods and several variants of the Huang method [2]. The obtained computational results agreed with the already known facts about the numerical stability of the ABS algorithms. The program has shown that implicit LU is numerically unstable and that the modi ed Huang method has better stability properties than the original Huang method and the famous MGS (modi ed Gram-Schmidt) method. We also tested three famous fast matrix multiplication algorithms with the following results: The classical Winograd scalar product based matrix multiplication algorithm of O (n 3 ) operation cost is highly unstable in accordance with the common belief, that has never been published. Both the Strassen and Winograd recursive matrix multiplication algorithms of O (n 2:81 ) operation costs are numerically unstable. The comparative testing indicates that the numerical stability of Strassen s algorithm is somewhat better than those of Winograd. 6 The obtained results support the common but disputed opinion that these fast matrix multiplication methods are numerically unstable (for reference, see, e.g. Higham [27]). Upon the basis of our testing, we may think that the new software called Miller Analyzer for Matlab will be useful for numerical people or algorithm developers to analyze the e ects of rounding errors. 7 6 List of references [1] Aba y, J., Broyden, C. G., Spedicato, E., A class of direct methods for linear systems, Numerische Mathematik, 45, 1984, [2] Aba y J., Spedicato, E.: ABS Projection Algorithms: Mathematical Techniques for Linear and Nonlinear Equations, Ellis Horwood, 1989 [3] Bini, D., Lotti, G.: Stability of fast algorithms for matrix multiplication, Numerische Mathematik, 36, 1980, [4] Bischof, C.H., Bücker, H.M., Hovland, P., Naumann, U., Utke, J.(eds.): Advances in Automatic Di erentiation, Springer, 2008 [5] Bliss, B.: Instrumentation of FORTRAN programs for automatic roundo error analysis and performance evaluation, MSc thesis, University of Illinois at Urbana-Champaign, 1990 [6] Bliss, B., Brunet, M.-C., Gallopoulos, E.: Automatic parallel program instrumentation with applications in performance and error analysis, In Expert Systems for Scienti c Computing, E. N. Houstis, J. R. Rice, and R. Vichnevetsky, editors, North-Holland, Amsterdam, The Netherlands, 1992, [7] Brent, R.P.: Algorithms for matrix multiplication, Technical Report STAN- CS , Computer Science Department, Stanford University, 1970 [8] Brent, R.P.: Error analysis of algorithms for matrix multiplication and triangular decomposition using Winograd s identity, Numerische Mathematik, 16, 1970, [9] Byatt, D.: Convergent variants of the Melder-Nead algorithm, MSc thesis, University of Canterbury, 2000 [10] Castano, B., Heintz, J., Llovet, J., Martinez, R.: On the data structure straight-line program and its implementation in symbolic computation, Mathematics and Computers in Simulation, Volume 51, Number 5, February 2000, pp (32) [11] Chaitin-Chatelin, F., Frayssé, V.: Lectures on Finite Precision Computations, SIAM, Philadelphia, 1996 [12] Coleman, T.F., Verma, A.: ADMAT: An Automatic Di erentiation Toolbox for MATLAB, Computer Science Department, Cornell University, [13] Conn, A.R., Scheinberg, K. Vicente, L.N.: Introduction to derivative-free optimization, SIAM, Philadelphia, 2008 [14] Coppersmith, D., Winograd, S.: Matrix multiplication via arithmetic progressions, Journal of Symbolic Computation, 9, 1990, [15] Cormen, T.H., Leiserson, C.E., Rivest, R.L., Stein, C.: Introduction to Algorithms, 2nd edition, The MIT Press, McGraw Hill, Cambridge, 2001 [16] D Alberto, P., Nicolau, A.: Adaptive Strassen s matrix multiplication, In Proceedings of the 21st Annual International Conference on Supercomputing, ACM, New York, NY, 2007, [17] D Alberto, P., Nicolau, A.: Adaptive Winograd s matrix multiplications, ACM Transactions on Mathematical Software, Vol. 36, No. 1, Article 3, DOI / [18] Demmel, J., Dumitriu, I., Holtz, O., Kleinberg, R.: Fast matrix multiplication is stable, Numerische Mathematik, 106, 2007, , DOI /s [19] Douglas, C.,, Heroux, M., Slishman, G., Smith, R.M.: GEMMW: A Portable Level 3 BLAS Winograd Variant of Strassen s Matrix-Matrix Multiply Algorithm, J. Comp. Phys., 110, 1994, 1 10 [20] Dumitrescu, B.: Improving and estimating the accuracy of Strassen s algorithm, Numerische Mathematik, 79, 1998, [21] Einarsson, B. (ed): Accuracy and reliability in scienti c computing, SIAM, Philadelphia, 2005 [22] Forth, S.A.: An e cient overloaded implementation of forward mode automatic di erentiation in MATLAB, ACM Trans. Math. Softw., 32, 2006, [23] Gao, F., Han, L.: Implementing the Nelder-Mead simplex algorithm with adaptive parameters, Comput. Optim. Appl., DOI /s [24] Golub, G.H., Van Loan, C.F.: Matrix Computations, 2nd ed., Johns Hopkins University Press, 1989 [25] Griewank, A.: Evaluating Derivatives: Principles and Techniques of Algorithmic Di erentiation, SIAM, [26] Higham, N. J.: Exploiting Fast Matrix Multiplication Within the Level 3 BLAS, ACM Transactions on Mathematical Software, 16, 4, 1990, [27] Higham, N. J.: Accuracy and Stability of Numerical Algorithms, SIAM, 1996 [28] Kaporin, I.: A practical algorithm for faster matrix multiplication, Numer. Linear Algebra Appl., 6, 1999, [29] Kaporin, I.: The aggregation and cancellation techniques as a practical tool for faster matrix multiplication, Theoretical Computer Science 315, 2004, [30] Kolda, T.G., Lewis, R.M., Torczon, V.: Optimization by Direct Search: New Perspectives on Some Classical and Modern Methods, SIAM Review, 45, 2003, [31] Krämer, W.: Constructive Error Analysis, Journal of Universal Computer Science, vol. 4, no. 2 (1998), [32] Krämer, W., Bantle, A.: Automatic Forward Error Analysis for Floating Point Algorithms, Reliable Computing 7, 2001, [33] Lagarias, J.C., Reeds, J.A., Wright, M.H., Wright, P. E.: Convergence Properties of the Nelder Mead Simplex Method in Low Dimensions, SIAM J. Optim., 9 (1999) [34] Lakshmivaran, S., Dhall, S.K.: Analysis and Design of Parallel Algorithms, McGraw-Hill, 1990 [35] Larson, J.L., Pasternak, M.E.,Wisniewski, J.A.: Algorithm 594: Software for Relative Error Analysis, ACM Trans. Math. Softw., 9, 1983, [36] Lewis, R.M., Torczon, V., Trosset, M.W.: Direct search methods: then and now, Journal of Computational and Applied Mathematics 124 (2000) [37] Miller, W.: On the stability of nite numerical procedures, Numerische Mathematik, 19, 1972, [38] Miller, W.: Computational complexity and numerical stability, in Proceeding STOC 74 Proceedings of the sixth annual ACM symposium on Theory of Computing, 1974, [39] Miller, W.: Computational complexity and numerical stability, SIAM J. Comput., 4, 2, 1975, [40] Miller, W.: Computer search for numerical instability, JACM, 22, 4, 1975, [41] Miller, W.: Software for roundo analysis, ACM Trans. Math. Softw., 1, 2, 1975, [42] Miller, W.: Roundo analysis by direct comparison of two algorithms, SIAM Journal on Numerical Analysis, 13, 3, 1976, [43] Miller, W.: Roundo Analysis and Sparse Data, Numerische Mathematik, 29, 1977, [44] Miller, W., Spooner, D.: Software for roundo analysis, II, ACM Trans. Math. Softw., 4, 4, 1978, [45] Miller, W., Spooner, D.: Algorithm 532: software for roundo analysis [Z], ACM Trans. Math. Softw., 4, 4, 1978, [46] Miller, W., Wrathall, C.: Software for roundo analysis of matrix algorithms, Academic Press, New York, 1980 [47] Monte Leon, V.J.: Automatic error analysis in nite digital computations, using range arithmetic, MSc thesis, U.S. Naval Postgraduate School, Monterey, California, 1966 [48] Moore, R.E.: Automatic Error Analysis in Digital Computation, Technical report, LMSD-48421, Lockheed, 28 January, 1959 [49] Moore, R.E.: Methods and Applications of Interval Analysis, SIAM, Philadelphia, 1979 [50] Muller, J.-M., et al.: Handbook of Floating-Point Arithmetic, Birkhäuser, 2010 [51] Mutrie, M.P.W., Bartels, R.H., Char, B.W.: An approach for oating-point error analysis using computer algebra, Proceeding ISSAC 92 Papers from the international symposium on Symbolic and algebraic computation, ACM New York, NY, USA, 1992, [52] Nelder, J.A., Mead, R.: A simplex method for function minimization, Computer Journal, 7 (1965) [53] Overton, M.L.: Numerical computing with IEEE oating point arithmetic, SIAM, Philadelphia, [54] Pan, V.: How can we speed up matrix multiplication?, SIAM Review, 26, 3, 1984, [55] Rall. L.B.: Automatic Di
Related Search
We Need Your Support
Thank you for visiting our website and your interest in our free products and services. We are nonprofit website to share and download documents. To the running of this website, we need your help to support us.

Thanks to everyone for your continued support.

No, Thanks

We need your sign to support Project to invent "SMART AND CONTROLLABLE REFLECTIVE BALLOONS" to cover the Sun and Save Our Earth.

More details...

Sign Now!

We are very appreciated for your Prompt Action!