Tuesday, June 4, 2019
Solving Large Systems of Linear Simultaneous Equations
Solving Large Systems of Linear Simultaneous EquationsNICOLE LESIRIMAMETHODS OF SOLVING  broad SYSTEMS OF LINEAR SIMULTANEOUS EQUATIONSPROJECT DESCRIPTIONLinear  agreements simulate real-world problems using  utilize numerical procedure. The main aim of this project is to consider what factors affect the  strength of the mixed  manners of solving  one-dimensional simultaneous equations. So far, one of the main factors is rounding errors that can produce inaccurate  results. Moreover, MATLAB programs have been produced to time the calculation  revive to determine the efficiency of the methods. Generally, these methods  be subdivided into two direct and iterative methods. Direct methods  be commonly  utilize to solve small systems of equations. The iterative methods are used to solve real-world problems that produce systems of equations for which the coefficient matrices are sparse.The relevance of studying these methods have its real world applications. The real world applications can    be seen in various  handle such as science and engineering, accounting and finance, business management and in operational research. The approach provides a logical framework for solving complex decisions in a  big range of industries. The advantage is that, decisions are founded on data analysis.Environmentalists and meteorologists may use  bombastic systems of simultaneous linear equations to   verbal expression to future outcomes. For instance, to predict weather patterns or climate change, a large volume of data is collected over a long span of time on  numerous variables including, solar radiation, carbon emissions and ocean temperatures. Queen Mary University of London (2015). This data is represented in the form of a transition  hyaloplasm that has to be  course of instruction  centerd into a probability   ground substance that can then be used in the prediction of climate change.The objective of an enterprise is to maximize returns  sequence maintaining minimum costs. Where   as the use of large systems of simultaneous linear equations may provide a basis for evidence based business decision making in an enterprise, it is  authorised to know which linear systems are most appropriate in order to minimize undesirable outcomes for an enterprise.PROJECT REPORT OUTLINEChapter 1 aditLarge systems of linear simultaneous equations are used to simulate real-world problems using applied numerical procedure. The real world applications can be seen in various fields such as science and engineering, accounting and finance, business management. The approach provides a logical framework for solving complex decisions in a wide range of industries. The advantage is that decisions are founded on data analysis. The aim of this project is to explore the efficiency of a large systems of linear simultaneous equations in the  best decision making of an enterprise.Chapter 2Direct Methods Gaussian Elimination and LU FactorisationDirect methods of solving linear simultaneous equa   tions are introduced. This chapter will look at the Gaussian Elimination and LU Factorisation methods. Gaussian Elimination involves representing the simultaneous equations in an augment form, performing  mere(a) row operations to reduce the upper   angular form and finally back substituting to form the solution vector. LU Factorisation on the other hand is where a matrix A finds a lower triangular matrix L and an upper triangular matrix U such that A = LU. The purpose of this lower triangular matrix and upper triangular matrix is so that the  previous and backward substitutions can be directly applied to these matrices to obtain a solution to the linear system. An operation count and computing times using MATLAB is calculated so as to determine the best method to use.Chapter 3Cholesky FactorisationIntroduction to the Cholesky method. This is a procedure whereby the matrix A is factorised into the product of a lower triangular matrix and its transpose the forward and backward substi   tutions can be directly applied to these matrices to obtain a solution. A MATLAB program is written to compute timings. A conclusion can be drawn by comparing the three methods and  ascertain which is the most suitable method that will produce the most accurate result as well as take the shortest computing time.Chapter 4iterative Methods Jacobi Method and Gauss-SeidelThis chapter will introduce the iterative methods that are used to solve linear systems with coefficient matrices that are large and sparse. Both methods involve splitting the matrix A into lower triangular, diagonal and upper triangular matrices L, D, U respectively. The main difference comes down to the way the x values are calculated. The Jacobi method uses the previous x values (n) to calculate the next iterated x values (n+1). The Gauss-Seidel uses the new x value (n+1) to calculate the x2 value.Chapter 5Successive Over Relaxation and Conjugate  sideOther iterative methods are introduced. The Successive Over Relaxa   tion method over relaxes the solution at each iteration. This method is calculated using the  weighted sum of the values from the previous iteration and the values form the Gauss-Seidel method at the current iteration. The Conjugate Gradient method involves improving the approximated value of xk to the  postulate solution which may be reached after a finite  come up of iterations usually smaller than the size of the matrix.Chapter 6ConclusionAll the project   determinations and results are summarised in this chapter. Conclusion can be made from both direct methods and iterative methods whereby the most accurate method with the shortest computing time can be found. Drawbacks from each method will be mentioned as well its suitability for solving real world problems.PROGRESS TO DATEThe project to date has cover the direct methods of solving simultaneous equations.Gaussian EliminationThis involves representing the simultaneous equations in an augmented form, performing elementary row op   erations to reduce the upper triangular form and finally back substituting to form the solution vector. For example, to solve an mxn matrixAx = bThe aim of the Gaussian elimination is to manipulate the augmented matrix Ab using elementary row operations by adding a multiple of the pivot rows to the rows beneath the pivot row i.e. Ri Ri +kRj. Once the augmented matrix is in the row echelon form, the solution is found using back substitution.The following general matrix equation has been reduced to row echelon formThis corresponds to the linear systemRearranging the final solution is given byFor all other equations i = n  1, . . .,The operation count and timing the Gaussian Elimination was performed. The total number of operations for an nxn matrix using the Gaussian elimination is with O(N3).LU FactorisationThis is where a matrix A finds a lower triangular matrix L and an upper triangular matrix U such that A = LU. The purpose of this lower triangular matrix and upper triangular matr   ix is so that the forward and backward substitutions can be directly applied to these matrices to obtain a solution to the linear system.In general,L and U is an m x n matrixL = U = For higher order matrices, we can derive the calculation of the L and U matrices. Given a set of n elementary matrices E1, E2,, Enapplied to matrix A, row reduce in row echelon form without permuting rows such that A can be written as the product of two matrices L and U that isA = LU,WhereU = EnE2E1A,L = E1-1 E2-1En-1 For a general nxn matrix, the total number of operations is O(N3). A Matlab program has been produced to time the LU Factorisation. So far, this method has proven more efficient than the Gaussian Elimination.Cholesky FactorisationThis is a procedure whereby the matrix A is factorised into the product of a lower triangular matrix and its transpose i.e. A = LLT or =  The Cholesky factorisation is only possible if A is a positive definite. Forward and backward substitution is employed in findi   ng the solutions.The method was also timed at it can be concluded that it is the most effective and efficient direct method for solving simultaneous equations.The  verificatory methods have been introduced with a short outline of what each method entails.Work Still to be Completed  As from the objectives layed out from the terms of reference, the following are the objectives that are yet to be completed.Week 13  16 Evaluating the convergence rate of the iterative methods in detail as well as finding out which method improves the solution efficiency. Production of MATLAB programs analysing the  distinct methods and other methods. Over the next 3 weeks, the conditions for convergence will be analysed. One of the most important conditions that will be studied is the spectral radius. This is a condition applied on the indirect methods to determine how fast or slow a method takes to achieve the state of convergence. Moreover, the project will also produce Matlab programs for the iterativ   e methods and employ the spectral radius on these programs to determine the speed of convergence for large sparse matrices.Weeks 17  19 Introduction to the Successive Over-Relaxation (SOR) method and the Conjugate Gradient method. Successive Over-Relaxation method improves the rate of convergence of the Gauss-Siedel method by over-relaxing the solution at every iteration. While the Conjugate Gradient improves the approximated value of x to the exact solution. Matlab programs will be produced for the two methods together with the speed of convergence of different sizes of matrices.Week 20  24Writing the findings and conclusions of the report, finalising on the bibliography and doing a review of the project as a whole. Preparing oral and poster presentation.  
Subscribe to:
Post Comments (Atom)
 
 
No comments:
Post a Comment
Note: Only a member of this blog may post a comment.