Difference between revisions of "MAT3633"

From Department of Mathematics at UTSA
Jump to navigation Jump to search
 
(15 intermediate revisions by 2 users not shown)
Line 6: Line 6:
 
==Topics List==
 
==Topics List==
 
{| class="wikitable sortable"
 
{| class="wikitable sortable"
! Date !! Sections !! Topics !! Prerequisite Skills !! Student Learning Outcomes
+
! Date !! Sections [Sauer 3rd ed] !! Topics !! Prerequisite Skills !! Student Learning Outcomes
 
|-
 
|-
|Week 1
+
| Week.1
 
||
 
||
Section 0.2 & 1.1
+
0.2 and 1.1
 +
||
 +
* Loss of significant digits
 +
* Bisection Method
 +
* Brief introduction to matlab
 +
||
 +
* binary number system;
 +
* Taylor's theorem;
 +
* intermediate value theorem
 +
||
 +
* Nested multiplication for evaluating polynomials
 +
* Machine representation of real numbers
 +
* Loss of significant digits in numerical computing
 +
* Review of Taylor's Theorem
 +
* Bisection method and implementation
 
||
 
||
* [[Loss of Significant Digits]]
 
* [[Nested Multiplication for Evaluating Polynomials]]
 
* [[Machine Representation of Real Numbers]]
 
* [[Loss of Significant Digits in Numerical Computing]]
 
* [[Review of Taylor's Theorem]]
 
  
* [[Bisection Method]]
+
|-
* [[Bisection Method and Implementation]]
+
| Week.2
* [[Brief Introduction to Matlab]]
 
 
||
 
||
* [[Binary Number System]]
+
1.2 and 1.3
* [[Taylor's Theorem]]
 
* [[Intermediate Value Theorem]]
 
 
||
 
||
* (To be Entered)
+
* Fixed-Point Iteration
|-
+
* Limits of Accuracy: Conditioning of problems
|Week 2
 
 
||
 
||
Sections 1.2 & 1.3
+
* limit of sequences
 +
* multiplicity of solution of equations.
 
||
 
||
* [[Fixed-Point Iteration]]
+
* Geometric interpretation
* [[Geometric Interpretation]]
+
* Convergence of fixed point iterations
* [[Convergence of Fixed Point Iterations]]
+
* Order of convergence of iterative methods
* [[Order of Convergence of Iterative Methods]]
+
 
 +
* Wilkinson polynomial and other examples
 +
* Sensitivity analysis of root-finding
 +
* Error magnification factor for solution of equations
  
* [[Limits of Accuracy: Conditioning of Problems]]
+
|-
* [[Wilkinson Polynomial]]
+
| Week.3
* [[Sensitivity Analysis of Root-Finding]]
 
* [[Error Magnification Factor for Solution of Equations]]
 
 
||
 
||
* [[Limit of Sequences]]
+
1.4 and 1.5
* [[Solution Multiplicity of Equations]]
 
 
||
 
||
* (TBD)
+
* Newton's Method
|-
+
* Root-Finding without Derivatives
|Week 3
 
 
||
 
||
Sections 1.4 & 1.5
+
* Remainder of Taylor's series
 +
* intermediate value theorem.
 
||
 
||
* [[Newton's Method]]
+
* Algebraic and geometric interpretation of Newton's method
* [[Algebraic and Geometric Interpretation of Newton's method]]
+
* Error analysis for Newton's method based on Taylor's theorem
* [[Error Analysis for Newton's Method Based on Taylor's Theorem]]
+
* Newton's method as a fixed point iteration
* [[Newton's Method as a Fixed Point Iteration]]
+
* Modified Newton's method and its rate of convergence
* [[Modified Newton's Method and its Rate of Convergence]]
 
  
* [[Root-Finding Without Derivatives]]
+
* Secant Method and its convergence,
* [[Secant Method and its Convergence]]
+
* Method of False Position, Muller's Method:
* [[Method of False Position, Muller's Method]]
+
* Stopping criteria for iterative methods
* [[Stopping Criteria for Iterative Methods]]
+
 
 +
 
 +
|-
 +
| Week.4
 
||
 
||
* [[Remainder of Taylor's Series]]
+
2.1 and 2.2
* [[Intermediate Value Theorem]]
 
 
||
 
||
* (TBD)
+
* Solve Systems of Linear Equations: Gaussian Elmination
|-
+
* Solve System of Linear Equations: LU Decomposition
|Week 4
 
 
||
 
||
Sections 2.1 & 2.2
+
* Matrix-matrix products and matrix-vector products
 +
* inverse matrix
 +
* elementary row operations
 +
* product and inverse of matrices for elementary row operations.
 
||
 
||
* [[Solve Systems of Linear Equations: Gaussian Elimination]]
+
* Gaussian elimination and its operation counts
* [[Gaussian Elimination and its Operation Counts]]
+
* Gaussian elimination with pivoting
* [[Gaussian Elimination with Pivoting]]
+
* Implementation of Gauss elimination
* [[Implementation of Gauss Elimination]]
 
  
* [[Solve System of Linear Equations: LU Decomposition]]
+
* Matrices for elementary row operations
* [[Matrices for Elementary Row Operations]]
+
* Gauss elimination as matrix products
* [[Gauss Elimination as Matrix Products]]
+
* Advantages of solutions by LU decomposition
* [[Advantages of Solutions by LU Decomposition]]
+
 
 +
|-
 +
| Week.5
 
||
 
||
* [[Matrix-Matrix Products]]
+
2.3 and 2.4
* [[Matrix-Vector Products]]
 
* [[Inverse Matrix]]
 
* [[Elementary Row Operations]]
 
 
||
 
||
* (TBD)
+
* Error Analysis for Solution of Ax=b
|-
+
* Iterative Methods for Solving Ax=b
|Week 5
 
 
||
 
||
Sections 2.3 & 2.5
+
* Length of vectors
 +
* eigenvalue and eigenvectors of matrix
 
||
 
||
* [[Error Analysis for Solution of Ax=b]]
+
* various norms for vectors and matrices: compatibility of vector and matrix norms.
* [[Various Norms for Vectors and Matrices: Compatibility of Vector and Matrix Norms]]
+
* Error Analysis for the solution of Ax=b
* [[Error Analysis for Solution of Ax=b]]
+
* Error magnification factor and condition number of matrix
* [[Error Magnification Factor and Condition Number of Matrix]]
 
  
* [[Iterative Methods for Solving Ax=b]]
+
* Jacobi method, Gauss-Seidel method, Successive-Over-Relaxation (SOR) method
* [[Jacobi Method]]
+
* Convergence of Jacobi Method, GS method and SOR method.
* [[Gauss-Seidel Method]]
+
* spectral radius of matrix
* [[Successive-Over-Relaxation (SOR) Method]]
+
* convergence of general iterative method for solving system of linear equations,
* [[Convergence of Iterative Methods]]
+
* Sparse Matrix
* [[Spectral Radius of Matrix]]
+
* Comparison of Gauss Elimination and iterative methods
* [[Convergence of General Iterative Method for Solving System of Linear Equations]]
+
 
* [[Sparse Matrix]]
+
 
* [[Comparison of Gauss Elimination and Iterative Methods]]
+
|-
 +
| Week.6
 
||
 
||
* [[Length of Vectors]]
+
2.6 and 2.7
* [[Eigenvalues of a Matrix]]
 
* [[Eigenvectors of a Matrix]]
 
 
||
 
||
* (TBD)
+
* Conjugate Gradient Method
|-
+
* Nonlinear System of Equations
|Week 6
 
 
||
 
||
Sections 2.6 & 2.7
+
* scalar product of vectors
 +
* determinant and eigenvalues of matrix
 +
* quadratic polynomials of n-variables
 +
* partial derivatives and gradients
 +
* chain rule for partial derivatives.
 
||
 
||
* [[Conjugate Gradient Method]]
+
* Symmetric positive definite matrix and properties
* [[Symmetric Positive Definite Matrix and Properties]]
+
* Construction of Conjugate Gradient (CG) Method
* [[Construction of Conjugate Gradient (CG) Method]]
+
* Propertise of CG Method
* [[Properties of CG Method]]
+
* Preconditioning for CG method
* [[Preconditioning for CG Method]]
 
  
* [[Nonlinear System of Equations]]
+
* Taylor's Theorem for multi-variate vector valued functions:
* [[Taylor's Theorem for Multi-Variate Vector Valued Functions]]
+
* Newton's Method:
* [[Newton's Method]]
+
* Broyden's Method
* [[Broyden's Method]]
+
 
 +
|-
 +
| Week.7
 
||
 
||
* [[Scalar Product of Vectors]]
+
3.1 and 3.2
* [[Determinant of a Matrix]]
 
* [[Eigenvalues of a Matrix]]
 
* [[Quadratic Polynomials of n-variables]]
 
* [[Partial Derivatives]]
 
* [[Gradients]]
 
* [[Chain Rule for Partial Derivatives]]
 
 
||
 
||
* (TBD)
+
* Data and Interpolating Functions
|-
+
* Interpolation Error and Runge Phenomenon
|Week 7
+
* Chebyshev interpolation
 
||
 
||
Sections 3.1 & 3.2
+
* Fundamental theorem of algebra
 +
* Rolle's theorem.
 
||
 
||
* [[Data and Interpolating Functions]]
+
* Lagrange Basis Functions:
* [[Lagrange Basis Functions]]
+
* Properties of Lagrange basis functions:
* [[Properties of Lagrange Basis Functions]]
+
* Lagrange form of the interpolation polynomials
* [[Lagrange Form of the Interpolation Polynomials]]
+
 
* [[Newton's Divided Differences]]
+
* Newton's Divided Differences:
* [[Properties of Newton's Divided Differences]]
+
* Properties of Newton's divided differences:
* [[Newton's Form of the Interpolation Polynomials]]
+
* Newton's Form of the interpolation polynomials
 +
 
 +
* Interpolation error analysis
 +
* Runge phenomenon
 +
 
 +
* Chebyshev Polynomial
 +
* Error estimates for Chebyshev interpolation
  
* [[Interpolation Error and Runge Phenomenon]]
+
 
* [[Interpolation Error Analysis]]
+
|-
* [[Runge Phenomenon]]
+
| Week.8
* [[Chebyshev Polynomial]]
 
* [[Error Estimates for Chebyshev Interpolation]]
 
 
||
 
||
* [[Fundamental Theorem of Algebra]]
+
3.4, 3.5 and 4.1
* [[Rolle's Theorem]]
 
 
||
 
||
* (TBD)
+
* Cubic Splines
|-
+
* Bezier Curves
|Week 8
+
* Least Square Method
 
||
 
||
Sections 3.4, 3.5, & 4.1
+
* one-sided limits
 +
* continuity of functions
 +
* indefinite integrals
 +
* extremum values of multivariate quadratic functions.
 
||
 
||
* [[Cubic Splines]]
+
* Cubic splines
* [[Cubic Splines]]
+
* construction of cubic splines for interpolation
* [[Construction of Cubic Splines for Interpolation]]
+
* end conditions
* [[End Conditions]]
+
* properties of cubic spline interpolation
* [[Properties of Cubic Spline Interpolation]]
 
  
* [[Bezier Curves]]
+
* Bezier Curve and fonts
* [[Bezier Curve and Fonts]]
 
  
* [[Least Square Method]]
+
* Least square method for solving inconsistent system of linear equations.
* [[Least Square Method for Solving Inconsistent System of Linear Equations]]
+
* Basic properties of least square solutions:
* [[Basic Properties of Least Square Solutions]]
+
 
 +
 
 +
|-
 +
| Week.9
 
||
 
||
* [[One-Sided Limits]]
+
4.2 and 4.5
* [[Continuity of Functions]]
 
* [[Indefinite Integrals]]
 
* [[Extremum Values of Multivariate Quadratic Functions]]
 
 
||
 
||
* (TBD)
+
* Mathematical Models and Data Fitting
|-
+
* Nonlinear Least Square Fitting
|Week 9
 
 
||
 
||
Sections 4.2 & 4.5
+
* linear spaces, basis functions
 +
* product rule and chain rule for vector valued multivariate functions.
 
||
 
||
* [[Mathematical Models and Data Fitting]]
+
* Least square method for curve fitting and statistical modeling.
* [[Least square method for curve fitting and statistical modeling]]
+
* Survey of Models: linear model, periodic model, exponential models, logistic model, etc
* [[Survey of Models]]: linear model, periodic model, exponential models, logistic model, etc
 
  
* [[Nonlinear Least Square Fitting]]
+
* Taylor's theorem for vector valued multivariate functions.
* [[Taylor's Theorem for Vector Valued Multivariate Functions]]
+
* Gauss-Newton Method
* [[Gauss-Newton Method]]
+
* Levenberg-Marquardt Method
* [[Levenberg-Marquardt Method]]
+
 
 +
 
 +
 
 +
|-
 +
| Week.10
 
||
 
||
* [[Linear Spaces]]
+
5.1, 5.2 and 5.3
* [[Basis Functions]]
 
* [[Product Rule for Vector Valued Multivariate Functions]]
 
* [[Chain Rule for Vector Valued Multivariate Functions]]
 
 
||
 
||
* (TBD)
+
* Numerical Differentiation
|-
+
* Numerical Integration: Newton-Cotes Formulas
|Week 10
+
* Numerical Integration: Romberg's Technique
 
||
 
||
Sections 5.1, 5.2, & 5.3
+
* Taylor's theorem
 +
* interpolation error estimates
 +
* properties of definite inetgrals
 
||
 
||
* [[Numerical Differentiation]]
+
* Finite difference (FD) approximations of 1st order derivative and their error analysis
* [[Finite difference (FD) Approximations of 1st order Derivative and Their Error Analysis]]
+
* FD approximations of 2nd order derivatives and their error analysis
* [[FD approximations of 2nd order Derivatives and Their Error Analysis]]
+
* Undetermined coefficient method for FD approximation
* [[Undetermined Coefficient Method for FD Approximation]]
+
* Extropolation technique for improving the order of approximation
* [[Extrapolation Technique for Improving the Order of Approximation]]
+
 
 +
* Midpoint rule, trapezoid rule and Simpson's rule;
 +
* Error analysis based on Taylor's Theorem and interpolation errors
 +
* Degree of precision of quadrature rules
 +
* Composite quadrature rules
 +
 
 +
* Motivation, construction and implementation of Romberg's technique.
  
* Numerical Integration: [[Newton-Cotes Formulas]]
 
* [[Midpoint rule]]
 
* [[Trapezoid rule]]
 
* [[Simpson's rule]]
 
* [[Error Analysis based on Taylor's Theorem]]
 
* [[Error Analysis based on Interpolation Errors]]
 
* [[Degree of Precision of Quadrature Rules]]
 
* [[Composite Quadrature Rules]]
 
  
* Numerical Integration: [[Romberg's Technique]]
 
* Motivation, construction and implementation of [[Romberg's Technique]].
 
||
 
* [[Taylor's Theorem]]
 
* [[Interpolation Error Estimates]]
 
* [[Properties of Definite Integrals]]
 
||
 
* (TBD)
 
 
|-
 
|-
|Week 11
+
| Week.11
 
||
 
||
Sections 5.4 & 5.5
+
5.4 and 5.5
 
||
 
||
* [[Adaptive Numerical Integration]]
+
* Adaptive Numerical Integration
* [[Implementation of Adaptive Numerical Integration Techniques]]
+
* Gauss Quadrature Formulas
 
 
* [[Gauss Quadrature Formulas]]
 
* [[Orthogonal Polynomials]]
 
* [[Legendre polynomials]]
 
* [[Gauss Quadrature Rule]]
 
 
||
 
||
* [[Long Divisions]]
+
* long divisions
* [[Substitution Methods]] for definite integrals
+
* changing variables for definite integrals
 
||
 
||
* How to estimate the error on a sub interval
+
* How to estimate the error on a subinterval
* How to mark sub intervals to be further refinement?
+
* How to mark subintervals to be further refinement?
 +
* Implementation of adaptive numerical integration techniques.
  
* Motivation and difficulties with straightforward approach
+
* Motivation and difficulties with straightforward approach.
* Legendre polynomials and their basic properties
+
* Orthogonal polynomials,
* Gauss Quadrature rule based on Legendre polynomials  
+
* Legendre polynomials and their basic properties;
 +
* Gauss quadrature rule based on Legendre polynomials
 
* Degree of precision of Gauss Quadrature
 
* Degree of precision of Gauss Quadrature
 
* Gauss quadrature formula on general interval and composite Gauss rules
 
* Gauss quadrature formula on general interval and composite Gauss rules
 +
 +
 
|-
 
|-
|Week 12
+
| Week.12
 +
||
 +
10.1 and 11.1
 +
||
 +
* Discrete Fourier Transform and FFT
 +
* Discrete Cosine Transform (optional)
 +
* Image Compression  (optional)
 
||
 
||
Sections 10.1, 11.1, & 11.2
+
* complex numbers and complex variables
 +
* integration by parts
 +
* convergence of sequences and series.
 
||
 
||
* [[Discrete Fourier Transform and Fast Fourier Transform (FTT)]]
+
* Fourier Series,
* [[Fourier Series]]
+
* Discrete Fourier Transform
* [[Discrete Fourier Transform]] (DFT)
+
* Matrix Form of Discrete Fourier Transform:
* [[Matrix Form of Discrete Fourier Transform]]
+
* Inverse Discrete Fourier Transform:
* [[Inverse Discrete Fourier Transform]]
+
* DFT and Trigonometric interpolation
* [[DFT and Trigonometric Interpolation]]
+
* Algorithm for computing DFT: Fast Fourier Transform (FFT)
* [[Fast Fourier Transform (FFT)]]
+
 
 +
*  Discrete Cosine Transform (DCT),
 +
*  DCT and Interpolation by Cosine Functions
 +
*  Relation between DFT and DCT:
 +
*  Fourier Transform of 2-Dimensional Functions
 +
*  DCT of 2-Dimensional Functions:
 +
*  Interpolation Theorem for 2-Dimensional DCT
 +
 
 +
*  Digital Gray scale images and color color images:
 +
*  RGB format:
 +
*  YCbCr (or YUV) format:
 +
*  Convertion between RGB and YUV formats:
 +
*  Quantization, Image Compression and Decompression
  
* [[Discrete Cosine Transform]](optional)
 
* [[Discrete Cosine Transform]](DCT)
 
  
* [[Image Compression]](optional)
 
* [[Quantization]]
 
* [[Image Compression]]
 
* [[Image Decompression]]
 
||
 
* [[Complex Numbers]]
 
* [[Complex Variables]]
 
* [[Integration by Parts]]
 
* [[Convergence of Sequences]]
 
* [[Convergence of Series]]
 
||
 
* DCT and Interpolation by Cosine Functions
 
* Relation between DFT and DCT
 
* Fourier Transform of 2-Dimensional Functions
 
* DCT of 2-Dimensional Functions
 
* Interpolation Theorem for 2-Dimensional DCT
 
  
* Digital Gray scale images and color color images
 
* RGB format
 
* YCbCr (or YUV) format
 
* Convertion between RGB and YUV formats
 
 
|-
 
|-
|Week 13
+
| Week.13
 
||
 
||
Sections 12.1 & 12.2
+
12.1 and 12.2
 
||
 
||
* [[Power Iteration Methods]]
+
* Power Iteration Methods
* [[Power Iteration Methods]]
+
* QR Algorithm for Computing Eigenvalues
* [[Convergence of Power Iteration Methods]]
 
* [[Inverse Power Iteration]]
 
* [[Inverse Power Iteration with Shift]]
 
* [[Rayleigh Quotient Iteration]]
 
 
 
* [[QR Algorithm for Computing Eigenvalues]]
 
* [[Orthogonal Matrices]]
 
* [[QR-Factorization]]
 
* [[Normalized Simultaneous Iteration]](NSI)
 
* [[Unshifted QR Algorithm]]
 
* [[Shifted QR Algorithm]]
 
 
||
 
||
* [[Eigenvalues]]
+
* properties of eigen values and eigenvectors
* [[Eigenvectors]]
+
* Gram-Schmidt orthogonalization
* [[Orthonormal Bases and the Gram-Schmidt Process]]
 
 
||
 
||
* Definition and basic properties of orthogonal matrices
+
* Power iteration and its rate of convergence.
* QR-Factorization based on Gram-Schmidt Orthogonalization
+
* Inverse Power Iteration,
 +
* Inverse Power Iteration with Shift
 +
* Rayleigh Quotient Iteration
 +
 
 +
 
 +
* Definition and basic properties of orthogonal matrices:
 +
* QR-Factorization based on Gram-Schmidt Orthogonalization:
 +
* Normalized Simultaneous Iteration (NSI).
 +
* Unshifted QR Algorithm:
 +
* Shifted QR Algorithm:
 +
 
 +
 
 +
 
 
|-
 
|-
|Week 14
+
| Week.14
 
||
 
||
Sections 12.2
+
12.2
 
||
 
||
* [[QR Algorithm for Computing Eigenvalues]]
+
* Algorithm for Computing Eigenvalues: Speed up of QR-algorithm:
* [[Upper Hessenberg Form]] (UHF)
 
* [[Householder Reflector]]
 
 
||
 
||
* [[Matrices for Orthogonal Projection]]
+
* matrices for orthogonal projection and reflection
* [[Matrices for Reflection]]
+
* block matrices and their products
* [[Block Matrices]]
+
* similar matrices.
* [[Similar Matrices]]
 
 
||
 
||
 +
* Upper Hessenberg form (UHF)
 +
* Householder Reflector
 
* Convert a matrix into UHF by Householder reflectors
 
* Convert a matrix into UHF by Householder reflectors
 +
 
|}
 
|}

Latest revision as of 14:38, 17 August 2020

Course Catalog

MAT 3633. Numerical Analysis. (3-0) 3 Credit Hours.

Prerequisites: MAT2233, MAT3213, and one of the following: CS1063, CS1714, or CS2073. Solution of linear and nonlinear equations, curve-fitting, and eigenvalue problems. Generally offered: Fall, Spring. Differential Tuition: $150.

Topics List

Date Sections [Sauer 3rd ed] Topics Prerequisite Skills Student Learning Outcomes
Week.1

0.2 and 1.1

  • Loss of significant digits
  • Bisection Method
  • Brief introduction to matlab
  • binary number system;
  • Taylor's theorem;
  • intermediate value theorem
  • Nested multiplication for evaluating polynomials
  • Machine representation of real numbers
  • Loss of significant digits in numerical computing
  • Review of Taylor's Theorem
  • Bisection method and implementation
Week.2

1.2 and 1.3

  • Fixed-Point Iteration
  • Limits of Accuracy: Conditioning of problems
  • limit of sequences
  • multiplicity of solution of equations.
  • Geometric interpretation
  • Convergence of fixed point iterations
  • Order of convergence of iterative methods
  • Wilkinson polynomial and other examples
  • Sensitivity analysis of root-finding
  • Error magnification factor for solution of equations
Week.3

1.4 and 1.5

  • Newton's Method
  • Root-Finding without Derivatives
  • Remainder of Taylor's series
  • intermediate value theorem.
  • Algebraic and geometric interpretation of Newton's method
  • Error analysis for Newton's method based on Taylor's theorem
  • Newton's method as a fixed point iteration
  • Modified Newton's method and its rate of convergence
  • Secant Method and its convergence,
  • Method of False Position, Muller's Method:
  • Stopping criteria for iterative methods


Week.4

2.1 and 2.2

  • Solve Systems of Linear Equations: Gaussian Elmination
  • Solve System of Linear Equations: LU Decomposition
  • Matrix-matrix products and matrix-vector products
  • inverse matrix
  • elementary row operations
  • product and inverse of matrices for elementary row operations.
  • Gaussian elimination and its operation counts
  • Gaussian elimination with pivoting
  • Implementation of Gauss elimination
  • Matrices for elementary row operations
  • Gauss elimination as matrix products
  • Advantages of solutions by LU decomposition
Week.5

2.3 and 2.4

  • Error Analysis for Solution of Ax=b
  • Iterative Methods for Solving Ax=b
  • Length of vectors
  • eigenvalue and eigenvectors of matrix
  • various norms for vectors and matrices: compatibility of vector and matrix norms.
  • Error Analysis for the solution of Ax=b
  • Error magnification factor and condition number of matrix
  • Jacobi method, Gauss-Seidel method, Successive-Over-Relaxation (SOR) method
  • Convergence of Jacobi Method, GS method and SOR method.
  • spectral radius of matrix
  • convergence of general iterative method for solving system of linear equations,
  • Sparse Matrix
  • Comparison of Gauss Elimination and iterative methods


Week.6

2.6 and 2.7

  • Conjugate Gradient Method
  • Nonlinear System of Equations
  • scalar product of vectors
  • determinant and eigenvalues of matrix
  • quadratic polynomials of n-variables
  • partial derivatives and gradients
  • chain rule for partial derivatives.
  • Symmetric positive definite matrix and properties
  • Construction of Conjugate Gradient (CG) Method
  • Propertise of CG Method
  • Preconditioning for CG method
  • Taylor's Theorem for multi-variate vector valued functions:
  • Newton's Method:
  • Broyden's Method
Week.7

3.1 and 3.2

  • Data and Interpolating Functions
  • Interpolation Error and Runge Phenomenon
  • Chebyshev interpolation
  • Fundamental theorem of algebra
  • Rolle's theorem.
  • Lagrange Basis Functions:
  • Properties of Lagrange basis functions:
  • Lagrange form of the interpolation polynomials
  • Newton's Divided Differences:
  • Properties of Newton's divided differences:
  • Newton's Form of the interpolation polynomials
  • Interpolation error analysis
  • Runge phenomenon
  • Chebyshev Polynomial
  • Error estimates for Chebyshev interpolation


Week.8

3.4, 3.5 and 4.1

  • Cubic Splines
  • Bezier Curves
  • Least Square Method
  • one-sided limits
  • continuity of functions
  • indefinite integrals
  • extremum values of multivariate quadratic functions.
  • Cubic splines
  • construction of cubic splines for interpolation
  • end conditions
  • properties of cubic spline interpolation
  • Bezier Curve and fonts
  • Least square method for solving inconsistent system of linear equations.
  • Basic properties of least square solutions:


Week.9

4.2 and 4.5

  • Mathematical Models and Data Fitting
  • Nonlinear Least Square Fitting
  • linear spaces, basis functions
  • product rule and chain rule for vector valued multivariate functions.
  • Least square method for curve fitting and statistical modeling.
  • Survey of Models: linear model, periodic model, exponential models, logistic model, etc
  • Taylor's theorem for vector valued multivariate functions.
  • Gauss-Newton Method
  • Levenberg-Marquardt Method


Week.10

5.1, 5.2 and 5.3

  • Numerical Differentiation
  • Numerical Integration: Newton-Cotes Formulas
  • Numerical Integration: Romberg's Technique
  • Taylor's theorem
  • interpolation error estimates
  • properties of definite inetgrals
  • Finite difference (FD) approximations of 1st order derivative and their error analysis
  • FD approximations of 2nd order derivatives and their error analysis
  • Undetermined coefficient method for FD approximation
  • Extropolation technique for improving the order of approximation
  • Midpoint rule, trapezoid rule and Simpson's rule;
  • Error analysis based on Taylor's Theorem and interpolation errors
  • Degree of precision of quadrature rules
  • Composite quadrature rules
  • Motivation, construction and implementation of Romberg's technique.


Week.11

5.4 and 5.5

  • Adaptive Numerical Integration
  • Gauss Quadrature Formulas
  • long divisions
  • changing variables for definite integrals
  • How to estimate the error on a subinterval
  • How to mark subintervals to be further refinement?
  • Implementation of adaptive numerical integration techniques.
  • Motivation and difficulties with straightforward approach.
  • Orthogonal polynomials,
  • Legendre polynomials and their basic properties;
  • Gauss quadrature rule based on Legendre polynomials
  • Degree of precision of Gauss Quadrature
  • Gauss quadrature formula on general interval and composite Gauss rules


Week.12

10.1 and 11.1

  • Discrete Fourier Transform and FFT
  • Discrete Cosine Transform (optional)
  • Image Compression (optional)
  • complex numbers and complex variables
  • integration by parts
  • convergence of sequences and series.
  • Fourier Series,
  • Discrete Fourier Transform
  • Matrix Form of Discrete Fourier Transform:
  • Inverse Discrete Fourier Transform:
  • DFT and Trigonometric interpolation
  • Algorithm for computing DFT: Fast Fourier Transform (FFT)
  • Discrete Cosine Transform (DCT),
  • DCT and Interpolation by Cosine Functions
  • Relation between DFT and DCT:
  • Fourier Transform of 2-Dimensional Functions
  • DCT of 2-Dimensional Functions:
  • Interpolation Theorem for 2-Dimensional DCT
  • Digital Gray scale images and color color images:
  • RGB format:
  • YCbCr (or YUV) format:
  • Convertion between RGB and YUV formats:
  • Quantization, Image Compression and Decompression


Week.13

12.1 and 12.2

  • Power Iteration Methods
  • QR Algorithm for Computing Eigenvalues
  • properties of eigen values and eigenvectors
  • Gram-Schmidt orthogonalization
  • Power iteration and its rate of convergence.
  • Inverse Power Iteration,
  • Inverse Power Iteration with Shift
  • Rayleigh Quotient Iteration


  • Definition and basic properties of orthogonal matrices:
  • QR-Factorization based on Gram-Schmidt Orthogonalization:
  • Normalized Simultaneous Iteration (NSI).
  • Unshifted QR Algorithm:
  • Shifted QR Algorithm:


Week.14

12.2

  • Algorithm for Computing Eigenvalues: Speed up of QR-algorithm:
  • matrices for orthogonal projection and reflection
  • block matrices and their products
  • similar matrices.
  • Upper Hessenberg form (UHF)
  • Householder Reflector
  • Convert a matrix into UHF by Householder reflectors