Difference between revisions of "MAT4XXX"

From Department of Mathematics at UTSA
Jump to navigation Jump to search
(Created page with "=Mathematical Physics - MAT4XXX/5XXX= ==Course description== The course intends to be a basic introduction to the mathematical and computational techniques in applied mathema...")
 
Line 50: Line 50:
 
*  
 
*  
 
||
 
||
Strang's 4 special matrices (Part 1)
+
Strang's 4 special matrices  
 
||
 
||
  
 
||
 
||
* Strang's 4 special matrices (continued)
+
*  
 
|-
 
|-
 
|Week 2
 
|Week 2
Line 60: Line 60:
 
*  
 
*  
 
||
 
||
Differences, Derivatives, BC.  
+
Differences, Derivatives, BC. Gradient, Divergence. Laplace equation.
 
||
 
||
  
 
||
 
||
*Gradient, Divergence. Laplace equation.
+
*
 
|-
 
|-
 
|Week 3
 
|Week 3
Line 70: Line 70:
 
*  
 
*  
 
||
 
||
Inverses.  
+
Inverses. Positive Definite Matrices
 
||
 
||
  
 
||
 
||
* Positive De�nite Matrices
+
*  
 
|-
 
|-
 
|Week 4
 
|Week 4
Line 80: Line 80:
 
*  
 
*  
 
||
 
||
Stiffness Matrices.  
+
Stiffness Matrices. Oscillations & Newton's Laws.
 
||
 
||
  
 
||
 
||
* Oscillations & Newton's Laws.
+
*  
 
|-
 
|-
 
|Week 5
 
|Week 5
Line 90: Line 90:
 
*  
 
*  
 
||
 
||
Graph Models. Networks.  
+
Graph Models. Networks. Clustering and k-means.
 
||
 
||
  
 
||
 
||
* Clustering and k-means.
+
*  
 
|-
 
|-
 
|Week 6
 
|Week 6
Line 100: Line 100:
 
*  
 
*  
 
||
 
||
Fourier Series.  
+
Fourier Series. Chebyshev, Legendre, and Bessel
 
||
 
||
  
 
||
 
||
* Chebyshev, Legendre, and Bessel
+
*  
 
|-
 
|-
 
|Week 7
 
|Week 7
Line 110: Line 110:
 
*  
 
*  
 
||
 
||
Fast Fourier Transform (FFT).  
+
Fast Fourier Transform (FFT). Convolution and Signal Processing.
 
||
 
||
 
   
 
   
 
||
 
||
* Convolution and Signal Processing.
+
*  
 
|-
 
|-
 
|Week 8
 
|Week 8
Line 120: Line 120:
 
*  
 
*  
 
||
 
||
Fourier Integrals. Deconvolution, Integral Equations.  
+
Fourier Integrals. Deconvolution, Integral Equations. Wavelets, Signal Processing.
 
||
 
||
  
 
||
 
||
* Wavelets, Signal Processing.
+
*  
 
|-
 
|-
 
|Week 9
 
|Week 9
Line 130: Line 130:
 
*  
 
*  
 
||
 
||
Computational implementation of Laplace and z- Transforms.  
+
Computational implementation of Laplace and z- Transforms. Spectral Methods.  
 
||
 
||
  
 
||
 
||
* Spectral Methods.
+
*  
 
|-
 
|-
 
|Week 10
 
|Week 10
Line 140: Line 140:
  
 
||
 
||
Finite Difference for ODEs.  
+
Finite Difference for ODEs. Accuracy & Stability. Conservation Laws, diffusion, fluids
 
||
 
||
  
 
||
 
||
* Accuracy & Stability. Conservation Laws, diffusion, fluids
+
*  
 
|-
 
|-
 
|Week 11
 
|Week 11
Line 150: Line 150:
 
*  
 
*  
 
||
 
||
Elimination with reordering, multigrid methods, conjugate gradients
+
Elimination with reordering, multigrid methods, conjugate gradients, Krylov subspaces
 
||
 
||
  
 
||
 
||
* Krylov subspaces
+
*  
  
 
|-
 
|-
Line 161: Line 161:
  
 
||
 
||
Regular. least sq. Linear programming.  
+
Regular. least sq. Linear programming. Adjoint. Stoch. Gradient Descent. ADAM.
 
||
 
||
 
*   
 
*   
 
||
 
||
* Adjoint. Stoch. Gradient Descent. ADAM.
+
*  
 
|-
 
|-
 
|Week 13
 
|Week 13
Line 171: Line 171:
 
*  
 
*  
 
||
 
||
Matrix-matrix Multiplication. 4 Fundamental Subspaces. Orthogonal Matrices.
+
Matrix-matrix Multiplication. 4 Fundamental Subspaces. Orthogonal Matrices. Best low rank matrix. Rayleigh quotients. Factoring matrices and tensors.
 
||
 
||
 
   
 
   
 
||
 
||
Best low rank matrix. Rayleigh quotients. Factoring matrices and tensors.
+
*   
 
|-
 
|-
 
|Week 14
 
|Week 14
 
||
 
||
* Randomized Linear Algebra. Low rank signals. Singular values. Compressed sensing.
+
*  
 
||
 
||
Covariance Matrices. Multivariate Gaussian. Weighted least squares. Markov chains.
+
Randomized Linear Algebra. Low rank signals. Singular values. Compressed sensing. Covariance Matrices. Multivariate Gaussian. Weighted least squares. Markov chains. Neural Networks (Convolutional, Deep). Backpropagation. Machine Learning.
 
||
 
||
Neural Networks (Convolutional, Deep).
+
*   
 
||
 
||
Backpropagation. Machine Learning.
+
*   
 
|}
 
|}

Revision as of 14:16, 24 January 2025

Mathematical Physics - MAT4XXX/5XXX

Course description

The course intends to be a basic introduction to the mathematical and computational techniques in applied mathematics, computational science & engineer�ing, and data science & machine learning. This course will stress then how the methods of mathematical modeling in the STEM disciplines have transitioned from the analytical (as in Theoretical Physics) to the numerical (as in traditional methods in Computational Science and Engineering) and more recently to Data-based methods (as in current developments in Data Science and Machine Learning). The student will acquire the basic skills needed broadly in Computational Science and Engineering, of which Computational Physics, Data Science, Machine Learning, and Numerical Modeling in the Mathematical Sciences are a subset.


Catalog entry

Prerequisite: Calculus III MAT2214 and Differential Equations I MAT3613 with a letter grade of C- or better, or successful completion of at least three credits of equivalent courses.

Content: 1. Computational Science, Engineering, and Mathematics (a) Linear Algebra and Computational Science & Engineering (b) Applied Math and Computational Science & Engineering (c) Fourier Series and Integrals (d) Laplace Transform and Spectral Methods (e) Initial Value Problems (f) Conjugate Gradients and Krylov Subspaces (g) Minimum Principles 2. Data Science and Machine Learning: a Mathematical Perspective (a) Principal Components and the Best Low Rank Matrix (b) Randomized Linear Algebra (c) Low Rank and Compressed Sensing (d) Markov Chains (e) Stochastic Gradient Descent and ADAM (f) Introduction to Machine Learning: Neural Networks


Textbooks:

  • Strang, G. Computational Science & Engineering. USA, Wellesley-Cambridge, 2007.
  • Strang, G. Linear Algebra and Learning from Data. Wellesley-Cambridge Press, 2019.


Topics List

Date Sections Topics Prerequisite Skills Student Learning Outcomes
Week 1

Strang's 4 special matrices

Week 2

Differences, Derivatives, BC. Gradient, Divergence. Laplace equation.

Week 3

Inverses. Positive Definite Matrices

Week 4

Stiffness Matrices. Oscillations & Newton's Laws.

Week 5

Graph Models. Networks. Clustering and k-means.

Week 6

Fourier Series. Chebyshev, Legendre, and Bessel

Week 7

Fast Fourier Transform (FFT). Convolution and Signal Processing.

Week 8

Fourier Integrals. Deconvolution, Integral Equations. Wavelets, Signal Processing.

Week 9

Computational implementation of Laplace and z- Transforms. Spectral Methods.

Week 10

Finite Difference for ODEs. Accuracy & Stability. Conservation Laws, diffusion, fluids

Week 11

Elimination with reordering, multigrid methods, conjugate gradients, Krylov subspaces

Week 12

Regular. least sq. Linear programming. Adjoint. Stoch. Gradient Descent. ADAM.

Week 13

Matrix-matrix Multiplication. 4 Fundamental Subspaces. Orthogonal Matrices. Best low rank matrix. Rayleigh quotients. Factoring matrices and tensors.

Week 14
Randomized Linear Algebra. Low rank signals. Singular values. Compressed sensing. Covariance Matrices. Multivariate Gaussian. Weighted least squares. Markov chains. Neural Networks (Convolutional, Deep). Backpropagation. Machine Learning.