Physical and Mathematical Interpretations of an Adjoint Model with Application to ROMS Andy Moore University of Colorado Boulder.

Slides:



Advertisements
Similar presentations
General Linear Model With correlated error terms  =  2 V ≠  2 I.
Advertisements

The Inverse Regional Ocean Modeling System:
Solving Linear Systems (Numerical Recipes, Chap 2)
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Started detailed look at PCA Reviewed linear algebra Today: More linear algebra Multivariate.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Dan Witzner Hansen  Groups?  Improvements – what is missing?
Motion Analysis Slides are from RPI Registration Class.
Computer Graphics Recitation 5.
3D Geometry for Computer Graphics
ROMS/TOMS Tangent Linear and Adjoint Models Andrew Moore, CU Hernan Arango, Rutgers U Arthur Miller, Bruce Cornuelle, Emanuele Di Lorenzo, Doug Neilson.
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
The ROMS TL and ADJ Models: Tools for Generalized Stability Analysis and Data Assimilation Hernan Arango, Rutgers U Emanuele Di Lorenzo, GIT Arthur Miller,
Chapter 3 Determinants and Matrices
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
3D Geometry for Computer Graphics
Lecture 20 SVD and Its Applications Shang-Hua Teng.
Ordinary least squares regression (OLS)
Linear and generalised linear models
Class 25: Question 1 Which of the following vectors is orthogonal to the row space of A?
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Linear and generalised linear models
PHY 042: Electricity and Magnetism
Matrices CS485/685 Computer Vision Dr. George Bebis.
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 6. Eigenvalue problems.
1 Statistical Analysis Professor Lynne Stokes Department of Statistical Science Lecture 5QF Introduction to Vector and Matrix Operations Needed for the.
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Linear Algebra Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
1 February 24 Matrices 3.2 Matrices; Row reduction Standard form of a set of linear equations: Chapter 3 Linear Algebra Matrix of coefficients: Augmented.
Linear Algebra Lecture 25.
ROMS/TOMS TL and ADJ Models: Tools for Generalized Stability Analysis and Data Assimilation Andrew Moore, CU Hernan Arango, Rutgers U Arthur Miller, Bruce.
Algorithms for a large sparse nonlinear eigenvalue problem Yusaku Yamamoto Dept. of Computational Science & Engineering Nagoya University.
Computing a posteriori covariance in variational DA I.Gejadze, F.-X. Le Dimet, V.Shutyaev.
4 4.2 © 2012 Pearson Education, Inc. Vector Spaces NULL SPACES, COLUMN SPACES, AND LINEAR TRANSFORMATIONS.
SVD: Singular Value Decomposition
Least SquaresELE Adaptive Signal Processing 1 Method of Least Squares.
Method of Least Squares. Least Squares Method of Least Squares:  Deterministic approach The inputs u(1), u(2),..., u(N) are applied to the system The.
Quantum One: Lecture Representation Independent Properties of Linear Operators 3.
Hanjo Lim School of Electrical & Computer Engineering Lecture 2. Basic Theory of PhCs : EM waves in mixed dielectric.
Vector Norms and the related Matrix Norms. Properties of a Vector Norm: Euclidean Vector Norm: Riemannian metric:
Orthogonalization via Deflation By Achiya Dax Hydrological Service Jerusalem, Israel
Elementary Linear Algebra Anton & Rorres, 9th Edition
A Note on Rectangular Quotients By Achiya Dax Hydrological Service Jerusalem, Israel
Chapter 10 Real Inner Products and Least-Square
4 © 2012 Pearson Education, Inc. Vector Spaces 4.4 COORDINATE SYSTEMS.
Lecture 26 Molecular orbital theory II
Review of Linear Algebra Optimization 1/16/08 Recitation Joseph Bradley.
Review of Matrix Operations Vector: a sequence of elements (the order is important) e.g., x = (2, 1) denotes a vector length = sqrt(2*2+1*1) orientation.
1. Systems of Linear Equations and Matrices (8 Lectures) 1.1 Introduction to Systems of Linear Equations 1.2 Gaussian Elimination 1.3 Matrices and Matrix.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Instructor: Mircea Nicolescu Lecture 8 CS 485 / 685 Computer Vision.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Singular Value Decomposition and Numerical Rank. The SVD was established for real square matrices in the 1870’s by Beltrami & Jordan for complex square.
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
MA5233 Lecture 6 Krylov Subspaces and Conjugate Gradients Wayne M. Lawton Department of Mathematics National University of Singapore 2 Science Drive 2.
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
EE611 Deterministic Systems Vector Spaces and Basis Changes Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
CS479/679 Pattern Recognition Dr. George Bebis
Singular Value Decomposition
Some useful linear algebra
CS485/685 Computer Vision Dr. George Bebis
Singular Value Decomposition SVD
Chapter 3 Linear Algebra
Feature space tansformation methods
Subject :- Applied Mathematics
NULL SPACES, COLUMN SPACES, AND LINEAR TRANSFORMATIONS
Presentation transcript:

Physical and Mathematical Interpretations of an Adjoint Model with Application to ROMS Andy Moore University of Colorado Boulder

Linearized Systems  Consider a state-space (the ocean) with state vectors  Denote  ROMS is just a set of operators:  In general will be nonlinear.  For many problems it is of considerable theoretical and practical interest to consider perturbations to

 Let  In which case:  For many problems, it is sufficient to consider small perturbations: and negligible and negligible  The Tangent Linear Equation (TLE):  TLE forms core of many analyses (e.g. normal modes, linear iteration of nonlinear problems (data assimil))

Matrix-Vector Notation  ROMS solves the primitive equations in discrete form: NLROMS trajectory

Important Questions  Now that we have reduced the linearized ROMS (TLROMS) to a matrix, what would we like to know?  We should perhaps ask of what value is since in reality ROMS (and the real ocean) is nonlinear?

Justification for TLROMS  All perturbations begin in the linear regime.  Linear regime often continues to provide useful information long after nonlinearity becomes important.  Since the action of is to merely “scatter” energy, linear regime yields stochastic paradigms.

The Propagator  It is more convenient to work in terms of the TLROMS propagator:  So, what would we like to know about ?

Dimension  The ocean is a very large and potentially very complicated place!  But just how complicated is it?  What is it’s effective dimension?  Low dimension described by just a few d.o.f? or high dimension?  Does dimension depend on where we look?

R is BIG!!!  R is a monster!  Can we reduce R to something more managable?

Enter the Adjoint!  Eckart-Schmidt-Mirsky theorem: the most efficient representation of a matrix: where and are the orthonormal singular vectors of where and are the orthonormal singular vectors of  Singular Value Decomposition (SVD).

 By definition: where = singular values where = singular values = transpose propagator or = transpose propagator or adjoint (wrt Euclidean adjoint (wrt Euclidean norm) norm)  Clearly:

 Consider:  So looks like an EOF (i.e. something that we “observe”).

The Question of Dimension  The dimension of is equal to the “range” of (i.e. the set of singular vectors with ).  Dimension=“rank”=maximal # of independent rows and columns of  SVD is the most reliable method for determining numerically the rank of a matrix.

 In the limit to its continuous counterpart. to its continuous counterpart. Hypothesis  The rank of will provide fundamental information about the dimensionality of the real ocean circulation

The Active and Null Space  i.e. Transforms from v-space to u-space v-space to u-space  Suppose is an (NxN) matrix of rank P (i.e. )  If, (i.e. nothing is observed).  is the Null Space of  is the Activated Space of Initial state Final/Observed state

Null Space Activated Space

 Recall:  transforms from “observed u” back to “activated v-space”.  So if we observe “u” the adjoint tells us from whence it came! (cf Green’s functions). (cf Green’s functions). Observed State Initial State

Generating Vectors  Let be an (NxM) matrix, where N<M.  SVD yields two fundamental spaces: N-space and M-space N-space and M-space M-space to N-space N-space to M-space

 Consider the underdetermined system, given; unknown., given; unknown.  Unique solutions exist if:  Then:  is called the “generating vector”.  is called the “natural solution”.

 Suppose that has only P non-zero singular values:  SVD:  So is ALWAYS in P-space (i.e. “activated space” identified by )

A Familiar Example  The QG barotropic vorticity equation:  Solve for : underdetermined!  Adjoint vorticity equation yields the generating (stream) function:

Adjoint Applications  Clearly the adjoint operator of ROMS yields information about the subspace or dimensions that are activated by.  There are many applications that take advantage of this important property.

Sensitivity Analysis  Consider a function  Clearly  But  So Sensitivity

 Clearly the action of the adjoint restricts the sensitivity analysis to the subspace activated by (i.e. to the space occupied by “natural” solutions).

Least-Squares Fitting and Data Assimilation  If then the gradient provided by can be used to find that minimizes.  The is the idea behind 4-dimensional variational data assimilation (4DVAR)  Clearly the that minimizes lies within the active subspace of

Traditional Eigenmode Analysis  We are often taught to use the eigenmodes of to explore properties and stability of ocean.  In general, the eigenmodes of are NOT orthogonal, meaning each mode has a non-zero projection on other modes.  What does this do to our notion of active and null space?

Null Space based on modes of Activated Space based on modes of The two spaces overlap!

 The amplitude of a particular eigenmode of is determined by its projection on the active subspace (i.e. by it’s projection on the corresponding eigenmode of ).

Basin Modes in a Mean Flow Basic State Circulation Eigenmode #6 Adjoint Eigenmode #6

SVD and Generalized Stability Analysis  Recall from SVD that:  Time evolved SV is  Ratio of final to initial “energy” is:  So of all perturbations, is the one that maximizes the growth of “energy” over the time interval.

 Consider the forced TL equation:  If is stochastic in time, more general forms of SVD are of interest.  Assume unitary forcing:  Of particular interest are: Controllability Grammiam Observability Grammiam

 Eigenvectors of are the EOFs.  Eigenvectors of are the Stochastic Optimals.  Variance:  Eigenvectors of are balanced truncation vectors.  All have considerable practical utility and applications that go far beyond traditional eigenmode analysis!

Covariance Functions and Representers  The “controllability” Grammiam is nothing more than a covariance matrix.  Note that looks a lot like: which yields the “natural solution”. which yields the “natural solution”.  Operations involving yield only natural solutions related to “Representer Functions”.

Norm Dependence  The adjoint is norm dependent.  For the Euclidean norm,  Changing norms is simply equivalent to a rotation and/or change in metric Activated Subspace Null Space

Activated Space Summary The adjoint identifies the bits of state-space that actually do something!

The Adjoint of ROMS is a Wonderful Thing!

The Cast of Characters - played by NLROMS - played by TLROMS - played by ADROMS

Acknowledgements  Lanczos, C., 1961: Linear Differential Operators, Dover Press.  Klema, V.C. and A.J. Laub, 1980: The Singular Value Decomposition: Its Computation and Some Applications. IEEE Trans. Automatic Control, AC- 25,  Wunsch, C., 1996: The Ocean Circulation Inverse Problem, Cambridge University Press.  Brogan, W.L., 1991: Modern Control Theory, Prentice Hall.  Bennett, A.F., 1992: Inverse Methods in Physical Oceanography, Cambridge University Press.  Bennett, A.F., 2002: Inverse Modeling of the Ocean and Atmosphere, Cambridge University Press.  Benner, P., P. Mehrmann and D. Sorensen, 2005: Dimension Reduction of Large-Scale Systems. Lecture Notes in Computational Science and Engineering, Vol. 45, Springer-Verlag.  Trefethen, L.N and M. Embree, 2005: Spectra and Pseudospectra, Princeton University Press.  Farrell, B.F. and P.J. Ioannou, 1996: Generalized Stability Theory, Part I: Autonomous Operators. J. Atmos. Sci., 53,  Farrell, B.F. and P.J. Ioannou, 1996: Generalized Stability Theory, Part II: Nonautonomous Operators. J. Atmos. Sci., 53,

R is BIG!!!  R is a monster!  Can we reduce R to something more managable?