What is the motivation of adjoints and normal operators. By "motivation," I mean an example, such as a proof, where it is natural to use them.


The idea of adjoint came from Lagrange who used integration by parts to differentiations and multiplications from one function to another in the integral. $$ \int (Lf)gdx-\int f(L^*g)dx = \mathscr{L}(f,g). $$ $L^*$ was the Lagrange adjoint obtained by integrating by parts. Lagrange used his formula to come up with variation of parameters in order to reduce an ODE to lower order, and his formula was later used to study the first symmetric differential equations arising from Fourier's separation of variables technique for solving his Heat equation. Fourier's application, however, naturally involved cases where $L=L^*$, which singled this case out for further study.

Sturm, along with Liouville initiated a study of "symmetric" ODEs in this context of Fourier and Lagrange, and they studied the associated orthogonal eigenfunction expansions as well. Endpoint conditions were imposed that would force the evaluation terms $\mathscr{L}(f,g)$ to vanish. Such conditions arose naturally in the context of Fourier's study of the Heat Equation. This led to operators that were symmetric on the domain of functions that were sufficiently differentiable and satisfied the endpoint conditions:

$$ \int_a^b (Lf)g dx = \int_a^b f(Lg)dx. $$ They realized that, just as Fourier had found, there were resulting eigenfunctions, discrete eigenvalues, and functions could be expanded in these orthogonal eigenfunctions. It was quite a remarkable thing considering that linear space had not been defined yet, and they were working in an infinite-dimensional space. It wasn't until decades later that symmetry was used to study matrices, and to find similar orthogonal expansions in eigenvectors of symmetric matrices.

So it all seems a little unnatural because infinite-dimensional analysis of symmetry and eigenfunctions came well before the analysis of finite-dimensional cases, which makes the natural applications inaccessible in the study of finite-dimensional Linear Algebra. The most abstract came first, which is also rather unusual in Mathematics.

By the way, I'm not sure where the study of normal operators started, but a normal $N$ can be written as $N=A+iB$ where $A$, $B$ are selfadjoint and commute with each other.


In some applications, the adjoint operator has a meaning. For instance, in signal processing, there are linear operators $T$ that map a signal (think: a sample of human speech) into a different representation (think: a digital representation), and then the adjoint $T^{*}$ maps the corresponding representations back into signals (e.g. digital signals into audio). A simple example is the Fourier transform for periodic signals.

If you're familiar with Fourier series, then an example of the previous type is the Fourier transform: if $T : L^{2}(\mathbb{T}) \to \ell^{2}(\mathbb{Z})$ is given by $Tf = (\hat{f}(n))_{n \in \mathbb{Z}}$, where $\hat{f}(n) = \int_{\mathbb{T}} f(x) e^{-i 2 \pi x} \, dx$, then $T^{*}a = \sum_{j \in \mathbb{Z}} a_{j} e_{j}$, where $e_{j}(x) = e^{i 2 \pi j x}$, is the adjoint. The interpretation is periodic "speech" signals get mapped to their "digital" Fourier coefficients by $T$, and $T^{*}$ takes sequences of coefficients to periodic signals.

Another nice example of this is in the theory of diffusion processes. Roughly speaking, a diffusion process is determined by the partial differential equation $\frac{\partial u}{\partial t} - L u = 0$. If $L^{*}$ is the adjoint operator of $L$, then the partial differential equation $\frac{\partial u}{\partial t} - L^{*}u = 0$ governs the time reversal of the diffusion. (The time reversal of a diffusion is another diffusion that looks like the original going backwards.) In other words, in this case, taking adjoints is interpreted as reversing time.