# Foreword

What is functional analysis? A satisfactory answer requires going back to where it all started.

"All are present; the meeting convenes," intoned Fredholm. Intent were the gathered faces, their thoughts fixed on their students. "

What do we know of their weaknesses?"

Hilbert leaned back, torch's light flickering across his features. "Lots of dimensions, especially when they need to find the Hessian. What if… what if we made them deal with

infinitelymany dimensions?"...

It was Banach who finally spoke. "David, they already know about the vector space for the polynomials".

Hilbert smirked. "Who said anything about

countablyinfinite?". More silence, then glances, then grins.

It was Riesz's voice which next broke the silence. "And we can make them do analysis in that space. And linear algebra, but not the easy parts. Of course, they'll need to also deal with complex numbers. Sprinkle a little topology and abstract algebra on top, because... they deserve –"

"Frigyes, some of them might actually be able to do that. We need more." After a pause, Fredholm continued: "We'll tell them that they only need to know basic calculus."

# A Friendly Approach to Functional Analysis

I didn't actually find the book overly hard (it took me seven days to complete, which is how long it took for my first book, *Naïve Set Theory*), although there were some parts I skipped due to unclear exposition. it's actually one of my favorite books I've read in a while – it's for sure my favorite since the last one. That said, I'm very glad I didn't attempt this early in my book-reading journey.

## My brain won't stop line to me

Some part of me *insisted* that the left-shift mapping

is "non-linear" because it incinerates ! But wait, brain, this totally *is* linear, and it's also continuous with respect to the ambient supremum norm!

Formally, a map is linear when .

Informally, linearity is about being able to split a problem into small parts which can be solved individually. It doesn't have to "look like a line", or something. In fact, lines^{[1]} are linear *because* putting in more gets you more !

## Linearity and continuity

Two things surprised me.

First, a(n infinite-dimensional) linear function can be discontinuous. (?!)

Second, a linear function is continuous if and only if it is bounded; that is, there is an such that .

- The
**if**is easy: this is just Lipschitz continuity, which obviously implies normal continuity. - The other direction follows because the continuity implies that for , we can bound how much it's expanding the volume of some -ball and then apply linearity.

## What the hell are functional derivatives?

Derivatives tell you how quickly a function is changing in each input dimension. In single-variable calculus, the derivative of a function is a function .

In multi-variable calculus, the derivative of a function is a function – for a given -dimensional input vector, the real-valued output of can change differently depending on in which input dimension change occurs.

You can go even further and consider the derivative of , which is the function – for a given -dimensional input vector, again can change its vector-valued output differently depending on in which input dimension change occurs.

But what if we want to differentiate the following function, with domain and range :

How do you differentiate with respect to a function? I'm going to claim that

It's not clear why this is true, or what it even means. Here's an intuition: at any given point, there are uncountably many partial derivatives in the function space – there are many, many "directions" in which we could "push" a function around. gives us the partial derivative at with respect to .

This concept is important because it's what you use to prove e.g. that a line is the shortest continuous path between two points.

Below is an exchange between me (in plain text) and TheMajor (quoted text), reproduced and slightly edited with permission.

I'm having trouble understanding functional derivatives. I'm used to thinking about derivatives as with respect to time, or with respect to variations along the input dimensions. But when I think about a derivative on function space, I'm not sure what the "time" is, even though I can think about the topology and the neighborhoods around a given function.

And I know the answer is that there isn't "time", but I'm not sure what there *is*.

An interesting concept that comes to mind is thinking about a functional derivative with respect to e.g. a straight-line homotopy, where you really *could* say how a function is changing at every point with respect to time. But I don't think that's the same concept.

The concept is as follows:

Let's say we have some (a priori non-linear) map , which takes a function as an input and gives a number as an output. I.e. it maps from a vector space of functions to the complex numbers . Now fix a function , and a second function . We can then consider the 1-dimensional linear subspace . The map on this subspace is just a normal map, and if it is differentiable at the point in this subspace then its derivative is called

the functional derivative of at with respect to.

By normal map, is that something like a normal operator?

sorry, I didn't mean normal in a technical context. Since the subspace I introduced is one-dimensional (as a complex vector space), and it maps to the complex numbers as well, we have good old introduction to complex analysis derivatives here. If you like you can work with reals instead of complex variables too, in which case it would be the familiar real derivative.

Wouldn't it still output a function, maybe? wait. Would the derivative wrt just be ?

there is no derivative with respect to .

ah ya. duh (ETA: my brain was still acting as if differentiation had to be from the real numbers to the real numbers, so it searched for a real/complex number in the problem formalization and found .)

let me know if this part is clear, because unfortunately its the next few steps where it gets really confusing.

Unfortunately, I don't think it's clear yet. So I see how this is a one-dimensional subspace,^{[2]} because it's generated by one basis function ().

But I don't see how this translates to a normal complex derivative, in particular, I don't quite understand what the range of this function is.

No problem, and it's very good that you share that it's unclear. The range of is the complex numbers, maps from (our vector space of functions) to (the complex numbers).

I guess I'm confused why we're using that type signature if we're taking a derivative on the whole function – but maybe that'll be clear after I get the rest.

that is exactly the heart of the confusion surrounding functional derivatives, and we'll have to get there in a few steps.we'll start with defining functional derivatives for easy maps, i.e. the ones that take on complex values, and then work towards more complicated settings.

so back to the example above; we have a vector space (our 'function space'), we have a (possibly non-linear) map . we will now introduce the derivative of at with respect to , with . This derivative is just a complex number.

To find this we consider the 1-dimensional subspace that I introduced above, and we note that the map from to this subspace, given by , is a bijection that goes through at 0. this gives us a map from to , by sending to . We take the derivative of that at , and that is

the derivative of at with respect to.

Okay, that makes sense so far.

Nice 😃 this map has a few properties that I just want to remark and then ignore. For example it need not be linear in (which makes sense, since is only the point we're evaluating at). And by doing some work with chain rules it does have some linear properties in .

now there are two ways in which we can make this story complicated again, and most authors do both simultaneously.

Firstly we can try to extend the "derivative of at wrt " to something like "derivative of at ". We'll do this first. Secondly we can try to take a different map, say , which maps from into another vector space (instead of the complex numbers). We can then try and define a derivative of at wrt .

The first step is conceptually simple, but formally and computationally very difficult. Given a point and our map from before, we can simply say that "the derivative of at " is the map that sends to "the derivative of at with respect to ". So "the derivative of at " is a map from to .

this is formally difficult because usually you want this derivative to have some nice properties, but because it was defined pointwise it's very difficult to establish this! Frequently these derivatives are not continuous, and mathematicians resort to horrible tricks (like throwing out a bunch of points of the domain X on which our derivative is annoying) to recover some structure here.

So, given some arbitrary function which is "differentiable" at , we define a function (derivative of at with respect to )?

yes, exactly.

You could even maybe think of each input as projecting the derivative of at ? Or specifying one of many possible directions.

Yes, this is 100% correct. This is related to the "nice linear properties in " that I mentioned above

I also stated that this is computationally difficult. This is actually quite funny - the best way to find "The derivative of at " is to take a 'test function' (arbitrarily), compute (the derivative of at with respect to ), and then tahdah, you have now found the map that sends to (the derivative of at wrt ), i.e. exactly what you were looking for.

this sounds pretty computationally easy? Or are you calculating for a general test function , in which case, how do you get any nontrivial information out of that?

Yes, you need to calculate it for a general test function.

also something that may help with gaining insight: in multivariable calculus (lets say 2 dimensions, that's already plenty difficult) there is a clear divide between the [existence of a partial derivative of a function at a point] and [the function being differentiable at that point].

ETA: Back in my *Topology* review, I discussed a similar phenomenon: continuity in multiple input dimensions requires not just continuity in each input variable, but in *all* sequences converging to the point in question:

"Continuity in the variables says that paths along the axes converge in the right way. But for continuity overall, we need all paths to converge in the right way. Directional continuity when the domain is is a special case of this: continuity from below and from above if and only if continuity for all sequences converging topologically to ."

Similarly, for a function to be differentiable, the existence of all of its partial derivatives isn't enough – you need derivatives for every possible approach to the point in question. Here, the existence of all of the partials automatically guarantees the derivatives for every possible approach, because there's a partial for every function.

here we have the same, except we have (in an infinite-dimenional function space X) infinitely many 'partial derivatives'. so from that point of view it's not that surprising that a function "having a derivative at " is actually quite rare/complicated.

yeah, because has to exist for… all ? That seems a little tough.

It exists for all , and then exists as a formal map. But usually you want something stronger, for example that is continuous.

as an important but relatively trivial aside: if is a linear map, then does not actually depend on . So usually it is just called "the derivative of " instead of "the derivative of at ". This is confusing, because for non-linear there is also something called "the derivative of ", namely "the map that sends to [the derivative of at ]".

hm. That's because of the definition of linearity, right? it's a homomorphism for both the operations of addition and scalar multiplication... Wait, I intuitively understand why linearity means it's the same everywhere, but I'm having trouble coming up with the formal justification…

Yes, the point is that when we look at the definition of "derivative of at wrt " that is given by ...

ah, got it!

ok, so this was all the first way to make it confusing again. Ready for the second?

I'm ready to be reconfused.

Ok, so now let's pick a range not inside the complex numbers , but inside a second normed vector space . So we have a map , not necessarily linear. Again fix points . We are going to define the derivative of at wrt .

so we repeat our trick from before, consider the map from via to given by . We wish to differentiate it at .

unfortunately, its image is now in , not in , so we don't really know what the derivative means. But because is a normed vector space, the expression makes sense for all non-zero .

if this function can be continuously extended to then we define its image at 0 as the derivative of at wrt . Note that this notion of continuity has to do with the norm of .

this is now a vector in , so if this works we have: [the derivative of at wrt ] which is an element of , [the derivative of at ] which is a (linear! usually horrible and not continous!) map from to .

btw if the "continuously extending" part is new, you can also just think of it as the limit of that fraction as approaches 0. The only point is that (as long as we're working with complex vector spaces) there are a lot of different ways for to approach 0, and it has to work for all of them.

if we're working over the reals its simply the notion of "right limit" and "left limit" (the only two ways to approach 0 in ) that you may have seen before, except that the convergence is now happening in .

## Other notes

- The operator norm is really cool.
- Linear combinations always involve finitely many terms, but using the orthonormal basis of an infinite dimensional space, you can take the limit as .
- I was really happy to see watered-down versions of symmetry/conservation law correspondences (aka Noether's theorem). Can't wait to learn the real version.

## Final thoughts

The book is pretty nice overall, with some glaring road bumps – apparently, the Euler-Lagrange equation is one of the most important equations of all time, and Sasane barely spends any effort explaining it to the reader!

And if I didn't have the help of TheMajor, I wouldn't have understood the functional derivative, which, in my opinion, was *the* profoundly important insight I got from this book. My models of function space structure feel qualitatively improved. I can look at a Fourier transform and see what it's doing – I can *feel* it, to an extent. Without a doubt, that single insight makes it all worth it.

# Forward

I'm probably going to finish up an epidemiology textbook, before moving on to complex analysis, microeconomics, or... something else – who knows! If you're interested in taking advantage of quarantine to do some reading, feel free to reach out and maybe we can work through something together. 🙂

Lines () aren't actually linear functions, because they don't go through the origin. Instead, they're affine. ↩︎

To be more specific, is often an affine subspace, because the zero function is not necessarily a member. ↩︎