One of the most exquisite minds of our times shares his insights of how a Mathematician perceives the worlds of Mathematics and Physics.

A *must read* for all ages and all specializations, from students to retirees.

One of the most exquisite minds of our times shares his insights of how a Mathematician perceives the worlds of Mathematics and Physics.

A *must read* for all ages and all specializations, from students to retirees.

It so happened that at the beginning of 2016 I gave a talk on the conference “Geometric aspects of modern dynamics” in Porto, delivered a minicourse at Journées Louis Antoine in Rennes and wrote an expository paper for the European Mathematical Society Newsletter, all devoted to the same subject. The subject, provisionally dubbed as “Analgebraic geometry”, deals with algebraic-like properties (especially from the point of view of intersection theory) of real and complex analytic varieties defined by ordinary and Pfaffian differential equations with polynomial right hand sides. Thus

analgebraic=un-algebraic +analytic +algebraic(background) + weak algebraicity-like properties.

It turns out that this analgebraic geometry has very intimate connections with classical problems like Hilbert 16th problem, properties of periods of algebraic varieties, analytic number theory and arithmetic geometry.

For more details see the presentation prepared for the minicourse (or the shorter version of the talk) and the draft of the paper.

Any remarks and comments will be highly appreciated.

P.S. Video records (in French) are available from this page.

The exam is posted online on **Feb 1, 2016**, and must be submitted on the last day of the exams’ period, **February 26**. Its goals are, besides testing your acquired skills in the Analysis, to teach you a few extra things and see your ability for logical reasoning, not your proficiency in performing long computations. If you find yourself involved in heavy computations, better double check whether you understand the formulation of the problem correctly. Remember, small details sometimes matter!

Please provide argumentation, better in the form of logical formulas, not forgetting explicit or implicit quantifiers and . They really may change the meaning of what you write!

Problems are often subdivided into items. The order of these items is not accidental, try to solve them from the first till the last, and not in a random order (solution of one item may be a building block for the next one).

To get the maximal grade, it is *not necessary* to solve all problems, but it is imperative *not to write* stupid things. Please don’t try to shoot in the air.

The English version is the authoritative source, but if somebody translates it into Hebrew (for the sake of the rest of you) and send me the translation, I will post it for your convenience, but responsibility will be largely with the translator.

If you believe you found an error or crucial omission in the formulation of a problem, please write me. If this will be indeed the case (*errare humanum est*), the problem will be either edited (in case of minor omissions) or cancelled (on my account).

That’s all, folks!© Good luck to everybody!

**Yes, and feel free to leave your questions/talkbacks here, whether addressed to Michal/Boaz/me or to yourself, if you feel you want to ask a relevant question. **

The formulation of Problem 1 was indeed incorrect. The set was intended to be the set of accumulation points for a set . The formal definition is as follows.

*Definition*. A point belongs to to the set of limit points if and only if >0 the intersection is **infinite**. The point itself may be or may be not in .

Isolated points of are never in , but may contain points .

Apologies for the hasty formulation.

**The statement requested to prove in Problem 3(b) is wrong**, and I am impressed how fast did you discover that. Actually, the problem was taken from the textbook by Zorich, vol. 1, where it appears on p. 169, sec. 4.2.3, as Problem 4.

The assertion about existence of the common fixed point of two commuting continuous functions becomes true if we require these functions to be continuously differentiable on (in particular, for polynomials), but the proof of this fact is too difficult to be suggested as a problem for the exam.

Thus **Problem 3(b) is cancelled.**

Questions concerned integrability of discontinuous functions, notions of improper integrals (how and when they can be defined), topological properties (equivalent definitions of compactness, connectedness etc.)

Here are some textbooks that I recommend for preparing when working on the exam. Keep them on your virtual bookshelf: they cover much more that I explained in the course, but who knows what questions related to analysis you might have.

- V. Zorich, vol. 1: Chapters 1-6, pp.1-371.
- V. Zorich, vol. 2: Parts of Chapter 9 (continuous maps) and the first part of Chapter 18 on Fourier series.
- W. Rudin: Chapters 1-6, pp.1-165.

The problems for exam will be posted on February 1st (at least the English version).

- Low regularity functions: example, discussion. Stock market, home electricity meter…
- Euclidean geometry in the -space, real and complex. Normed bases and their advantage. Projections.
- Scalar product on the space of functions as the limit of finite-dimensional scalar products.
- Fourier coefficients in the real and complex form. Rate of their decay for functions of different regularity.
- Fourier series and their main properties (without proofs). Completeness of the exponential/trigonometric system, Parseval equality.
- Connection with holomorphic functions. Converging power series as a source of fast converging Fourier series.

The field is naturally extending the field , which means that all arithmetic operations on extend naturally as operations on . In particular, any polynomial can be interpreted as a map . Geometrically, this can be visualized as a map of the 2-plane to the 2-plane.

We discussed the maps and .

Smooth functions are those which can be accurately approximated by the -affine maps near each point . Functions that can be accurately approximated by -affine maps (the same formula, but over the complex numbers), are called holomorphic (or complex analytic). Such maps are characterized by the property that small circles are mapped into small *almost-circles*, that is,

- angles are preserved, and
- lengths are scaled

Sometimes these local conditions become global. Examples: the affine maps send lines to lines and circles to circles. The map maps circles and lines into circles or lines (depending on whether the circles/lines pass through the origin ).

Integration is carried over smooth (or piecewise smooth) paths in , using Riemann-like sums. It depends linearly on the function which we integrate, but in contrast with the real case we have much more freedom in choosing the paths.

- If is a constant, and is a closed triangle, then the integral is zero as the sum .
- If is non-constant, then the integral identity is valid only for “very small triangles” near a point with .
- This implies that as long as the paths share the common endpoints and can be continuously deformed one into the other.

Integrals over closed loops are zeros, unless there are singular points (where the function is non-holomoprhic) inside.

Example: is non-holomorphic at , and .

If is holomorphic inside a domain bounded by a closed curve and , then .

In other words, the value of on the boundary uniquely determine its values inside the domain. This is in wild contrast with functions of real variable!

As a function of , the expression admits a converging Taylor expansion (in fact, the same old geometric progression series) in powers of for any . Thus if we choose off the path , then the series will converge for any (warning! note that the “variable” and the “parameter” exchanged their roles!!!), the Cauchy integral can be expanded in the converging series of powers , hence the function gets expanded in the series .

Conclusion: functions that are -differentiable can be expanded in the convergent Taylor series (hence are “polynomials of infinite degree”) and vice versa, “polynomials of infinite degree” are infinitely -differentiable. This is a miracle that so many functions around us are actually holomorphic!

The way how logarithmic, exponential and trigonometric functions are usually introduced, is not very satisfactory and appears artificial. For instance, the mere definition of the non-integer power , , is problematic. For , one can define the value as the root , but the choice of branch/sign and the possibility of defining it for negative is speculative. For instance, the functions and may turn out to be different, depending on whether the latter is defined as (makes sense for negative ) or as which makes sense only for positive . But even if we agree that the domain of should be restricted to positive arguments only, still there is a big question why for two close values and the values, say, and should also be close…

The right way to introduce these functions is by looking at the differential equations which they satisfy.

A differential equation (of the first order) is a relation, usually rational, involving the unknown function , its derivative and some known rational functions of the independent variable . If the relation involves higher derivatives, we say about higher order differential equations. One can also consider systems of differential equations, involving several relations between several unknown functions and their derivatives.

**Example. ** Any relation of the form implicitly defines as a function of and can be considered as a trivial equation of order zero.

**Example.** The equation with a known function is a very simple differential equation. If is integrable (say, continuous), then its solution is given by the integral with variable upper limit, for any meaningful choice of the lower limit . Any two solutions differ by a constant.

**Example.** The equation with a known function . Even the case where is a constant, there is no, say, polynomial solution to this equation (why?), except for the trivial one . This equation is linear: together with any two functions and any constant , the functions and are also solutions.

**Example.** The equation has a family of solutions for any choice of the constant (check it!). However, any such solution “explodes” at the point , while the equation itself has no special “misbehavior” at this point (in fact, the equation does not depend on at all).

The transcendental function satisfies the differential equation : this is the only case of the equation , which has no rational solution. In fact, all properties of the logarithm follow from the fact that it satisfies the above equation and the constant of integration is chosen so that . In other words, we show that the function defined as the integral possesses all what we want. We show that:

- is defined for all , is monotone growing from to as varies from to .
- is infinitely differentiable, concave.
- transforms the operation of multiplication (of positive numbers) into the addition: for any .

The above listed properties of the logarithm ensure that there is an inverse function, denoted provisionally by , which is inverse to . This function is defined for all real , takes positive values and transforms the addition to the multiplication: . Denoting the the value by , we conclude that for all , and for all rational values . Thus the function , defined as the inverse to , gives interpolation of the exponent for all real arguments. A simple calculation shows that satisfies the differential equation with the initial condition .

Consider the integral operator which sends any (continuous) function to the function defined by the formula . Applying this operator to the function and using the differential equation, we see that is a “fixed point” of the transformation : . This suggests using the following approach to compute the function : choose a function and build the sequence of functions , . *If there exists a limit* , then this limit is a fixed point for .

Note that the action of $\Phi$ can be very easily calculated on the monomials: (check it!). Therefore if we start with , we obtain the functions $\latex f_n=1+x+\frac12 x^2+\cdots+\frac1{n!}x^n$. This sequence converges to the sum of the infinite series which represents the solution on the entire real line (check that). This series can be used for a fast approximate calculation of the number .

The function satisfies the differential equation . The corresponding “motion on the complex plane”, , is rotation along the (unit) circle with the unit (absolute) speed, hence the real and imaginary parts of are cosine and sine respectively. In fact, the “right” definition of them is exactly like that,

.

Thus, the Euler formula “cis” in fact is the *definition* of sine and cosine. Of course, it can be “proved” by substituting the imaginary value into the Taylor series for the exponent, collecting the real and imaginary parts and comparing them with the Taylor series for the sine and cosine.

In fact, both sine and cosine are in turn solutions of the real differential equations: derivating the equation , one concludes that . It can be used to calculate the Taylor coefficients for sine and cosine.

For more details see the lecture notes.

Not completely covered in the class: solution of linear equations with constant coefficients and resonances.

- Area under the graph as a paradigm
- Definitions (upper and lower sums, integrability).
- Integrability of continuous functions.
- Newton-Leibniz formula: integral and antiderivative.
- Elementary rules of antiderivation (linearity, anti-Leibniz rule of “integration by parts”).
- Anti-chain rule, change of variables in the integral and its geometric meaning.
- Riemann–Stieltjes integral and change of variables in it.
- Integrability of discontinuous functions.

Not covered in the class: Lebesgue theorem and motivations for transition from Riemann to the Lebesgue integral.

The sketchy notes are available here.

We discussed a few issues:

- Lagrange interpolation formula: how to estimate the difference through the derivative ?
- Consequence: vanishing of several derivatives at a point means that a function has a “root of high order” at this point (with explanation, what does that mean).
- Taylor formula for polynomials: if you know all derivatives of a polynomial at some point, then you know it everywhere.
- Peano formula for -smooth functions: approximation by the Taylor polynomial with asymptotic bound for the error.
- Lagrange formula: explicit estimate for the error.

The notes (updated) are available here.

Continuity of functions (and maps) means that they can be nicely approximated by constant functions (maps) in a sufficiently small neighborhood of each point. Yet the constant maps (easy to understand as they are) are not the only “simple” maps.

Linear maps naturally live on *vector spaces*, sets equipped with a special structure. Recall that is algebraically a *field*: real numbers cane be added, subtracted between themselves and the ratio is well defined for .

**Definition.** A set is said to be a vector space (over ), if the operations of addition/subtraction and multiplication by constant are defined on it and obey the obvious rules of commutativity, associativity and distributivity. Some people prefer to call vector spaces *linear spaces*: the two terms are identical.

**Warning.** There is no “natural” multiplication !

**Examples.**

- The field itself. If we want to stress that it is considered as a vector space, we write .
- The set of tuples is the Euclidean -space. For it can be identified with the “geometric” plane and space, using coordinates.
- The set of all polynomials of bounded degree with real coefficients.
- The set of all polynomials without any control over the degree.
- The set of all continuous functions on the segment .

**Warning.** The two last examples are special: the corresponding spaces are not finite-dimensional (we did not have time to discuss what is the dimension of a linear space in general…)

Let be two (different or identical) vector spaces and is a function (map) between them.

**Definition. ** The map $f$ is linear, if it preserves the operations on vectors, i.e., .

Sometimes we will use the notation .

**Obvious properties of linearity.**

- (Note: the two zeros may lie in different spaces!)
- For any two given spaces the linear maps between them can be added and multiplied by constants in a natural way! If , then we define for any (define yourselves). The result will be again a linear map between the same spaces.
- If and , then the composition is well defined and again linear.

**Examples.**

- Any linear map has the form (do you understand why the notations are used?)
- Any linear map has the form for some numbers . Argue that all such maps form a linear space isomorphic to back again.
- Explain how linear maps from to can be recorded using -matrices. How the composition of linear maps is related to the multiplication of matrices?

The first example shows that linear maps of to itself are “labeled” by real numbers (“*multiplicators*“). Composition of linear maps corresponds to multiplication of the corresponding multiplicators (whence the name). A linear 1-dim map is invertible if and only if the multiplicator is nonzero.

**Corollary.** *Invertible linear maps constitute a commutative group (by composition) isomorphic to the multiplicative group . *

Maps of the form for a fixed vector (the domain and source coincide!) are called *shifts* (a.k.a. translations). **Warning:** The shifts are not linear unless ! Composition of two shifts is again a shift.

**Exercise.**

*Prove that all translations form a commutative group (by composition) isomorphic to the space itself.* (Hint: this is a tautological statement).

**Definition.**

A map between two vector spaces is called *affine*, if it is a composition of a linear map and translations.

**Example.**

Any affine map has the form for some . Sometimes it is more convenient to write the map under the form : this is possible for any point . Note that the composition of affine maps in dimension 1 is not commutative anymore.

**Key computation.** Assume you are given a map in the sense that you can evaluate it at any point . Suppose an oracle tells you that this map is affine. How can you restore the explicit formula for ?

Obviously, . To find , we have to plug into it any point and the corresponding value . Given that , we have for any choice of .

The expression for a non-affine function is in general not-constant and depends on the choice of the point .

**Definition.** A function is called differentiable at the point , if the above expression for , albeit non-constant, has a limit as , where is a function which tends to zero. The number is called the **derivative** of at the point and denoted by (and also by half a dozen of other symbols: , …).

Existence of the limit means that near the point the function admits a reasonable approximation by an affine function : , i.e., the “non-affine part” is small not just by itself, but also relative to small difference .

See the notes and their earlier version.

The only non-obvious moment is differentiability of the product: the product (unlike the composition) of affine functions is not affine anymore, but is immediately differentiable:

, but the quadratic term is vanishing relative to , so the entire sum is differentiable.

**Exercise.** Derive the Leibniz rule for the derivative of the product.

Affine functions have no (strong) maxima or minima, unless restricted on finite segments. Yet absence of the extremum is a strong property which descends from the affine approximation to the original function. Details here and here.