Got math?

This is the E2 usergroup e2, which was originally a proper subset of the e2science group (e^2e2science). At first, the group name was e π i + 1 = 0, but some whiners thought that would be too hard to send a message:

/msg <var>e</var><sup>_&pi;_<var>i</var></sup>_+_1_=_0 After typing that, I forgot what I was going to say.

So here we are instead with a simpler (but more boring) name e2theipiplus1equalszero. Update: more complainers. Now we're just e^2. (Now does that means e² or e XOR 2 ? That is my secret.) Tough luck for those without a caret key.

e^2 often erupts into long mathematical discussions, giving members more /msgs than they care to digest. So, you have a few other options if the math is going to get seriously hairy:

  • Send to only those members of the group currently online:

    /msg? e^2 Brouwer was a Communist!.
     
  • Speak in the chatterbox. But be prepared to give non-math noders headaches.
  • Add the room Mathspeak to the list of rooms you monitor in squawkbox or Gab Central. Mathspeak died of loneliness.

You may want to read some of these while you are calculating ln π.


Venerable members of this group:

Wntrmute, cjeris, s19aw, Brontosaurus, TanisNikana, abiessu, Siobhan, nol, flyingroc, krimson, Iguanaonastick, Eclectic Scion, haggai, redbaker, wazroth, small, Karl von Johnson, Eidolos, Ryouga, SlackinWhileSleepin, ariels, quantumlemur, futilelord, Leucosia, RPGeek, Anark, ceylonbreakfast, fledy, Oolong@+, DutchDemon, jrn, allispaul, greth, chomps, JavaBean, waverider37, IWhoSawTheFace, DTal, not_crazy_yet, Singing Raven, pandorica, Gorgonzola, memplex, tubular, Tom Rook
This group of 45 members is led by Wntrmute

A tangent space is a useful structure which can be extracted from the differential structure of a differential manifold M. On such a manifold, it is handy to be able to define vectors, so that we can calculate things like, say, the velocity of a given curve defined on M. However, since a general manifold doesn't have to be defined as a subspace of flat euclidean space, we can no longer just think of a vector as some "arrow" living in the ambient space. In euclidean space, we would just draw an arrow, and state that it's "located" at the coordinates of the non-pointy end of the arrow, and it had a "direction" and "magnitude" given by the coordinates of the "point" of the arrow. We were able to get away with this nonsense because there was no curvature in euclidean space, and our vectors therefore are given by straight lines. The essential core of a vector is linearity, and it is that property of vectors that we cannot lose when generalizing to curved spaces. Therefore, we can no longer define our vectors as living in the manifold itself. If we try to represent our vectors as "curved" arrows in the manifold, there is no proper way of "adding" them together; we lose linearity. The bottom line is, we need a new place for our vectors to live. So, we seek to define the tangent space to M at any given point p in the manifold. This can be done in a few different ways.

If M is an n-dimensional differential manifold, the tangent space to M at p, denoted TpM, is an n-dimensional vector space. You might like to think of it as the "closest flat approximation to M at p", but so far, it's really just a copy of Rn where vectors at p can be defined.

Curves in the Manifold

Let's begin in a familiar setting. We look at the velocity vector for a curve in R3. Define the curve as a map C: (-1,1) → R3. In local coordinates on R3 (which we can just take to be the cartesian coordinates), we can represent C by (x1(t), x2(t), x3(t))*. The tangent vector to C inside of R3 is given by just taking the derivative with respect to the curve parameter, t: (dx1/dt|p, dx2/dt|p, dx3/dt|p). This can be interpreted as the velocity vector of the curve at the point p.

Now, there is no reason we can't do this for arbitrary manifolds; in a small enough neighborhood of a point p, we have a local coordinate representation of any curve passing through p, and thus we can compute the tangent vector to such a curve in those coordinates. We can then compare the tangent vectors to multiple curves passing through p in a natural way, since we can use the same coordinate representation to look at both curves.

All of this is useful to us, because in a way we can now associate vectors defined at a point p on M with curves in M passing through p; "vectors" are really just the velocities of curves. This definition is nice, because it is a completely coordinate-independent statement. Note, however, the relationship between vectors at p and curves passing through p is not a one-to-one mapping. For a given vector vp at p, there are many different curves passing through p which have the same velocity vp. Therefore, in order to associate the tangent space at p with the set of curves passing through p, we must first define an equivalence class of curves.

Consider the set of curves on M passing through p. An element Cj of this space is a map Ci: (-1,1) → M such that Cj(0) = p.

Define two such curves Cj and Ck to be "equivalent" when their tangent vectors at p agree in a local coordinate representation {xi} about p. You can check that this definition is independent of the choice of coordinates {xi}.

Now, we can define a tangent vector to M at p to be an element of the equivalence class of curves on M through p defined as above. The tangent space, TpM, is therefore the space of all such equivalence classes. Notice we use the notation vp, hilighting the fact that vectors at different points on the manifold live in different tangent spaces. Before defining additional structure, there is no way of comparing two vectors vp and vq at different points on the manifold.

This is one good way of defining the tangent space. It is nice because its definition does not rely on our choice of coordinates. However, it is a bit abstract, and therefore it is common to use the coordinate-dependent version, especially when doing calculations:

Coordinate Representations of Curves

Given a point p in M, and local coordinates {xi} on a neighborhood of p, the tangent space TpM is the space of all possible velocities (dC1/dt|p, dC2/dt|p, ..., dCn/dt|p) for curves C(t) mapping into M. The components of a vector v in M are given by the components of this coordinate representation for dC/dt;

vi = dCi/dt

Note that the components of v are coordinate-dependent. If we chose a different patch to represent the neighborhood of p in M, we would generally get different components vi. While v is a coordinate-independent object, its components are entirely coordinate-dependent.

Directional Derivatives

There is another way of defining a tangent vector space, which involves looking at derivatives of functions defined on M. At a point p, given a tangent vector, vp (defined in one of the two ways above), and a function f: M → R, we ask how quickly is f changing in the v-direction?

The notation we will use will be vp(f) = the directional derivative of f at p in the v-direction. To calculate vp(f), we choose an associated curve C in M through p (from the equivalence class of curves corresponding to v) with velocity v at p. Then, the rate of change of f in the v-direction is given by the rate of change of f o C with respect to the curve parameter t, evaluated at p:

vp(f) = d/dt(f(C(t)))|p.

This definition for vp(f) is manifestly coordinate-independent, since we haven't chosen a coordinate chart yet. However, we haven't yet shown that this definition is independent of our choice of curve, C (remember, we just picked one out of the set of curves with velocity v at p). We can do this by choosing a coordinate chart, {xi}, and carrying through the t-derivative using the chain rule:

vp(f) = d/dt(f(C(t)))|p = (df/dxi)(dxi/dt)|p

where xi(t) is just the coordinate representation of the curve C in M. Note we are implicitly summing over the coordinate index i, using the Einstein summation convention. Simplifying a little of the cumbersome notation, we write:

vp(f) = (dxi/dt) (∂if)

Thus, the directional derivative of f is specified by n parameters, dxi/dt, for i = {1,...,n}. These parameters are exactly the components vi in the coordinate representation given above. Now let us think of vp as an operator on functions in M. By the above equation, it is clear that this is a linear operator, and that its operation on functions is completely specified by the components vi. This means we have a new representation of our tangent space, TpM:

The tangent space TpM is the space of all directional derivative operators vp, acting on smooth functions f: M → R, and returning a real number given by the equation:

vp(f) = vi (∂if)|p.

So, now we can think of vectors as linear maps from functions into R. This concept is a bit abstract upon first viewing, so let's play with the algebra until you're a bit more convinced.

First of all, notice that by this definition adding two vectors together is the same thing as adding their components together, as one would expect. Secondly, multiplying a vector by a given constant just multiplies the components by a constant. Thus, directional derivative operators do form a linear space. Now, let's say we choose a given set of coordinates {xi} and choose a particular directional derivative, which we will call e1, given by e11 = 1, all other e1i = 0. We will show that this is one of a natural set of basis vectors for TpM. The equation above shows us how this operator acts on functions:

e1(f) = e1i (∂if) = ∂1f.

In other words, e1 = ∂1. Likewise, for any j, ej = ∂j, the j'th partial derivative. Since every directional derivative is a linear combination of partial derivatives, we can always express any vector in TpM as a linear combination of the ej's. Thus, we have established that we can think of {∂j} as a basis for TpM.

Change of Basis

Now that we're a bit more comfortable with the notion of a vector as a linear map, let's look at the consistency of this definition when going from one coordinate patch to another. Remember, when we first gave the definition for a directional derivative operator, we noted that it was manifestly coordinate-independent. However, when we chose a set of coordinates {xi}, we got a set of components for v, {vi}, which were coordinate-dependent. So, a natural question to ask would be: How exactly do the components vi for a vector v transform when we make a new choice of coordinates? This question is treated very thoroughly in the tensor node**, but the intuitive picture is that the components must somehow transform in a way "opposite" that of the basis vectors, so that the overall definition is coordinate-independent. Specifically, if the coordinate transformation results in a change of basis which can be represented as a matrix acting on basis vectors,

fi = ∂/∂yi = Mij ej = Mij ∂/∂xj

then the components of a given vector v must transform via the inverse matrix:

(v')i = (M -1)ij vj.

Since we are expressing the basis vectors as partial derivatives, the matrix is just given by the chain rule,

Mij = ∂xj/∂yi

and its inverse is given by using the chain rule in the opposite direction,

(M -1)ij = ∂yi/∂xj.

We have written down three equivalent approaches to understanding the tangent space TpM. In this third picture, we can think of TpM as all possible directional derivative operators at p on functions defined on M. In local coordinates, these are all expressible as linear combinations of partial derivatives, {∂/∂xj|p}, j = {1,...,n}. The transformation rule for a change of coordinates is just given by the chain rule on partial derivatives; the basis vectors transform like derivatives, and the components of a vector transform inversely to the basis vectors. If we wanted, we could now define vectors in a fourth way, i.e. that they are merely a set of components {vi} which transform via the inverse chain rule, as above. This definition for vectors would be entirely equivalent to the other three given above, but there is no corresponding intrinsic description of the space TpM. Vectors in this picture lose their intrinsic value as a mathematical object; they no longer "live" anywhere as an element of a topological space. This definition is the most common, however, because it's easiest for purposes of direct computation.

A Bigger Picture

Now that we've constructed the tangent space TpM from several different perspectives, it is a useful question to ask, "How is TpM related to TqM, for different points p and q in the manifold M?" We know that since the manifold is smooth, the vector spaces should somehow mesh smoothly with each other. We can also be sure about how this "meshing" should work on a global scale; this is how we define the tangent bundle, denoted simply TM. However, there is no god-given way of answering this question locally; we need to add additional structure to our manifold before we can compare tangent spaces. This leads to discussions of connections, parallel transport, covariant differentiation, and curvature. All of these notions can be defined before we add the further structure of a metric to the manifold.


*Try not to be confused by the fact that I'm using superscripts for the coordinate indices. This isn't supposed to represent "x squared" and "x cubed". It's just the most standard notation used for coordinate indices in this particular branch of mathematics and physics.
**This writeup was partially constructed as penance for my writeup in the tensor node. Since I produced that writeup, other noders created much better descriptions of tensors, describing them as multilinear maps acting on functions, rather than a set of components which transforms. It seemed silly to change my writeup at that point, considering that everything had been explained by others already, so I figured it would make better sense to try describing the place that vectors live, which is a similar but distinct concept.

Once again the opportunity has arisen to explain in simple language what's going on hereinabove.

This is topology for the common man.

GLOSSARY:

Topology: the study of bottle caps, screw-on jar lids, catsup bottle lids, and toupees. (You see, all part of the set X="tops").

Totally disconnected: the state of being out of touch with reality. Aunt Minerva's been in and out of psychiatric hospitals for years, and she's heavily medicated. She's a delightful old dear, but is indeed totally disconnected.

Cantor set: when the singing rabbi does a few numbers, then sits down and takes a break. The typical Cantor does two sets, one at the beginning of the service, and one at the end.

Cardinality: anything to do with a certain species of bird, the male of which displays fiery red-colored plumage.

Continuum: When one cleans with a vacuum cleaner, the act is referred to as "vacuuming." I will vacuum, she vacuums, I am vacuuming, we have vacuumed. However, if one stops this act and then resumes it, one may say "I will continuum."

R!: C'mon, silly, the middle initial of "Toys Я Us," with exclamation added.

Homeomorphic: I don't know what this is, but I'm pretty sure it has to do with that Brokeback Mountain stuff.

"Equipped with the product topology:" a line used by computer salesmen to get you to buy the top-of-the-line machine that's loaded with features.

_______

The definition is sublime. It makes so much sense. The phrase "each of its members is its own connected component" has to do with a gentleman's, ahem, member. Any man worth his salt should be pretty damn sure that his member is a connected component. Hey, remember the gal who cut her boyfriend's member off? His member wasn't connected anymore.

(Topology)

A set X is called totally disconnected if each of its members is its own connected component.

The most famous example of a totally disconnected set is the Cantor set. Note that it has the cardinality of the continuum ℵ=20. Totally disconnected sets can be big, even in R! This set is homeomorphic to the topological space {0,1}N; in fact, every set FN with F finite is totally disconnected when equipped with the product topology.

The most common method of representing numbers involves making strings out of a (small) finite alphabet, and using each position within the string as a multiplier with the symbol in that position. To be less abstract, a decimal positional notation is now in use by nearly every culture on Earth.

We'll get into the mathematics of positional notation later, but we'll start by saying the only really interesting mathematical property of a positional notation is its base, that is, the number of symbols in the alphabet that the strings are derived from. Our decimal positional system is base ten, that is, we use ten symbols (which vary from culture to culture) to represent numbers. In this system, a string position multiplies the contribution of a symbol there by a power of ten:

                    354 - represents "three hundred fifty-four"
three hundreds -----'||
five tens (fifty) ---'|
four -----------------'
The advent of computers has made systems with other bases (particularly two, eight, and sixteen) occasionally useful. (Now that I've introduced the concept, I'll drop the pretense of using words for numbers).

This little piggy goes to market...

It is far easier to perform complex calculations using positional notations than with physical analogues or earlier agglomerative symbols, primarily by allowing straightforward algorithms for doing them:

                            _______________
    735          DCCXXXV   |ooooo-----ooooo|
  x  49        x    XLIX   |ooooooo-----ooo|
  -----        ---------   |ooo-----ooooooo|
   6615        arrrrgh!!    
  2940                     |o-----ooooooooo| 
  -----                    |oooooo-----oooo|
  36015                         
                               go buy me a bunch of abaci...

Positional notations probably date from the earliest methods of counting using physical analogues. A crude form of positional notation was in use by preliterate people who, when counting objects using their fingers, ran out of fingers on one hand and used each finger on the other hand to represent a whole hands' worth of objects. Later methods involved putting small pebbles into cups or pits in the ground. Assigning positional importance to physical analogues continues with the use of the abacus in many places today.

Once people began writing things down, they invented symbols to stand in for numbers. The earliest number symbols resembled "tallies": New symbols for larger values piled up in the same way the smaller symbols did, and their order was not important.

The Babylonians invented the first written positional notation, using a base 60 (sexagesimal) system. Although keeping track of the system's 59 symbols seems unwieldy today, the symbols were actually aggolmerative collections of symbols representing 1, 5, and 20. At any rate, 59 symbols is not enough for a base 60 system, as 60 symbols are required. The missing symbol was a "zero" symbolizing no contribution at all for a particular position. As a result there was no way to tell the difference between 72, 702, and 720. However, the Babylonian system evolved to include a kind of zero: "placeholder" symbols eventually appeared in the middle of numbers, and Babylonian astrologers began using these placeholders at the beginnings and endings of numbers around the time of Buddha.

Two cultures later developed positional notations involving true zeroes: The Mayans who had a base 20 system, and Hindus who began using a base 10 system in the early centuries of the Common Era (to be more precise, the earliest archaeological evidence for decimal numbers in India comes from that time). This system spread west into Persia in the succeeding centuries, where it was picked up by Islamic scholars after the conquest. The most influential Islamic work on decimal numbers is an 825 treatise by Al-Khwarizmi. This was in turn read by Leonardo of Pisa, who introduced the notations and algorithms to Western Europe in his 1202 work Liber abaci. This method became known in Europe as augrim or algorism, attempts to represent the name al-Khwarizmi, and precusors to our modern word algorithm. During the Renaissance, algorism competed with other methods of calculation (including a graphical abacus) until the introduction of movable type made it the most feasible.

Other bases

Most of the time, other bases are for the sole purpose of demonstrating that you can use other bases. For example, base 3 uses the symbols {0, 1, 2}, and positions represent powers of 3 (1, 3, 9, 27, 81, …). 736 in base 3 would be 100021.

However, as described earlier, bases 2 ("binary"), 8 ("octal"), and 16 ("hexadecimal") have become useful for various aspects of computing.

Some mathematical stuff

Given an alphabet S = { S0, S1, S2,… Sb-1} containing b symbols, a number N can be represented by selecting one symbol Sdi from the set (a "digit") for each integer i, where

di = [b-iN] - [b1-iN]

where [x] represents "the greatest integer in x".

It should be evident that 0 <= di < b. Consequently, di is completely determined† for all i, and we can calculate the original number as N=Σbidi. Once all of the di have been determined we can represent the number of a string. For a positive integer:

…SdiSdi-1…Sd1Sd0

Nonintegers can be represented in this way, but since the digit strings could theoretically extend to infinity in both directions, it's necessary to "anchor" the string in some way. The usual way to do this is to introduce a "radix" symbol ("decimal point" in English-speaking countries) which is placed between Sd0 and Sd-1. This way, the symbols appearing after the radix symbol represent the fractional part of the number.

…SdiSdi-1…Sd1Sd0.Sd-1Sd-2

Every real number is "finite", and so there is an integer j such that bj-1 ≤ n < bj and so dk = 0 for all k≥j. It becomes unnecessary to include all of those "leading zeroes" in the string, but it's occasionally convenient to do so, and necessary after the radix symbol for representing fractions. In the fractional part (mantissa) of the string, digits can extend into infinity. If it's known that the digits past a certain point are all 0 they are customarily omitted, but some are frequently added to signify the precision of a measurement. If the fractional part contains an infinite repeating sequence, you will sometimes see a line drawn over the repeating part:

                         ______
1/7 = 0.142857142857…= 0.142857

Rational numbers

I went through all of this just so that I could show that every rational number has an infinite repeating sequence of symbols in its fractional part.

We can represent the reciprocal of an integer q > 0 in manner analogous to long division. If we define r0=1 and di+1=[bri/q]

br0 = d1q + r1
br1 = d2q + r2
br2 = d3q + r3
bri = di+1q + ri+1
where 0 ≤ di < b and 0 ≤ ri < q.

Multiply both sides of the last equation by 1/qbi+1:

ri/qbi = di+1/qbi+1 + ri+1/qbi+1

Since we set r0 = 1 at the outset,

1 / q = d1/b1 + r1/qb1
      = d1/b1 + d2/b2 + r2/qb2
      = d1/b1 + d2/b2 + … + di/bi

The definition of the "Greatest integer in" function determines di and ri for all i. In addition, ri completely determines di+1 and ri+1. Thus, ri = rj means di+1 = dj+1 and ri+1 = rj+1, such that whenever i < j and ri = rj, the series of digits Sdi..Sdj-1 is repeated forever.

Since q is an integer, all of the ri are integers. Thus, there are only q different values for the various ri. So eventually, ri=rj for some i, j where i < j, triggering the infinite sequence in 1/q's expansion.

Multiplying 1/q by some integer p will also result in an infinite sequence, but I'll leave that proof to you.

Not only that...

If the fractional part of a digit string contains an infinite repeating sequence, the resulting number must be rational. Let's say that a sequence of k digits repeats forever, starting at digit i. Thus:

r = d1/b1 + … + di-1/bi-1 
       + di/bi + … + di+k-1/bi+k-1
       + di/bi+k + … + di+k-1/bi+2k-1
       + di/bi+2k + … + di+k-1/bi+3k-1
       +…

We can group the terms such that the first group contains the first i-1 terms, and each group after that contains the next k terms. Then, we multiply the first group by bi-1/bi-1, and each successive group by bi+nk-1/bi+nk-1:

r = (d1/bi-2 + … + di-1/b0) / bi-1 
       + (dibk-1 + … + di+k-1b0) / bi+k-1
       + (dibk-1 + … + di+k-1b0) / bi+2k-1
       + (dibk-1 + … + di+k-1b0) / bi+3k-1
       +…

If we let S = d1/bi-2 + … + di-1/b0 and T=dibk-1 + … + di+k-1b0,

r = S / bi-1 + T / (bi-1 * Lim(n->∞) Σ(p=1..n, b-pk)) )

The sum of the geometric series Lim (n->∞)&Sigma(p=1..n, b-pk)) converges to 1/(bk-1) so

r = S / bi-1 + T / (bi-1(bk-1))
 = (S(bk-1)+T) / (bi-1(bk-1))

which is clearly†† a rational number.

Irrational numbers

A method similar to the one used for rational numbers can be used to show that a digit sequence exists for all real numbers, not just rationals. For a real number a, r0 = 1 and bri = di+1a + ri+1 still results in 1 / a = d1/b1 + d2/b2 + … + di/bi, but since a might not be integral, the various ri might not be integers and a pair of ri might never match. Since a repeating sequence appears only for a rational number, an irrational must be represented by an infinite non-repeating sequence.

Oh boy, the most tiresome flamewar on sci.math!

The positional representation of fractional real numbers introduces ambiguity into the system. The sum of the geometric series

Lim(j->∞)Σ(i=1..j, b-i)

converges to 1/(b-1), and so

Lim(j->∞)Σ(i=1..j, (b-1)b-i) = 1.

This means that whenever a number is a multiple of bi for some i, there are two ways of representing the number:

  1. If we assign di as above, dj = 0 for all j < i.
  2. However, when [b-iN] - [b1-iN] > 0, the limit above allows us to assign di = [b-iN] - [b1-iN] - 1, after which we can assign dj = b-1 for all j < i and still get the same number.

In base 10, this means the string "1" represents the same number as the string "0.99999…". Mathematicians care about this because: 1) cranks love to argue the point, and 2) it causes difficulty when mapping the power set of the integers to the reals.


†Well, opinions differ on the truth of this; see Law of Trichotomy for an explanation.

††A famous result of Euler's shows that for all integral a, b, there must be some k < a such that a divides bk-1. Remind me to node this some time.