Archive

Archive for 2012

How to lie with statistics, Merck style

In the pharmaceutical industry, where companies are making enormous bets with huge money and people’s lives, it makes sense that there are conflicting interests. The companies, who are in charge of testing their drugs for safety and for successful treatment, tend to want to emphasize the good and ignore the bad.

That’s why they are expected to describe beforehand how they are planning to do the tests. It stands to reason that, if they did a thousand tests and then only reported on the best ones, the public would get a biased view of the safety of their products.

For some reason, though, this standard doesn’t seem to be universally followed, and lying with statistics seems to be okay.

The newest example comes from Merck (see Pharmalot article here), which changed its statistical methods on testing Vioxx for Alzheimer’s patients from an intent-to-treat analysis to an on-treatment analysis even though their stipulated plans were the former. And even though the standard in the industry is the former.

Intent-to-treat means you choose people and stick with them, even if they get off the drug for some reason. And on-treatment only counts people that stay on the drug the whole time.

The difference is one of survivorship bias; there may be a good reason someone gets off the drug, and that may be because they got sick, and maybe they got sick because they were taking the drug.

What’s the difference in this case? From the article:

A subsequent intent-to-treat analysis found that as of April 11, 2002, when the FDA approved Vioxx labeling, there were 17 confirmed cardiovascular deaths on Vioxx compared with five on placebo in the same two trials.

With their on-treatment analysis, though, they didn’t see an elevated risk. So as it turns out the actual heart attacks happened a couple of weeks after people got off the pill.

So what happened there? Why were they allowed to change their stipulated method? Why were they allowed to not report their stipulated, gold-standard method? That’s complete bullshit and it must mean that someone at the FDA is either insanely stupid or very rich. Or both.

I’ve written about this issue before, specifically here. Just let me remind you of how we might assess the damage done by Merck through their statistical shenanigans:

Also on the Congress testimony I mentioned above is Dr. David Graham, who speaks passionately from minute 41:11 to minute 53:37 about Vioxx and how it is a symptom of a broken regulatory system. Please take 10 minutes to listen if you can.

He claims a conservative estimate is that 100,000 people have had heart attacks as a result of using Vioxx, leading to between 30,000 and 40,000 deaths (again conservatively estimated). He points out that this 100,000 is 5% of Iowa, and in terms people may understand better, this is like 4 aircraft falling out of the sky every week for 5 years.

According to this blog, the noticeable downwards blip in overall death count nationwide in 2004 is probably due to the fact that Vioxx was taken off the market that year.

Finally, I’d like to reiterate my question, why are pharmaceutical companies allowed to do their own trials?

Categories: rant, statistics

HCSSiM Workshop day 9

A continuation of this, where I take notes on my workshop at HCSSiM.

First we had a few proofs from the previous night’s problem set, including a proof of Hall’s Marriage Problem using Dilworth’s Theorem.

Counting stuff

We then proved an uncountable set union a countable set is uncountable, with the help of Lior’s comment from yesterday.

Then we proved the Cantor-Schroeder-Bernstein Theorem, which states that if you have two injective maps f and g in the opposite directions: f:X \rightarrow Y and g:Y \rightarrow X, then you can construct a bijection between X and Y, and in particular the two sets will have the same cardinality. It’s not that hard – consider the orbits of points in X and Y under repeated applications of the two injective maps f,g, and if possible, by pulling back by f^{-1} and g^{-1}.

It doesn’t take much thinking to convince yourself that these orbits come in three forms: an infinite list in both directions, a finite loop, or an infinite forward path but a finite backwards path, where at some point you can’t pull back any more. In the last case you could get “stuck” in either X or Y. Since the orbits form a partition of all of the points, you can independently decide how to define the bijection depending on what that orbit looks like. Namely, it takes an element x in X to f(x) unless it’s an orbit that gets going backwards in Y, in which case you take x to g^{-1}(x).

Now that I think of it, I’m pretty sure this proof uses the axiom of choice, and according to the wikipage on this theorem it doesn’t need to, but I don’t know a proof which avoids it. The truth is I can never tell. Please explain to me if you can, how you can verify if you’re using the axiom of choice in an argument where you make infinitely many decisions.

Homomorphisms

We defined a homomorphism of groups and wrote it in both additive and multiplicative notation, because it turns out some of the students were getting confused. We proved very basic properties that follow such as the fact that the identity element goes to the identity element under a homomorphism, and the image of the inverse of an element is the inverse of the image. We ended by asking them to consider the set of homomorphisms between a cyclic group of 6 elements and itself, or between a cyclic group of 6 elements to a cyclic group of 7 elements.

Fermat and Wilson

Next we whipped out some theorems modulo a prime p. We looked at the action defined on numbers mod 17 when you multiply them by 3, and noticed it just scrambles the non-zero ones up (and of course sends 0 to 0). We proved that this is true in general. But this means that the product of all the non-zero numbers mod p is the same if we pre-multiply by any a, which means that product is equal to itself times a^{p-1}, and since it’s invertible that means a^{p-1} \cong 1 \, \, (mod \; p). That’s just a hop skip and jump away from Fermat’s Little Theorem, which states that a^{p} \cong a \, \, (mod \; p) for every number a.

Next, we wondered, what was that product of all those non-zero numbers mod p? It turns out that each of those nonzero guys is invertible, so if you pair each up with its inverse their contribution to the product is just 1, and the leftovers are just the guys who are their own inverse, which is only 1 or -1 (which we proved, and which is most definitely not true modulo 8). So the whole product is -1. That’s Wilson’s Theorem, but we called it Wilevson’s Theorem since Lev came up with the argument.

Categories: math education

Mathematicians know how to admit they’re wrong

One thing I discussed with my students here at HCSSiM yesterday is the question of what is a proof.

They’re smart kids, but completely new to proofs, and they often have questions about whether what they’ve written down constitutes a proof. Here’s what I said to them.

A proof is a social construct – it is what we need it to be in order to be convinced something is true. If you write something down and you want it to count as a proof, the only real issue is whether you’re completely convincing.

Having said that, there are plenty of methods of proof that have been standardized and will help you in your arguments. There are things like proof by contradiction, or the pigeon hole principle, or proof by induction, or taking cases.

But in the end you still need to convince me; if you say there are three cases to consider, and I find a fourth, then I’ve blown away your proof, even if your three cases looked solid. If you try to prove something by induction, but your inductive step argument fails going from the case n=16 to n=17, then it’s not a proof.

Ultimately, then, a proof is a description of why you think something is true. The first half of your training is to problem solve (so, come up with a reason something is true) and construct a really convincing argument.

Coming at it from the other side, how can you check that what you’ve got is really a proof if you’ve written down the reason you think it’s true? That’s when the other half of your training comes in, to poke holes in arguments.

To be a really good mathematician you need to be a skeptic and to walk around with a metaphorical gun to shoot holes in other people’s arguments. Every time you hear a reasoned explanation, you look for the cases it doesn’t cover or the assumptions it’s making.

And you do the same thing with your own proofs to help yourself realize your mistakes before looking like a fool. Because putting out a proof of something is tantamount to asking for other people to shoot holes in your argument.

For that reason, every proof that one of these young kids offers up is an act of courage. They don’t know exactly how to explain their thinking, nor do they yet know exactly how to shoot holes in arguments, including their own. It’s an exercise in being wrong and admitting it. They are being trained to get shot down, to admit their mistake, and then immediately get back up again with better reasoning. The goal is to get so good at being wrong that it doesn’t hurt, that it’s not taken personally, and that it’s even fun to be wrong and to improve your argument.

Not every person gets trained in being wrong and admitting it. I’d wager that most people in the world, for most of their professional lives, are trained to do the opposite in the face of being wrong: namely, to wriggle out of it or deflect criticism. Most disciplines spend more time arguing they’re right, or at least not as wrong, or at least they have different mistakes, than other related fields. In math, you can at the most argue that what you’re doing is more interesting or somehow more important than some other field.

[I’ve never understood why people would think certain math is more important than other math. It’s almost never on the basis of having applications in the real world, or helping people in some way. It’s just some arbitrary snobbery, or at least that’s how it’s seemed. For my part I can’t explain why I love number theory more than analysis, it’s pure sense of smell.]

Most people never even say something that’s provably wrong in the first place. And that makes it harder to prove they’re wrong, of course, but it doesn’t mean they’re always right. Since they’ve not let themselves get pinned down on a provably wrong thing, they tend to stick with their wrong ideas for way too long.

I’m a huge fan of skepticism, and I think it’s generally undervalued. People who run companies, or universities, or government agencies, typically say they like healthy skepticism but actually want people to drink the kool aid. People who are skeptical are misinterpreted as being negative, but there’s a huge difference: negative means you’re not trying to solve the problem, skeptical means you care enough about the problem to want to solve it for real.

Now that I’ve thought about the training I’ve received as a mathematician, though, and that I’m now giving that training to these new students, I’ll add this to my defense of skepticism: I’m also a huge fan of people being able to admit they’re wrong. It’s the flip side of skepticism, and it’s why things get better instead of stay wrong.

By the way, one caveat: I’m not claiming that mathematicians are any better at admitting they’re wrong outside a strictly logical sphere.

Categories: math education, rant

HCSSiM Workshop day 8

A continuation of this, where I take notes on my workshop at HCSSiM.

We first saw presentations from the students from last night’s problem set. One was a four line proof that the last two digits of 17^{2012} are 61. The other was a beautiful proof that the only real numbers that have more than one decimal representation are rationals of the form a/10^k for some integer k.

Dillworth’s Theorem Revisited

After going over examples of chains and antichains, and making sure we knew that there must be at least as many chains as there are elements in an anti-chains if we want the chains to cover our set, we set up a proof by induction on the number of elements in our set. The base case is easy (one element, one chain, one element in a maximal anti-chain) and then to reduce to a smaller case we remove any maximal element m. Note this just means there’s nothing above it. But now the inductive hypothesis holds, and we cover the remaining set with chains C_1, C_2, \dots, C_k and moreover we define a_i \in C_i to be the largest element in C_i such that it is contained in some maximal anti-chain.

We then stated two lemmas. The first is that any maximal anti-chain must have a unique element in each chain $C_i$, and the second is that, after defining the elements a_i as above, they form a maximal anti-chain. We treat these lemmas as black boxes for the proof (the first we did yesterday, the second is on problem set tonight).

Now we put back the removed point m. There are two cases, either m is incomparable to any of the a_i‘s or it is comparable to at least one of them.

In the first case, we have an anti-chain of size k+1, namely the a_i‘s plus m, and we can form a (k+1)st chain consisting of just m itself.

In the second case, m is comparable to some a_i. Since m is maximal, and since a_i is maximal in its chain C_i with respect to being in a maximal antichain, we can form a new chain which is just the same thing as C_i for a_i and below, and goes straight up to m from a_i. Note we may be missing some stuff that used to be above a_i in C_i. But that doesn’t matter, because if we remove this new chain, we see that the maximal anti-chains leftover is only of size k-1, so by Strong Induction we can cover it with k-1 chains, and then bring back this chain to get an overall covering with k chains.

Dihedral Groups

After reviewing the formal definition of a Karafiol (group), we used rotations and flips of a folder and then of a regular 5-gon to come up with a non-commutative group. We defined a subgroup and explored the different subgroups of the dihedral group as well as other examples we’d seen coming from modular arithmetic.

Euler’s Totient function

We introduced the number of positive integers less than n relatively prime to n as a function of n called \phi(n) and wrote out a table up to 17. We observed that \phi(p) = p-1 and that, from the Chinese Remainder Theorem, we also could infer that for gcd(m, n) = 1 we have \phi(mn) = \phi(m) \phi(n). Writing out a prime factorization for n = \prod_i p_i^{a_i} we realized we’d have a formula for \phi(n) if we just knew \phi(p^a) for any prime p and any positive a. We wrote out a picture and decided \phi(p^a) = p^a - p^{a-1}.

We ended by proving Euler’s formula by induction on the number of distinct prime factors:

\sum_{d | n} \phi(d) = n.

Categories: math education

HCSSiM Workshop day 7

A continuation of this, where I take notes on my workshop at HCSSiM.

The real numbers are uncountable

Today we used Cantor’s diagonal argument to prove that the real numbers aren’t countable. Namely, we assumed they were, and that we had a bijection f: \mathbb{N} \rightarrow \mathbb{R} and then proved it didn’t contain the real number YP, whose ith digit we defined to be 1 if the ith digit of f(i) is 7, and 7 otherwise. One of our students pointed out that there are some numbers that have more than one decimal expansion, so the general argument that a certain decimal expansion isn’t in the list isn’t a total proof. But we decided that we’d be able to prove on homework that the only numbers that have more than one representative in decimal expansions are rational with their denominators divisible only by 2 and 5, which is not the case for the number YP we’d been talking about.

Then we wanted to show that \mathbb{R} has the same cardinality as \mathbb{R}\times\mathbb{R}. We tried to use a “splicing argument” where we create a new decimal expansion out of two decimal expansions (where the odd digits correspond to the first and the even to the second), but then we decided this map isn’t well defined, since we still have this “overlap” problem where 0.999… = 1 but the pair (0, 0.999...) doesn’t get mapped by the splicing map to the same decimal expansion as the pair (0, 1).

So then we decided to instead consider the set of decimal expansions, which is a bit bigger than the reals, and there the splicing map works. We reduced it to this case and are leaving it to homework to prove that the set of decimal expansions has the same cardinality as the reals. Although I don’t actually know how to do this without Cantor-Schroder-Bernstein, which we haven’t proven yet.

The Chinese Remainder Theorem 

We stated and proved the Chinese (Llama) Remainder Theorem. It was a nice example of a constructive proof that assumes it’s possible and then, when you follow your nose, it’s possible to completely characterize what the solution must look like, and then it falls out. We showed it described a bijection of sets \mathbb{Z}/\prod_i n_i \mathbb{Z} \rightarrow \prod_i \mathbb{Z}/n_i \mathbb{Z}. We haven’t talked about homomorphisms of groups yet though so we can’t prove it’s an isomorphism of groups.

Dilworth’s Theorem on chains and anti-chains in posets

We started the proof of Dilworth’s Theorem but didn’t finish yet (to be continued tomorrow). Turns out that this problem is really hard and that, moreover, there are lots of ways to convince yourself it’s true but be wrong about. The inductive proof in wikipedia, for example, is incomplete, but can be extended to a complete proof.

As an example, try to cover the following poset on the power set of five letters with 10 chains:

In particular you can see we solve a matching problem on the way, between subsets of size two and subsets of size 3 in the set of five letters, which isn’t the normal “take the complement” match, but rather is an inclusion of one in the other. I’m still wondering if there’s a more direct way to do this.

Categories: math education

Center for Popular Economics Summer Institute 2012

I’ll going to be giving a plenary talk on Tuesday, July 24th at the CPE Summer Institute 2012, which is being held this summer at Columbia the week of July 23rd – July 27th. I’ll be joined by Richard Wolff, an economist, radio show host, and author of multiple books, most recently Occupy the Economy. You can register for the Institute here (sliding scale).

The Summer Institute open to non-experts to teach them about the financial system and economics. They have two core courses, one based in the U.S. and the other international. From their web page:

The U.S. Economy/ Topics include

»     Roots of the economic meltdown and solutions
»     Speculation, finance and housing bubbles
»     Economy, race, class and gender
»     Economic histories – from personal to global
»     Labor and jobs
»     Democratizing the Federal Reserve and banks.
»     Economic alternatives, socialism and the solidarity economy

The International Economy/ Topics include

»     Roots of the economic meltdown and solutions
»     Brief history of the global economy
»     International trade, production and finance
»     The IMF, World Bank, WTO
»     Global climate change and the environment
»     Creating a new world economy

In addition to hosting this cool and open Summer Institute, the Center for Popular Economics also recently came out with a booklet explaining some economic history of the U.S. written for the non-expert; take a look here.

I’m about halfway through, and I’ve spotted things you usually don’t see in economics texts you might read in high school, such as the following phrase:

So What Caused This Crisis?
Neoliberal capitalism has had three features that both explain how it promoted 25 years of economic expansions and why it led to a massive crisis in 2008. First, inequality grew rapidly, as profits rose while workers’ wages actually fell. From 1979 to 2007, the average inflation-corrected hourly wage of non-supervisory workers declined by 1 percent, while inflation-corrected nonfinancial corporate profits after taxes rose by a remarkable 255 percent. While surging profits pleased the capitalists, it brought a problem: who could buy the growing output that comes with economic expansion? The solution was debt. Somehow, people would have to borrow more and more if a form of capitalism that brings skyrocketing profits and falling wages was to function.

I think it would be cool to have a typical high school “history of economics” text side by side with this one, and have students read both and argue them.

I’m going to try to go to as much of the Summer Institute as I can as a student. I hope I see you there!

Categories: finance

Toilet paper rant

I’ve been here at HCSSiM for almost exactly a week now, and I’ve been exclusively blogging about what mathematics we’ve been teaching this year’s brilliant crop of high school kids. Considering the fact that I usually have lots of opinions on important subjects such as financial reform, data science, and the incorrigible misuse of statistics, you might think I’m dying to also post about such things now that it’s Sunday and I’ve finally had time to catch up on some sleep.

You’d be wrong.

What I really need to vent about this afternoon is toilet paper dispensers. You see, I’ve been using lots of bathrooms with stalls and with those new-fangled huge toilet paper dispensers.

Do you remember in the olden days when a toilet paper dispensing system was relatively easy to understand? There’d be room for at most two rolls, the normal smallish kind, and if that wasn’t enough there’d be extra rolls somewhere for you to use. Granted, sometimes there weren’t, and sometimes there were but they got wet or dirty rolling on the floor.

Nowadays there are they enormous plastic cases which contain about 4 huge rolls of toilet paper, and I guess it’s a good thing in terms of how often toilet paper runs out, although it’s not an excellent idea in terms of the overall cleanliness of the bathroom, since you can mostly fill those fuckers up and leave for vacation.

But I’m not here to complain about dirty bathrooms. What I’d like to complain about is that these huge toilet paper dispensers, which are now about 3 feet in diameter, are for some reason always placed at the same level, at their center, as their older counterparts which contained two small rolls and opened up in the front.

The old dispenser would allow you to get toilet paper at approximately shoulder level. It was a pretty good system.

But these new ones dispense out at the bottom, so now we’re immediately talking about having to bend down to even find a corner of paper, usually blind. God forbid if it’s a new roll.

And once you catch hold of the ephemeral toilet paper corner, you have to then pull out some paper, which sounds easy, but your natural inclination is to pull on the paper by pulling towards yourself. This causes your tiny little corner of toilet paper to be immediately cut off by the serrated edge of the dispenser mouth.

So what you need to do, unless you are satisfied with one square inch of toilet paper (which I am not, in general), is you need to devise a two-handed system of pulling where one hand acts as a soft corner, almost like a ball bearing pulley, directly below the dispenser mouth, and the other hand pulls on it, at first straight down and then around the other hand and up.

A metaphorical second hand when desperately grasping for toilet paper

But mind you, you’re already stooping over to get the paper. So at this point you are basically on hands and knees trying to get more than one square inch of goddamned toilet paper.

People. People. People who install bathrooms, I’m talking to you right now.

Don’t you ever go to the bathroom yourself? Can’t you modify your installation procedure now that these big toilet roll dispensers have been around now for 10 years? Can we get them to dispense at shoulder level some time in the near future? Is this some way of keeping people from using too much toilet paper? If so, it’s not working. I always take too much because I always figure, “what the hell, now that I’ve constructed a pulley system I might as well see what she can do. I’ma gonna let her rip.”

Categories: rant

HCSSiM Workshop day 6

A continuation of this, where I take notes on my workshop at HCSSiM.

What is a group

We talked about sets with addition laws and what that really means. We noted that associativity seems to be a common condition and that some weird operations aren’t associative. Example: define a*b for a pair of integers (a,b) to be the sum a + b^2. Then we have:

(1*1)*1 = 2*1 = 3 but 1*(1*1) = 1*2 = 5.

We decided those things would make them crappy generalized addition operators. We ended up by defining what a group is, although we call it a “Karafiol” so that when our final senior staff member P.J. Karafiol arrives in a couple of weeks he will already be famous.

We showed that \mathbb{Z}/n \mathbb{Z} is a Karafiol and that, if you remove all of the congruence classes with numbers that aren’t relatively prime to n, you can also turn \mathbb{Z}/n \mathbb{Z} into a group under multiplication. I was happy to hear them challenge us on whether that would be closed under multiplication. The kids proved everything, we were just mediating. They are awesome.

Graphs

We had already talked about graphs (“Visual Representations”) as defined by vertices and edges. Today we talked about being able to put vertices in different groups depending on how the edges go between groups, so we ended up talking about bipartite and tripartite graphs. We ended up being convinced that the complete bipartite graph on 6 vertices (so 3 on each side) is not planar. But we haven’t proven it yet.

Special Lecture

Saturday morning we have only two hours of normal class, instead of 4, and we have a special event for the late morning. Yesterday Johan was visiting so he talked to them about the projective plane over a finite field, and how every line has the same number of points. He talked to them a bit about his REU at Columbia and his Stacks Project and the graph of theorems t-shirt that he wore to the talk. I think it’s cool to show the students this kind of thing because they are the next generation of mathematicians and it’s great to get them into online collaborative math as soon as possible. They were impressed that the Stack Project is more than 3000 pages.

Categories: math education

HCSSiM Workshop day 5

A continuation of this, where I take notes on my workshop at HCSSiM.

Modular arithmetic

We examined finite sets with addition laws and asked whether they were the “same”, which for now meant their addition table looked the same except for relabeling. Of course we’d need the two sets to have the same size, so we compared \mathbb{Z}/2\mathbb{Z} \times \mathbb{Z}/2\mathbb{Z} and \mathbb{Z}/4\mathbb{Z}. We decided they weren’t the same, but then when we did it for \mathbb{Z}/3\mathbb{Z} \times \mathbb{Z}/4\mathbb{Z} and \mathbb{Z}/12 \mathbb{Z}, and decided those were. We eventually decided it worked the second time because the moduli are relatively prime.

We essentially finished by proving the base case of the Chinese Remainder Theorem for two moduli, which for some ridiculous reason we are calling the Llama Remainder Theorem. Actually the reason is that one of the junior staff (Josh Vekhter) declared my lecture to be insufficiently silly (he designated himself the “Chief Silliness Officer”) and he came up with a story about a llama herder named Lou who kept track of his llamas by putting them first in groups of n and then in groups of m and in both cases only keeping track of the remaining left-over llamas. And then he died and his sons were in a fight over whether someone stole some llamas and someone had to be called in to arbitrate. Plus the answer is only well-defined up to a multiple of mn, but we decided that someone in town would have noticed if an extra mn llamas had been stolen.

Cardinality

After briefly discussing finite sets and their sizes, we defined two sets to have the same cardinality if there’s a bijection from one to the other. We showed this condition is reflexive, symmetric, and transitive.

Then we stopped over at the Hilbert Hotel, where a rather silly and grumpy hotel manager at first refused to let us into his hotel even though he had infinitely many rooms, because he said all his rooms were full. At first when we wanted to just add us, so a finite number of people, we simply told people to move down a few times and all was well. Then it got more complicated, when we had an infinite bus of people wanting to get into the hotel, but we solved that as well by forcing everyone to move to the hotel room number which was double their first. Then finally we figured out how to accommodate an infinite number of infinite buses.

We decided we’d proved that \mathbb{N} \times \mathbb{N} has the same cardinality as \mathbb{N} itself. We generalized to \mathbb{Q} having the same cardinality as \mathbb{N}, and we decided to call sets like that “lodgeable,” since they were reminiscent of Hilbert’s Hotel.

We ended by asking whether the real numbers is lodgeable.

Complex Geometry

Here’s a motivating problem: you have an irregular hexagon inside a circle, where the alternate sides are the length of the radius. Prove the midpoints of those sides forms an equilateral triangle.

It will turn out that the circle is irrelevant, and that it’s easier to prove this if you actually Circle is entirely prove something harder.

We then explored the idea of size in the complex plane, and the operation of conjugation as reflection through the real line. Using this incredibly simple idea, plus the triangle inequality, we can prove that the polynomial

 3z^{17} + 2iz^{12} - (1+3i)z^{10} + .017z^{5} - 17

has no roots inside the unit circle.

Going back to the motivating problem. Take three arbitrary points A, B, C on the complex plane (i.e. three complex numbers), and a fourth point we will assume is at the origin. Now rotate those three points 60 degrees counterclockwise with respect to the origin – this is equivalent to multiplying the original complex numbers by e^{\frac{i \pi}{6}}. Call these new points A’, B’, C’. Show that the midpoints of the three lines AB’, BC’, and CA’ form an equilateral triangle, and that this result also is sufficient to prove the motivating problem.

Categories: math education

HCSSiM Workshop day 4

A continuation of this, where I take notes on my workshop at HCSSiM.

As usual, we started with the students showing us solutions to their problem sets. Today one of them showed a sharp lower bound on the Fibonacci numbers, although he hadn’t proved it was sharp.

Arithmetic modulo n

Then we talked more about how we can talk about addition, and now also multiplication, with a finite set of symbols \{0, 1, 2, \dots n-1\}. Then we wrote out the multiplication tables for n = 6 and n = 7. The students noticed and proved that there is a multiplicative inverse for a modulo n if and only if gcd(a, n) = 1, using what we did yesterday with the Edwinian Algorithm and the way we turned it around to express gcd’s as linear combinations. We defined some notation and the natural map:

\mathbb{Z} \rightarrow \mathbb{Z}/n\mathbb{Z}.

Finally, we wrote down the subsets of \mathbb{Z} which map to each element of \mathbb{Z}/n\mathbb{Z}.

Posets and graphs

We went back to the idea of a partial ordering, and came up with a bunch of examples (including the set of integers under “divides evenly into”). We talked for a while about how to represent partial orderings, and finally settled on a graph. We talked a bit about poset chains and antichains, and then we formally defined a graph (we voted and decided to call it a “visual representation”).

The complex plane

The founder and director of the program is David Kelly. The program has been going for 40 years now and for maybe the first time ever Kelly himself isn’t teaching a workshop, so I’ve invited him to do some guest lectures in my workshop on complex geometry. It’s always a treat to watch him teach.

Kelly came in and built on the idea of “modding out by an integer” by definine \mathbb{R}[x]/(x^2 +1), which he described as “modding out by a polynomial”. He asked the students to investigate this idea and they eventually discovered that if x^2 +1 = 0, it also must be true that x^2 = -1, which allowed them to write every polynomial with as a linear combination of 1 and x, so as a + bx. Then they thought about the addition law and multiplication law and decided they had the complex plane. So we decided to start calling x the symbol “i“.

We then defined e^{i \theta} to be the point on the unit circle cos(\theta) + i sin(\theta), discarding once and forever the notation cis(\theta) (we justified this definition in last night’s problem set). We showed we could recover useful trigonometric identities that way (I skipped trigonometry myself and this is the only way I ever knew how to derive those identities) and that we could alternatively write any point on the complex plane in polar coordinates, so as r e^{i \theta}. Finally, we noted that if we multiply anything by the number r e^{i \theta}, we end up stretching it by r and rotating it by \theta.

We heard a funny story Kelly told us about taking a test to get his pilot’s license. He was given 30 minutes and lots of suggestions once to compute a heading which involved a calculation in polar coordinates. Since he was a mathematician he was too proud to accept the props they offered him, and finished with 29 minutes to spare. Once aloft though he quickly realized his calculation simply couldn’t be correct, but fudged the test by eyeballing it and following a highway. Turns out that pilots use due north as the axis along which the angle is zero, and then they go clockwise from there. I’m not sure what the moral of the story is, but it’s something like, “don’t be arrogant unless it’s a clear day and you have a backup plan.”

Categories: math education

Nim

Yesterday I gave a “Prime Time” talk here at HCSSiM, which is an hour long talk to the entire program. I talked about the game of Nim.

Nim is an old game (that’s the kind of in-depth historical information you’re gonna get from me) where you have a certain number of piles and each pile has a certain number of stones in it. There are two players and you take turns removing as many stones from any one pile as you want. The last person to remove a stone wins. Or, to anticipate my confusion later on, the person who gives back an empty game to their opponent wins.

You can play Nim online here with 3 or more piles and where the stones are matchsticks.

A bunch of the kids had never played so I got them to come to the board to play 2- and 3-pile Nim. Eventually it was  discovered that, as long as you start out with uneven piles with 2-pile Nim, you have a winning strategy by making them even. But it wasn’t clear how to win with 3-pile Nim, so we put that question in our back pocket for later.

I then changed it up a bit and put a rook on a chessboard, and said the point was to land on the top left corner, and you could only go up and to the left. They quickly realized it was just two-pile Nim again and the winning strategy was to get the rook on the diagonal. Then I switched it up further and made it be a queen, not a rook, and allowed the piece to move up, left, and diagonally up and left. This was harder.

We then decided that, when you have a game like Nim which is two-player, and the players share the same pieces (so not chess) and moves, and when the game gets smaller every time someone makes a move, then every position can be considered either “winning” or “losing”, by growing the game up from smaller games. If you can only get to winning positions, then you’re at a losing position. If there’s an option to get to a losing position, then you’re at a winning position. A consequence of this way of thinking of things is that a “game” can be described by the options you have when given a chance to play (along with the rules that define the options).

We then discussed adding two games A and B, which just means you get to play from either A or B but not both. We decided that 2-pile Nim is already of the form A + A, where A is 1-pile Nim. And furthermore, 1-pile Nim is pretty dumb – the winning strategy is always to just take away all the stones. But in spite of this, 5 stones is not the same as 6 stones for 1-pile Nim, since you can get to 5 stones from 6 but not vice versa.

Then I defined A ~ B if for all other (impartial) games H, A + H always has the same winner as B + H. It’s easy to see ~ is an equivalence relation, and that G + G is always winning (again, by mimicking your opponent’s moves). It’s also pretty easy to see that if A is winning, then A + G ~ G for all G.

But it’s a bad definition of ~ in that it seems to take an infinite amount of work to decide if A ~ B, since you’d have to check something for all possible H. We decided to improve this by proving that G ~ G’ if and only if G + G’ is winning. It is pretty easy to do this.

Then it was time for action, namely to prove the Sprague-Grundy Theorem which states that:

Every impartial game G has G ~ [N] for some N, where [N] denotes 1-pile Nim with N stones.

We prove this by showing recursively on the size of the game that N above (also called the “Nimber” of the position) is just the “mex” function, which is the minimum excluded non-negative integer. In other words, we designate the winning position as having N = 0, and then we grow the game up from there. If a given position can get to a “0” position, then its Nimber is at least 0 – in fact it’s the minimum number that it can’t reach in one move.

In particular, if you are at a position with Nim number non-zero, you can get to a zero position (i.e. a winning position), as well as any smaller Nim position, and if you are at a position with Nim number zero, you can only get to non-zero positions. This is similar to the losing-winning dichotomy except slightly more nuanced.

We then drew a Nimber addition table, which is the same as the chessboard problem with a rook. We used this to reduce 3-pile Nim to 2-pile Nim and worked out a winning strategy for the 3-pile Nim game (2, 5, 3).

Next we drew a Nimber table for the queen on a chessboard problem. We decided we knew how to play that game plus a 3-pile Nim game.

I was running out of time by this time but I ended with showing them a fast way to find the Nimber of the sum of a bunch of games whose Nimbers you already know, namely the bitwise XOR function. I didn’t have time to prove it (it’s not hard to see with induction on the number of games you’re adding up) but it’s easy to see this recovers our “mimicking” strategy with two games.

Categories: math education

HCSSiM Workshop day 3

A continuation of this, where I take notes on my workshop at HCSSiM.

Equivalence Relations and Partial Orderings

After going over many proofs of the geometry problem from last night’s problem set, I corrected the mistake in the “antisymmetric” property and we went through plenty of examples of equivalence relations and partial orderings. We ended with linear orderings, the real numbers, and the less than or equal relation.

Adding modulo n

Next we went back to the idea of maps f: X \times X \rightarrow X and got the kids to come up with a whole bunch of examples for X = \mathbb{Z}, such as f(a, b) = \lceil \frac{a}{2 \dot b +1} \rceil. We eventually got them to come up with stupider examples like (a, b) \mapsto a and (a, b) \mapsto a+b. Then we switched it up to the finite set [n] = \{1, 2, 3, \dots, n\} and got them to generalize addition. Since n really started to look like an identity element under this operation, we decided to define notation for the set \{0, 1, 2, \dots, n-1\}, which is like [n] but on its side.

Pigs-in-hole Principle

We introduced the pigeonhole principle (but since our camp mascot is a yellow, pig, we call it pigs-in-hole). We actually defined the slightly more general idea that, if you have n holes and k pigs which need to get put into holes, at least one of the n holes to contain at least \lceil k/n \rceil pigs. With that we proved that at least 2 people in New York City have the same number of hairs on their head, that five points in a 1 by 1 square are withing distance \sqrt{2}/2 of each other, that in a group of $n$ people at least two people will have had the same number of handshakes, and others.

Greatest Common Divisor

We asked the kids what the greatest common divisor of a and b is (denoted gcd(a, b)) and how to compute it. We spent a long time chasing down rabbit holes and proving other things that didn’t help us find the greatest common divisor but were true. For example, we showed that if you divide a and b by their greatest common divisor, you end up with numbers that are relatively prime. We even showed there are representations of a and b as products of primes, but since we couldn’t yet prove those were unique representations, we could use that to come up with a way to find the “common primes.”

Eventually we thought of a trick to reduce the problem, namely the division algorithm. Actually Edwin thought of the trick, and it eventually became the Edwinian Algorithm (not like it’s usually called, namely the Euclidean Algorithm). Edwin observed that gcd(a, b) = gcd(a - bq, b) when you write a = bq + r for 0 \leq r <b.

Once we had the Edwinian Algorithm, we realized we could go backwards and express gcd(a, b) as a linear combination of a and b, and we used that to prove a very important property of primes, namely that if p is a prime and p | ab then either p|a or p|b or both. This allowed us to show that the prime factorizations we’d found before for a and b were in fact unique up to relabeling, which we left for homework.

So it turns out I’m not going to be able to make the homeworks public, but we had an awesome problem set with lots of pigeon hole problems and Edwinian Algorithm problems. We asked them to decide which approach to calculating gcd(a,b) is faster, through the Edwinian Algorithm or via prime factorizations.

Categories: math education

HCSSiM Workshop day 2

A continuation of this, where I take notes on my workshop at HCSSiM.

The watermelon cutting problem revisited

We prove that the maximum number of pieces of watermelon you can cut with n slices, assuming a watermelon of dimension d, denoted by M_d(n), is given by:

\sum_{k=0}^d \binom{n}{k}.

First we proved it with a combinatorial argument, then with induction. I decided the first one is better because you figure out the answer as you go, whereas the induction route requires that you already know the formula. They both require you to use the recursion relation we talked about yesterday, and the first one involved writing a 2-d chart and showing how to unpack the value of M_d(n), using the recurrence relation, into paths going up to the top row consisting of all ones, and then the question becomes, “how many ways can you get to the top row?” and of course the answer is something like \binom{n}{k}, where you go up n times and left k times.

All pigs are yellow

Next we proved, using induction, that all pigs are the same color, and then we exhibited a yellow pig so a corollary was that all pigs are yellow. The base case is that a single pig is the same color as itself, and then assuming we have n pigs of the same color, we get to the statement that n+1 pigs are all the same color by first putting one pig in the barn, and then some other pig in the barn, and the leftover pigs are the same color as each of those so they’re all the same color.

This argument, of course, doesn’t work when you’re moving from “1 pig” to “2 pigs” and exhibits how careful you have to be with working through enough base cases so that your inductive step actually applies.

Strong Induction

We then went to using the Principle of Strong Induction (after showing that there’s no Principle of Induction over the real numbers). We proved that all numbers can be written as the sum of powers of 2, that the Fibonacci numbers grow exponentially, that every positive integer at least 2 is divisible by a prime, and that every planar polygon can be diagonalized using strong induction.

Notation

Incidentally, instead of “Strong Induction” the students voted to call it “Thor Induction”, and instead of the standard end-of-proof symbol, which is a box with an “x” inside, we voted to use the symbol “(see next page)”. They had lots of fun with that one. As a corollary, they decided that if they wanted someone to actually see the next page, they’d use the “Q.E.D.” symbol.

Cross product of Sets

Finally, we talked about the notation X \times Y, which denotes the cross product of sets, and made a bunch of examples, mostly of the form X \times X, specifically when X = \mathbb{R} and \mathbb{Z}, which we then drew as the plane and the lattice points. We ended by showing an injection from \mathbb{Z} into the lattice points, which incidentally showed that \mathbb{Z} and \mathbb{Z} \times \mathbb{Z} have the same cardinality, which we didn’t really define but we will.

Categories: math education

HCSSiM workshop day 1

So I’ve decided to try to explain what we’re doing in class here at mathcamp. This is both for your benefit and mine, since this way I won’t have to find my notes next time I do this.

Notetakers

We started the math, after intros, by assigning note-takers. In one row we wrote down the students’ names (14 of them), and in the other we wrote down the numbers 1 through 14. We drew lines from names down to numbers. These were the assigments for the days they’d take notes.

But to make it more interesting, we added pipes between different vertical lines. The pipes can be curly (my favorite ones were loopedy-loop) but have to start at one vertical line and end at another at “T” crosses.

Then the algorithm to get from a name to a number was this: start at the name, go down the vertical line til you hit a “T”, follow the “T” pipe til you hit another vertical line, and then go down.

This ends up matching people with numbers in a one-to-one fashion, but why? We promised to prove this by the end of the workshop.

Map of Math

We next had the kids talk about what “math” is. We had them throw up terms and we drew a collage on the board with everything they said. We circled the topics and connected them with lines if we could make the case they were related fields. We drew lines from the terms to the topics that used that a lot – like the symbol \pi got pointed at Trigonometry and Geometry, for example. I think it was useful. Lots of terms were clarified or at least people got told they would learn stuff about it in the next few weeks.

Cutting Watermelons

Next, we asked the kids how many pieces you can cut a watermelon into with 17 cuts. Imagine the watermelon plays nice and stays the shape of a watermelon as you continue cuts, and you can’t rearrange the watermelon’s pieces either.

If you do a few cuts it quickly gets hard to imagine.

So go down to a 2-dimensional watermelon, which could be called a pizza or a flattermelon. We called it a flattermelon. In this case you’re trying to see how many pieces you can achieve with 17 cuts. But also you may notice that a single slice of a 3-d watermelon looks, to the knife’s edge, like you are spanning a flattermelon.

Similarly, you may notice that a cut of the flattermelon looks like a 1-dimensional watermelon, otherwise known as a flatterermelon. And there the problem is easy: if you have a one dimensional watermelon, i.e. a line, then n cuts gives you maximum n+1 pieces. But going back to a pizza a.k.a. flattermelon, any cut looks from the point of view of the knife like a 1-d watermelon, which is to say it is cutting n+1 regions into half assuming the lines are in general position. So we get a recursion. If we denote by M_d(n) the max number of pieces you can get in d dimensions with n cuts, then we can see that

M_d(n) = M_d(n-1) + M_{d-1}(n-1).

Since we know M_1(n) = n+1, this recursion relation generates everything, although not in closed form.

Notation

Next, I went on at length about the utility and frustration of notation. Namely, notation is only useful if everyone agrees on what it means. I like standard notation because it’s more, well, useful, but Hampshire is a place where kids absolutely adore making up their own notation. As long as we are consistent it’s ok with me, and I like the fact that they own it. So instead of the standard notation for “n choose k” we are using a pacman symbol with n inside the pacman and k being eaten by the pacman. We call it “n chews k”.

Combinatorial Argument

We talked about putting balls in baskets, and defined that pacman figure to be the number of ways we can do it. Then we proved the pascal’s triangle recursion relation using the argument where you isolate one basket and talk about the two cases, one where there’s a ball inside it and the other when there’s not. Then we identified Pascal’s triangle as being equivalent to this concept of counting. I described this as an example of a combinatorial argument, which I like because it doesn’t involve formulas and I’m lazy.

Induction

Finally, I introduced Mathematical Induction and did the standard first proof, namely to show the sum of the first n positive integers is \frac{n (n+1)}{2}.

Categories: math education

How much of data science is busy work?

I’m at math camp, about to start the first day (4 hours of teaching a day, 3 hours of problem session) with my three junior staff (last year I only had one!). I expect I’ll be blogging quite a bit in the next few days about math camp stuff but today I wanted to respond to this blog post, entitled “The Fallacy of the Data Scientist Shortage”. I found this on Data Science Central which I had never known about but looks to be a good resource.

The author, Neil Radan, makes the point that, although we seem to have a shortage of data scientists, mostly what they do can be done by non-specialists. Just as you waste your time during a plane trip on things like security, waiting to board, and taxiing, the average data scientist spends most of her time cleaning data and moving it around.

If I understand this post correctly, they are saying that, because data scientists don’t spend that much time doing creative stuff, they can be replaced by someone who is good with data.

Hmm… let’s first go back to the idea that data scientists spend most of their time cleaning and moving data. This is true, but what do we conclude from it? It’s something like saying concert cellists spend most of their time practicing scales and rosining their bows, and don’t do all that much actual performing. Or, you could compare it to math professors who spend most of their time meeting (or avoiding) students and not much time proving new theorems.

My point is that this fact of time management is maybe a universal rule. Or even better, it may be a universal rule for creative endeavors. If you’re a truck driver then you can fairly said you worked the whole time you drove across the country, at a pretty consistent pace. But if you’re doing something that requires thought and puzzling then the nature of things is that it isn’t an 8-hour-a-day activity.

It’s more like, as a data scientist, you work hard to see the data in a certain way, which takes lots of time depending on how much data you have, then you make a decision based on what you’ve seen, then you set up the next test.

And I don’t think this can be done by someone who is strictly good at moving around data but isn’t trained as a modeler or statistician or the like. Because the hard part isn’t the data munging, it’s the part where you decide what test to perform that will give you the maximum information, and also the part where you look at the results and decipher them – decide whether they are what you expected, and if not, what could explain what you’re seeing.

I do think that data scientists can and should be paired with people who are experts at data moving and cleaning, because then the whole process is more efficient. Maybe data scientists can be brought in as 2-hour-per-day consultants or something, and the rest of the time there can be some engineers working on their tests. That might work.

Categories: data science

Mixing colors: pigment vs. light

Today we will address another topic in a list of “things I’m kind of ashamed I don’t understand considering I am a professional scientist of sorts” (please make suggestions!).

Why is it that when you mix light blue (cyan) and yellow paint you get green paint, but when you mix cyan and yellow light you get white light?

Unlike with yesterday’s analemma post, where I couldn’t find a satisfactory write-up on another blog, today’s blog is actually pretty nicely explained and beautifully illustrated here. I will crib their illustrations and summarize the explanations but it’s really out-and-out plagiarism for the moment.

First, you’ve got the so-called “hue wheel” (which sounds more sophisticated than “color wheel”, don’t you agree?):

This is illustrating the following. There are three basic pigments: yellow, cyan and magenta. There are three basic colors of light, namely green, blue, and red. And if you mix the fundamental pigments pair-wise (as in, you get paints and mix them) you get the fundamental colors of lights.

And vice versa as well, although this time you’re mixing as in splicing them together but keeping them separate, like we use pixels on our screen. This means, specifically, that you can combine green and red to get yellow. That’s majorly unbelievable until you see this miraculous picture, also from this webpage:

See how that works? I just can’t get over this picture. The little piece of yellow on the left is just stripes of green and red. Really incredible. The purple I get because it’s blue and red just like it’s supposed to be.

So, why?

The first thing to understand is that this isn’t just a relationship between us and the object we are looking at. It is instead a three-part relationship between us (or more specifically, our eyes), the object, and the sun (or some other source of light, but it’s more traditional in explanations like this to use fundamental, macho objects of nature like the sun).

Nothing can happen without a source of light. Which begs the question, what is light anyway? Again a picture stolen from here:

The prism separates the white light into various wavelengths, where red is at 700 nanometers and violet at 400 nanometers. More on the visible spectrum here. Note that the hidden difficulty here is why a prism does this, which is explained here.

So when an apple looks red to us, we have to imagine white light from the sun hitting that apple, and the key is that the skin of the apple is absorbing everything except the red light:

That thing on top is the sun, and the thing on bottom is your eyeball. The point is the red part of the light is reflected off the apple skin into your eye. And even though white light from the sun is the whole spectrum, we are denoting it when just the fundamental three colors of light because other colors can be made from those. And this can be corroborated by looking at your computer screen with a magnifying glass, where you will see that the white background is actually made up of little pixels of green, red, and blue.

By the way, we are again sidestepping the actual hard part here, namely why some surfaces such as apple skins reflect some colors like red. I have no idea. But I don’t feel as guilty about not understanding that.

Finally, back to the first question, of why cyan and yellow paint make green whereas cyan and yellow light make white. Turns out the light one is actually easier, since our second picture above shows us that yellow light is actually a mix of red and green, and when you add cyan, you now have all three fundamental colors of light, which gives us white light.

If you have cyan paint, then it is reflecting blue and green light, so absorbing red light. If you have yellow paint then that’s a material which is reflecting both green and red, so absorbing blue. For some weird reason (a third moment of stuffing things under the rug), the mixture of the paint is additive on absorbing things, so absorbs both blue and red, leaving only green reflected.

In the end we get a kind of mini De Morgan’s Law for color.

I’ve convinced myself that, modulo the following three questions I understand this explanation:

  1. How does a prism separate white light into the colors really?
  2. How do different surfaces decide which lights to reflect and which to absorb? And a related question from Aaron, why do colors fade when they’ve been in the sun?
  3. Why is “absorbing light” an additive procedure when you mix materials? I feel like if I understood 2 then I’d get 3 for free.
Categories: musing

Analemma

Today is a day of new things, since I finished my last day at my job yesterday and I’m going to math camp tomorrow. It’s exciting, and I’m going to kick off this first day of new things with a silly but fun thing I recently learned about the earth and the sun.

Some people know this already, but some people don’t, so sorry in advance if I bore you, but it’s super interesting the first time you think about it.

Namely, have you ever noticed, on your globe, a weird figure eight looking thing?

Nobody could be blamed for their curiosity, because there are so many important looking notches and then of course there’s the phrase “Equation of Time” next to it looking both pompous and intriguing. What is that thing??

After a few moments of contemplation, you’ve probably noticed there are months indicated, and since it’s a closed loop it’s probably describing something that is periodic with a one year period. Plus there are two axes, the vertical axis looks to be measured in degrees and the horizontal is called the “scale of time”.

Whenever I see north/south degrees I think of the earth’s tilt, and when I see something about time, it makes me think about how we measure time, which is vague to me, but probably has something to do with the sun, and orbiting around the sun, and spinning while we do it, again at a tilt. And if I want to be expansive at a time like this I’ll remember that the (pretty much circular) orbit of the earth lies on some plane where the sun also lives.

Now as soon as I get to this point I get nervous. What is time, anyway? How do we know what time it is? What with time zones, and daylight savings time, we’ve definitely corrupted the idea of it being noon when the sun is at its highest in the sky or anything as definitive as that.

So let’s imagine there are no time zones, that you are just in some specific place on the earth. You never move from that spot, because you’re afraid of switching time zones or what have you, and you’re’r wondering what time it is. If someone comes by and tells you it’s daylight savings time and to reset your clock, you tell them to go to hell because you’re thinking.

From this vantage point it’s definitely hard to know when it’s midnight, but you can for sure detect three things: sunrise, high noon, and sunset. I say “high noon” to mean as high as it gets, because obviously if you’re way north or way south of the equator the sun will never be totally overhead, as I noticed from living in the northeast my whole life.

But wait, even if you’re at the equator, the sun won’t be directly overhead most of the time. This goes back to the tilt of the earth, and if you imagine your left fist is the sun and your right fist is an enormous earth, and you tilt your right fist and stick your finger out and make it move around the sun (with your finger staying stuck out in the same direction because the tilt of the earth doesn’t change). As you imagine the earth spinning, you realize a point on the equator is only going to be directly in line to see the sun straight overhead about twice a year, and even then only if things line up perfectly.

Similarly you can see that, for any point between the equator and some limit latitude, you see the sun straight overhead twice a year – at the limit it’s once.

Going back to the point of view of a single person looking for high noon at a single place, we can see the height of the sun when it reaches its apex, from her perspective, is going to move around every day, possibly passing overhead depending on her latitude.

This is starting to sound like a periodic loop with a one-year period – and it makes me think we understand the x-axis. But what’s with the y-axis, the so-called “scale of time”?

Turns out it’s a definition thing, namely about what time noon is. Sometimes it takes the earth less time to spin around once than other times, and so the definition of “noon” can either be what we’ve said, namely “high noon,” or when the sun is at its highest in the sky, or you could use a clock, which has, by construction, averaged out all the days of the years so they all have the same length (pretty boring!). The difference between high noon and clock noon is called the equation of time.

By the way, back when we used sundials, we just let different days have different lengths. And when they first made clocks, they adjusted the clocks to the equation of time to agree with sundials (see this). It was only after people got picky about all their days having the same length that we moved away from sundial time. So it’s really just a cultural choice.

But why are some days shorter than others in terms of high noon? There are actually two reasons.

The first one, quaintly named “The Effect of Obliquity,” is again about the tilt. Imagine yourself sitting at the equator, looking up at the sun. It might be better to think of your position as fixed and the sun as going around the earth. And for that matter, we will assume the orbit of the earth around the sun is a perfect circle for this part.

Then what is being held constant is the spin of the tilted earth, or in other words the speed of the sun in the sky from the point of view of an observer on earth (this point is actually not obvious, but I do think it’s true because we’ve assumed a fixed tilt and a perfectly circular orbit. I will leave this to another post).

You can decompose this motion, this velocity vector, at a given moment, into two perpendicular parts: the part going in the direction of the equator (so the direction of some ideal sun if there were no tilt to the earth) and the part going up or down, i.e. in a right angle to the equator. Since we already know the sun doesn’t stay the same height all year, we know there has to be some non-zero part to the second part of this vector.

But since we also know the total vector has constant length, that means that the first vector, in the direction of the equator, is also not constant. Which means the length of the days actually varies throughout the year. The extent to which it does vary is approximated by a sin curve (see here)

The second reason for a varying length of a day, also beautifully named “The Effect of Orbital Eccentricity,” is that we don’t actually have a circular orbit around the sun- it’s an ellipse, and the sun is one of the two foci of the ellipse.

The thing about the earth being on an elliptical orbit is that it goes faster when it’s near the focus, which also causes it to spin more, due to the Conservation of Angular Momentum, which also makes an ice skater spin faster when her legs and arms are close to her body. Update (thanks Aaron!): no, it doesn’t cause it to spin more, although that somehow made sense to me. It turns out it just traverses a larger amount of angle with respect to the sun that we would “expect” because it’s moving faster. Since it turns as it moves faster, the day is shorter than you’d expect (this only works because of the way the earth spins – it’s counterclockwise if you’re looking down at the plane on which the earth is orbiting the sun clockwise). We therefore have faster days when we are closer to the sun.

When you add up these two effect, both approximated by sin curves, you get a weird function.

This is the “x-axis” of the analemma.

You can take a picture of the analemma by shooting a picture of the sun every day at noon, like these guys in the Ukraine did.

And by the way, you can use stuff about the analemma to figure out when sunrise and sunset will be, and why on the longest day of the year it’s not necessarily the day of the earliest sunrise and latest sunset.

And also by the way, there are lots of old things written about this stuff (see here for example) and there’s an awesome CassioPeia Project video (here uploaded on YouTube) explaining how all of this stuff varies over long periods of time.

Categories: musing

Is a $100,000 pension outrageous?

There are lots of stories coming out recently about how public workers, typically police or firefighters, are retiring with “outrageous” pensions of $100,000. Here’s one from the Atlantic. From the article:

That doesn’t frustrate Maviglio, who insists that “people who put their lives on the line every day deserve a secure retirement.” But do they “deserve” more than twice the US median income? Do they “deserve” the sum the average California teacher makes, plus $32,000? Do they “deserve” pensions far higher than the highway workers whose jobs are much more dangerous? These aren’t idle questions, given the public safety worker retirements we can expect in the near future.

Okay, let’s go there. If the median income in the country is 38,000, then $100,000 is a lot. But the median income in the communities where these retired firefighters live is sometimes much higher. For example, in Orange County, where the pension system is getting lots of flak, the median incomes can be seen here. In only one community out of is it below $50,000, and in 8 it’s above $100,000. So if you look at it that way then it doesn’t seem so outrageous.

And maybe we should be paying our teachers and our highway workers more, for that matter.

Point #1: California is a rich state, and it costs a lot of money to live there.

Now let’s move on to articles like this, which frame the issue in a very specific way. The title:

Police and Firefighter Pensions Threaten Government Solvency

How about all the other things that have contributed? Why are we blaming these guys, who have worked all their lives to protect their community? Why aren’t we blaming the mafia behind the muni bond deals, or sometimes even the local politicians as well?

Point #2: This is all a political blame game, trying to manipulate you from thinking about who are the actual crooks behind the scenes here.

My momma always said double down, and this is the ultimate double-down opportunity. Instead of looking for where the money went, or why it was handled so badly, we are going to blame the guys on taking the boring public servant job, and doing it for their adult lives, and trying to retire. Basically, we are blaming them for being right, for making the better choice between public and private.

Point #3: They made the right choice and we can’t swallow it because we thought our whole lives they were suckers for working in public service instead of in finance.

And what about that? Why do we compare $100,000 pensions to median incomes but not to golden parachute retirement packages of failing CEOs? Where’s the real outrage? Here’s another list of some seriously outrageous golden parachutes.

Maybe it’s because we feel like private pay is not our business, as taxpayers. It’s a different arena, and we have no right to judge. Let me remind you then that we taxpayers paid for bonuses at too-big-to-fail banks:

Point #4: These pensions don’t look very big when you compare them to what happens in the private sector.

And yes, I’m talking about the extreme cases, but so does everyone else when they talk about “outrageous pensions”, so it’s extreme-case apples to extreme-case apples.

Categories: finance, rant

Free online classes: the next thing

I love the idea of learning stuff online, especially if it’s free.

So there’s this place called Code Academy. They teach you how to code, online, for free. They crowdsource both the content and the students. So far they focus on stuff to build websites, so javascript, html, and css.

I just found out about Udacity (hat tip Jacques Richer), which also seems pretty cool. They offer various classes online, also free unless you want an official certificate saying you finished the class. And they have 11 courses so far, including this one on basic statistics with Professor Thrun.

Then there’s Coursera, which is starting to have quite a few different options for free online classes. The thing I’d like to bitch about with this is that Andrew Ng’s Machine Learning class, which I took when it came out last year, is not being offered currently, which makes me confused. Why does it need scheduling if it’s already been made?

I also just discovered openculture, which lists lots of free online courses. When you search for “statistics,” it returns the Udacity course I already mentioned as well as a Berkeley stats course on YouTube, among others.

I know this stuff is the future, so I’m hoping there continues to be lots of competition from various small start-ups. We are bound to profit from such competition as a culture. What I’m worried about is that the model goes from “free” to “fee” if it gets crowded by large players who, say, pay their instructors a lot for the content.

Which is not to say the instructors shouldn’t get paid at all, but I hope the revenue can continue to come from advertising or through job matching.

The basic unit is risk

Today I’m going to gush over two excellent blog posts I read recently written over at Interfluidity. But first I’m going to state a pet theory of mine about what units we talk in.

In a mathematical sense, units make no difference. If I give you measurements in inches rather than feet, all I’m doing is multiplying by 12. If I say something in French rather than German, all I need is a translation and we’re talking about equivalent information.

But in a psychological sense, a choice of units can make an enormous different. Things sound bigger in inches, and sometimes you barely understand French and can make bad guesses.

I’d argue that speaking in terms of wealth is a mistake. We should instead speak in terms of risk. It’s a different unit, and it’s harder to quantify, but I think risk is what we actually care about. I claim it’s more basic than money.

For example, why are we afraid of not having money? It’s because we run the risk of not having resources to eat, sleep, or get medicine or treatment when we’re sick. If we didn’t have fears about this stuff then people would have a very different relationship to money. The underlying issue is the risk, not the money.

Financial markets putatively push around money, but I’d argue that why they exist and how they actually function is as a way to spread around risk. That’s why the futures market was developed, for farmers to have less risk, and that’s why the credit default swap market was created, to put a price on risk and sell it to people who think they can handle it.

It also kind of explains, to me at least, the weirdness of super rich people- people who have more money than they can ever use. Why do they continue to collect money so aggressively when they already have so much? My guess is that they are confused about their units- they think all their problems can be solved by money, but their remaining actual problems are problems of risk that can’t be controlled by money. Things like the fact that we all get old and die. Things like that people don’t like you if you’re an asshole or that your wife may leave you. These are risks that most people never get to the point of trying to solve through money, because they’re still stuck in a different part of reality where inflation could screw their retirement plans. But for super rich weirdos, we have the Singularity University where you get to learn how to transcend humanity and live forever.

I’m not making a deep statement here. I’m just suggesting that, next time you hear of a plan by politicians or regulators or Wall Street bankers, think not about where the money is flowing but where the risk is flowing.

A perfect example is when you hear bankers say they “paid back all the bailout”; perhaps, but note that the risk went to the taxpayers and is firmly fixed here with us. We haven’t given the risk back to the banks, and there doesn’t seem to be a plan afoot to do so.

Which gets me to Interfluidity’s first plan, namely to have the government protect up to $200,000 of an individual’s savings from inflation.

Now, on the face of it, this plan is not all that protective of the 99%, because it’s definitely benefiting people who have savings, where we know that the lowest 25% or so of the population is in net debt. Only people with savings to protect can actually benefit.

But if you think about it more, it is good for people like my parents, whose retirement from a state school does not rise with inflation, or more generally for people who have a fixed savings put aside for retirement. And it isn’t at all good for very rich people, who would see a benefit only on a small percentage of their savings (assuming it is possible, as Interfludity says it is, to outlaw the bundling of these inflation-protected accounts like some people now bundle life insurance policies).

Most economic policies in this country are made to benefit rich people, and are defended by saying we need to protect middle-class people nearing retirement with a modest nest-egg. As Interfluidity said, those middle guys are used as “human shields”. Very few policies go into to the weeks sufficiently to figure out how to protect that group without having outsized benefits at the top.

Said in terms of risk, this plan is pushing inflation risk to people who can handle it, and removing it from people who are extremely vulnerable to it.

Which brings me to the second post I want to rave about, namely this one in which Interfluidity dissects the lack of political will in the face of the current depression. From the post:

We are in a depression, but not because we don’t know how to remedy the problem. We are in a depression because it is our revealed preference, as a polity, not to remedy the problem. We are choosing continued depression because we prefer it to the alternatives.

The reason? Because no matter how much someone might say that we care about the middle class, the truth is we are protecting rich people from the risk of getting poor. We have, as he says, a population with individual power roughly weighted in proportion to their wealth (or, to be consistent with my theme, inversely proportional to their risk), and when you take a vote with those weightings, we get a “weighted consensus view,” manifested among the macroeconomists in charge of this stuff, that we should avoid inflation at all costs (ironic that the people with the least risk are also the people with the most influence).

In order to remedy this situation, we’d need to implement something like the inflation-protected bank accounts up to $200,000 for the individual. Then the weighted consensus may change – we might instead actually pull for a policy that would have some risk for inflation and would also possible create jobs.

But of course, in order to implement such a policy, we’d need to have the political will to change the risk profile, which goes back to the weighted consensus thing. Keeping in mind that this policy would push the risk to rich people, I’m guessing they wouldn’t vote for it.

On the other hand, smallish savers would. So it’s not a mathematical impossibility, because there may be enough people in favor of the inflation-protection plan to make it happen, and then the second question, of how to get us out of the current depression, would be easier to address. I’m definitely in favor of trying.

Categories: finance