## Supply and Demand

Until recently, I thought I understood the concept of supply and demand functions pretty well: for each possible price of a good, we look at how much of the good consumers would collectively purchase if it was offered at that price, and how much of it producers would sell if offered that price for it. Sounds simple enough.

### Ambiguity

Problem is, the amount of something that a consumer will buy or that a producer will sell depends on more factors than just its price. So in order to determine how much of a good would be demanded or supplied in a counterfactual where its price changes, we also need to know about these other factors. You might want to let everything else stay the same, but this cannot be done. Changing the price of a good must come with some other changes as well. For instance, if the good has a substitute, then the price and quantity of the substitute cannot both stay the same when the price of the good changes.

So for supply and demand functions to be well-defined, we need a counterfactual model, which tells us what exactly is happening in the counterfactual where we set the price of a good and ask how much is demanded or supplied.

Given multiple goods (for simplicity, two goods, $x$ and $y$), we could consider joint supply and demand functions for them, which take as input a pair consisting of the price of the first good and the price of the second, and outputs a pair consisting of the quantities of each good supplied (for the supply function) or demanded (for the demand function) given those prices. So if $p_x$ and $p_y$ are the prices of $x$ and $y$, respectively, then $D(p_x,p_y)_x$ and $D(p_x,p_y)_y$ are the quantities demanded of $x$ and $y$, respectively, and $S(p_x,p_y)_x$ and $S(p_x,p_y)_y$ are the quantities supplied of $x$ and $y$, respectively. Given joint supply and demand functions, how should we extract supply and demand functions for each good individually? That is, how do we determine the demand function $D_x(p_x)$ and supply function $S_x(p_x)$ for the good $x$? To answer this question, we need to know what we mean by supply and demand functions, and answering this question provides similar information about what we mean by supply and demand functions as providing a counterfactual model does.

I'm going to make a couple simplifying assumptions for this discussion, not because they are true, but because they are convenient. First, producers and consumers are two entirely separate groups of people. Consumers don't get the money they spend on consumption by producing something, and producers don't spend the money they get for production on consuming anything else. Consumers just magically have money, and producers just hoard money so they can swim in it like Scrooge McDuck. Second, a system of equations has exactly one solution whenever I want it to.

Wikipedia suggests that in the counterfactuals considered for demand functions, prices of substitutes and complements should stay the same. This gives an answer to the question of how to extract individual demand functions from a joint demand function. Let $p_y^*$ be the equilibrium price of $y$ (i.e. the price such that there is some $p_x^*$ such that $D(p_x^*,p_y^*)=S(p_x^*,p_y^*)$). Then $D_x(p_x)=D(p_x,p_y^*)_x$. A counterfactual model to go along with this is that all producers disappear, and then a monopoly enters the market and offers to the consumers an unlimited supply of the same goods for the same prices that were being offered previously, except for the price of the good in question, which they change. The analogous assumption to make for supply functions is that the prices of other goods that the producers could be making instead (and any goods that can be most efficiently produced along with the good in question) stay the same. Then individual supply curves can be obtained from joint supply curves by $S_x(p_x)=S(p_x,p_y^*)_x$. And a counterfactual model is that all consumers disappear, and a monopsony enters the market and offers to purchase from the producers unlimited quantities of goods for the same prices the consumers were previously paying, except for the good in question, which they offer a difference price for.

End of story? No; there are other answers that could be given to these questions. You may have noticed that the counterfactual models I suggested were not especially realistic, which could be considered a drawback, so let's look at some other possibilities.

Suppose the government plans to tax or subsidize a good, and we want to know what effect this will have on the quantity of the good, the price paid by consumers, and the price received by the producers. The usual story for how to figure this out from supply and demand functions $S$ and $D$ is that, where $t$ is the amount of tax (or negative the amount of subsidy) on each unit of the good, if $p$ is the price paid by the consumer, then $p-t$ is the price received by the producer, and the quantity demanded is $D(p)$ and quantity supplied is $S(p-t)$; since quantities supplied and demanded must be equal, then we can just solve the equation $D(p)=S(p-t)$ for $p$ to get the price paid by the consumer, and the common value of $D(p)$ and $S(p-t)$ will be the quantity traded. What do we have to mean by supply and demand functions in order for this story to be correct? The counterfactual model here is clear: the government institutes a tax or subsidy of exactly the right amount so that the consumer/producer ends up paying/receiving a price equal to the input of the demand/supply function, respectively. In other words, $D(p)$ means quantity traded if the tax or subsidy is adjusted so that the price paid by consumers is $p$, and $S(p)$ means the quantity traded if the tax or subsidy is adjusted so that the price received by producers is $p$. To figure out the corresponding way to extract individual supply and demand functions from joint supply and demand functions, note that if the government taxes or subsidizes the good $x$ such that the price paid by consumers is $p_x^d$ and the price received by producers is $p_x^s$ (using superscripts as indices, not exponents), and $p_y$ is the price of the good $y$, then $S(p_x^s,p_y)=D(p_x^d,p_y)$, as these are the pairs consisting of quantities supplied of $x$ and $y$, and quantities demanded of $x$ and $y$, respectively. So $D_x(p_x)=D(p_x,p_y)_x$, where $(p_y,p_x^s)$ is the solution to $S(p_x^s,p_y)=D(p_x,p_y)$, and $S_x(p_x)=S(p_x,p_y)_x$, where $(p_y,p_x^d)$ is the solution to $S(p_x,p_y)=D(p_x^d,p_y)$.

Another possible counterfactual model is that a new agent enters the market and offers to buy or sell unlimited amounts of a good for a fixed price $p$, thus forcing $p$ to be the new market price. $D(p)$ is then the quantity purchased by the original consumers, and $S(p)$ is the quantity sold by the original producers. This happens, for example, with backing currency, where a government promises to sell as much of some good as anyone will buy at some fixed price, as a way of ensuring that money remains valuable. A related example is a fixed exchange rate, where a government pegs the price of its currency relative to a foreign currency at a fixed rate by exchanging the currencies in both directions at the pegged price with anyone who wants to trade them at that price, though this example is atypical in that the good being exchanged for money is also money. To figure out the way of extracting individual supply and demand functions from joint supply and demand functions corresponding to this counterfactual model, note that if the government pegs the price of a good $x$ but doesn't get involved in the market for another good $y$, then supply and demand for $y$ still must balance; that is, $D(p_x,p_y)_y=S(p_x,p_y)_y$, where $p_x$ is the pegged price of $x$ and $p_y$ is the market price of $y$. So the supply and demand functions for $x$ should be given by $D_x(p_x)=D(p_x,p_y)_x$ and $S_x(p_x)=S(p_x,p_y)_x$, where $p_y$ is such that $D(p_x,p_y)_y=S(p_x,p_y)_y$.

### Cases in which the appropriate notion of supply and demand functions is unclear

There are other situations that economists use supply and demand functions to describe, and each situation may need a different notion of what supply and demand functions mean, and the right way to define them for a given purpose might not always be obvious. For instance, suppose we want to predict the effects of price controls. This may only give meaning to the demand function on prices greater than or equal to the market price, and the supply function on prices less than or equal to the market price, since if supply and demand aren't equal due to price controls, then whichever one is smaller will determine the quantity traded, so it isn't clear that there's an objective way to say what the larger one should be.

Supply and demand functions are supposed to help us understand the effects of changes to supply or demand on price and quantity. The demand curve describes the set of (price, quantity) pairs that can be obtained by changes to how easy the good is to produce, and the supply curve describes the set of (price, quantity) pairs that can be obtained by changes to how desirable the good is. This is not completely well-specified, so what the supply and demand functions should be depends on which sorts of shocks should be considered to only affect demand for the good, or to only affect supply of the good. This isn't as straightforward as it might sound, since, for instance, two supply shocks might have different effects on demand because of having different effects on the market for a complement or substitute for the good; the supply shock can't have no effect on the market for the complement or substitute, because it affects the market for the original good, and the markets are linked, so some decision must be made about what assumptions to make about how a supply shock is supposed to affect the markets for other goods, which will depend on, for example, the extent to which the supply shock is due to producers leaving the market for a substitute or complement in order to produce the good in question instead, or vice-versa.

The area of the region bounded by the supply and demand curves and the price axis is used as an estimate of how much wealth is created by the market for the good (neglecting externalities). It isn't clear to me what way of defining supply and demand functions makes this the best metric, and it's not going to be perfect in any case, as it measures wealth in units of money, and the value of a certain amount of money to someone can vary depending on circumstances (for instance, on how much money they already have). Another problem is that areas of smaller regions in a supply and demand graph are used as measures of deadweight loss caused by market distorsions, but the nature of the market distorsion in question would require us to use one particular meaning of supply and demand functions in order to correctly describe the effects of the distorsion; if this meaning is not one for which area in the supply and demand graph accurately represents value, then accurately representing deadweight loss as the area of a region in a graph isn't possible.

### Example

A possible explanation for why these subtleties are not emphasized is that they don't make much difference in practice, as supply and demand functions tend to not depend too much on how exactly they're defined. But we can construct examples in which they are quite different.

Consider the supply function in the market for left shoes. Assume no one actually buys one shoe at a time, so no matter what the prices of left and right shoes are, the demand for each will be the same. Also assume that making left shoes and making right shoes is always equally easy for the producers, so they would shift entirely to producing just the more expensive chirality of shoe if the prices differed. That is, there is a demand function for pairs of shoes, $D_{pair}$, and the joint demand function for left shoes and right shoes is given by $D(p_{left},p_{right})=(D_{pair}(p_{left}+p_{right}),D_{pair}(p_{left}+p_{right}))$, since $p_{left}+p_{right}$ is the price of a pair of shoes. And there is a supply function for total shoes, $S_{shoes}$, and the joint supply function for left shoes and right shoes is given by $S(p_{left},p_{right})=(S_{shoes}(p_{left}),0)$ if $p_{left}>p_{right}$, $(0,S_{shoes}(p_{right})$ if $p_{right}>p_{left}$, and, if $p_{left}=p_{right}=p$, can be any $(q_{left},q_{right})$ such that $q_{left}+q_{right}=S_{shoes}(p)$ (if you don't like multi-valued functions, you can pretend we're working with a continuous approximation to this joint supply function).

The market prices for left shoes and for right shoes must be the same, since otherwise the producers would only make shoes of one chirality. Let $p^*$ be this market price for shoes, and now let's figure out the supply function $S_{left}$ for left shoes according to each of the three conventions I described in the section on possible answers.

If we are to assume that the prices of right shoes are constant, then $S_{left}(p_{left})=0$ for $p_{left}, and $S_{left}(p_{left})=S_{shoes}(p_{left})$ for $p_{left}>p^*$, since all production shifts to the more expensive chirality of shoe when we vary the price of left shoes.

If we assume that the government taxes or subsidizes left shoes such that the price received by the producer for left shoes is $p_{left}$, then the price received by the producer for right shoes must also be $p_{left}$, since consumers will still be buying both chirality of shoe, which means prices must adjust so that producers will still make both chiralities of shoe (this will mean that consumers will pay different prices for left and right shoes, with the difference being the amount of the tax or subsidy). So the producers will make $S_{shoes}(p_{left})$ shoes in total. Since the consumers will buy equal numbers of each, the producers must make equal numbers of each in order to sell them all, so $S_{left}(p_{left})=\frac{1}{2}S_{shoes}(p_{left})$.

If we assume a new agent offers to buy or sell unlimited numbers of left shoes for $p_{left}$, then again right shoes are going to end up going for the same price (except in the extreme cases where the original producers don't end up making any left shoes, or the original consumers don't end up buying any shoes, but let's suppose that doesn't happen). The original producers will supply $S_{shoes}(p_{left})$ shoes in total. The original consumers will buy $D_{pairs}(2p_{left})$ pairs of shoes. In particular, the original consumers will buy $D_{pairs}(2p_{left})$ right shoes, which is the same as the number of right shoes sold by the original producers, since the new agent does not buy or sell any right shoes. Thus, the original producers make $S_{left}(p_{left})=S_{shoes}(p_{left})-D_{pairs}(2p_{left})$ left shoes.

Of these three purported supply functions, the first is radically different from the others, in that the latter two are continuous (if $S_{shoes}$ and $D_{pairs}$ are), but the first is discontinuous at $p_{left}=p^*$. The relationship between the latter two supply functions for left shoes depends on the relationship between $S_{shoes}$ and $D_{pairs}$, but they certainly can be quite different.

One could argue that this example doesn't count, because a pair of perfect complements can just be considered a single good sold as a bundle (as left and right shoes typically are in practice). But complements that are not perfect complements also exist, and producers do sometimes have some ability to shift between markets, so approximate versions of this example should exist in real life.

### Joint supply and demand functions for infinitely many types of goods

We can, of course, consider joint supply and demand functions of arbitrarily many goods, instead of just two. For $n$ goods, $D(p_1,...,p_n)$ and $S(p_1,...,p_n)$ would be the $n$-tuples consisting of the quantities demanded or supplied of each good when $p_i$ is the price of the $i$th good. But sometimes goods can vary across a continuum of possibilities, and in such cases, it might not be appropriate to think of the market as just involving finitely many types of goods. For example, goods may vary by quality, with higher-quality goods being both more preferred by the consumers and harder to produce, and goods of different qualities substituting for each other. Or in the case of goods that can't easily be moved, like buildings, consumers may have preferences about their location, and location may affect how easily the goods can be produced. Quality and location can each be modeled as varying across a continuum of possibilities.

Where $X$ is some space of goods, prices can be represented as a function $p:X\rightarrow\mathbb{R}_+$, where $p(x)$ is the price of the good $x$. Quantity can be represented as a finite measure $\mu$ on $X$, where for $Y\subseteq X$, $\mu(Y)$ is the combined quantity of all goods in $Y$.

So, where $\mathbb{R}_+^X$ is the set of measurable functions $X\rightarrow\mathbb{R}_+$ and $M(X)$ is the set of finite measures on $X$, joint supply and demand functions for $X$ should be functions $S:\mathbb{R}_+^X\rightarrow M(X)$ and $D:\mathbb{R}_+^X\rightarrow M(X)$. The market equilibrium price function is the solution $p^*$ to $S(p^*)=D(p^*)$ ($S$ and $D$ should satisfy some conditions which should ensure that this exists), and the common value of $S(p^*)$ and $D(p^*)$ is the measure representing the market equilibrium quantities traded of types of goods in $X$.

## Green Bus

[Inspired by this map]

“Everyone up! Company meeting in thirty minutes! Breakfast available in the mess hall until then,” yelled Sergeant Tucker. Even in my groggy state, I realized instantly that something unusual was happening. Getting woken up after no where near enough sleep wasn't unusual; in fact, it was becoming the norm. I suspected that the Californian Peoples' State's attacks had been carefully timed to fuck up our sleep as much as possible. They might not even have been seriously trying to capture territory, just wearing us down.

But the lack of urgency in Sergeant Tucker's voice was unprecedented, and he was giving us time to eat breakfast before doing anything else. There was no sound of gunfire, too.

I dressed and followed the rest of the platoon out of the barracks and into the mess hall. It was bright out already, but was in the middle of my sleep anyway because of a grueling firefight we'd had in the middle of the night, in which we'd repelled an attack that had threatened to cut off our last remaining access to the canal.

After a breakfast that was no more appetizing than what we’d been eating, though oddly generous in quantity compared to the carefully rationed meals we’d been getting, we were hurried out of the mess hall and into the briefing room.

Captain Smith began the briefing, “We have surrendered to the Californian Peoples' State.” Hearing that was a relief, to be honest. We should have surrendered months ago, but Colonel Fitch was such a hardass I’d thought it was never going to happen.

Captain Smith continued, “We've reached a deal where, in exchange for surrendering the Sutter pocket, we will be given transportation to Reno instead of being taken prisoner. The first of the buses will be arriving in Sutter in about an hour. Bring your gun on the bus with you, and for God's sake don't ruin the deal by firing at any Californian soldiers. UN peacekeeping forces will be present to ensure the Peoples' State doesn't break the deal. The Californians will try to convince you to surrender individually to be taken prisoner instead of coming with the rest of us to Reno. Do not listen to them. Conditions in Reno are much better than here, so don't think you're better off in a prison camp than in Reno or anything foolish like that. If you have a family here in the Sutter pocket, they'll be coming with you; you're dismissed now, so you can go get them. Everyone else, pack up and reassemble here in twenty minutes. We'll be marching to the bus stop together. Dismissed!”

I was in shock. Why would the Peoples' State offer us this deal? If they'd just kept up the attacks for another couple weeks, we would have collapsed, and I would've thought the Californians would've caught onto this. Maybe the Californians were afraid that Free States of America forces in the Tahoe area would break the siege on the Sutter pocket soon? But at the rate the war's been going, that didn't seem likely, and besides, if Colonel Fitch suspected the same thing, there'd be no way he would have accepted the deal. Maybe the Californians were just bending over backwards to avoid incurring a few casualties (some of which would inevitably be civilians) in the process of taking the pocket by force. Or maybe it was a trick and we were all going to get taken prisoner anyway. The promise of UN peacekeepers made that last possibility seem somewhat unlikely; we all figured the UN was a bit biased in favor of the Peoples' State, but it was unlikely they'd let them get away with breaking an evacuation agreement.

Like most of us, I didn't have family with me in the Sutter pocket (or family at all, for that matter). I packed my things and regrouped with the others, and we marched into town. Another group of soldiers were already waiting at the bus stop when we arrived, and more groups joined us shortly thereafter. Soon, six green buses pulled up and stopped in front of us. It looked like the buses collectively just about fit all of us gathered in front of them, which was a small fraction of the total number of Free States of America soldiers in the pocket, even accounting for the fact that those with families in the pocket weren't present. Probably others were getting picked up in the small portion of Yuba City we still controlled instead of consolidating us all in Sutter first.

Some Californian soldiers and UN peacekeepers got out of the buses. There weren't very many UN peacekeepers; not enough to make much of a difference in a fight if anyone broke the agreement, anyway. But their presence was still useful, since no one wanted to piss off the UN.

A Californian officer held up a megaphone, and spoke, “To encourage you all to give up your arms instead of going to Reno, we're sweetening the deal. If you stay, instead of becoming a prisoner of war, you will gain the rights of California citizens, able to live and work freely in California, and exempt from the draft, with the only additional restrictions being that, until the war ends, you will not be able to own weapons, and someone'll check in on you occasionally to make sure you're not up to anything fishy. If you'd like to go to the war zone in Reno anyway, you may now board the buses. If you'd like to stay, you can just walk right past the buses and hand over your weapons to any Californian officer on the other side of the buses.”

The bus doors opened, and we were ordered onto the bus, officers and senior NCOs standing to the side and glaring at us lest anyone think of not boarding. For the most part, this seemed to work. Everyone near the front of the lines boarded the buses, though once someone walked past the buses, a few more people down the line followed them. I got on the bus. Looking out the window, I saw that the proportion of people choosing to stay was increasing towards the end of the line, but still most people got on the buses. After the rank and file boarded, the officers and NCOs followed. Most of them anyway; many stayed behind, not to give up their arms, but because they had civilian family in the Sutter pocket and would take a later bus with them (a higher fraction of officers and NCOs than rank-and-file soldiers were married, so in particular, more of them had family in the pocket, though still most didn't). Though I did see one sergeant in another company approach a bus, hesitate, and then run past it, to a visibly negative reaction from his company's officers.

The bus's engine started and the doors closed. “All right, let's go,” said the bus driver, “We'll be taking a slightly roundabout route so that we can stay within California-controlled territory until we reach the front line in Reno, but we will get you there in a few hours.” That sounded slightly suspicious, but the officers didn't seem worried, and there were a couple UN peacekeepers on the bus, so I was pretty sure we weren't getting kidnapped.

The bus pulled away, and we were on our way. It was a slow, very bumpy ride across the Sutter pocket and through Yuba City; the war had not been kind to highway 20. But once we left Yuba City and turned South, the ride was pretty smooth. Many of the other soldiers fell asleep. I wished I could do that, but I've never been able to sleep on the road, apparently not even in my current exhausted state.

Just past Placerville, the bus pulled over and stopped. “We're gonna have to stop for about 45 minutes to recharge the bus,” announced the bus driver, “Meanwhile, you can get out, stretch your legs, and have some lunch.” Most people had woken up, and I could tell by looking around at everyone's faces that I wasn't the only one to be surprised by that announcement. It wasn't surprising that these silly electric buses would have to regularly stop to recharge for an extended period of time, but no one had mentioned that they'd feed us, and I wouldn't've expected the Californians to give enemy soldiers free food if the deal didn't require them to. I don't think even the officers on board had seen this coming.

The doors opened, and the food aroma was overpowering. Whatever it was, it smelled delicious. I was ravenous despite having had a larger than usual breakfast about three hours previously, and I think everyone else was too. Some soldiers at the front didn't feel the need to wait for the officers to confirm to us that it was okay to leave the bus, and once they got up, the rest of us followed.

We were parked near what appeared to be an impromptu outdoor kitchen staffed by beautiful young women. We approached them, and were each handed generous servings of food, and immediately started scarfing them down. I stood in silence while I ate, next to Jones and Johnson, who were making small chat while they ate, though I wasn't listening to what they were saying. When I was about three-quarters of my way through the meal and starting to slow down noticeably, a lady with a dazzling smile approached the three of us.

“Hey, I'm Trisha,” she introduced herself, extending her hand. We each shook her hand and introduced ourselves. “How y'all doing?” she asked. There was a bit of a pause as we all processed how to answer that.

I came up with an answer first. “Relieved, but also exhausted,” I said.

“Yup, that,” echoed Johnson, and Jones nodded.

“We've got some cots nearby if you want to lie down for a bit,” she said.

“Sure, that'd be great,” I said. She glanced at Jones and Johnson, but they both declined, and Jones mumbled something about not wanting to miss the bus.

Trisha gestured for me to follow her, and I did. “Do you want to be woken up before the bus leaves?” she asked. I thought about it for a while without answering.

“If I say no, is it the same deal we were offered back in Sutter?” I asked.

“Mhm,” she said.

“What about all my stuff that's still on the bus?”

“We'll get it for you.”

I didn't say anything after that, and she didn't press for a real answer until we reached the door of a building that I gathered was where the promised cots were, and she gave me a questioning glance.

“I'm not getting back on that bus,” I said.

Trisha smiled, said “welcome to the Californian Peoples' State,” and left. I entered the building, fell onto a cot, and fell asleep almost instantly.

I awoke, feeling only somewhat refreshed, but desperately needing to pee. There were several other Free States soldiers on other cots now. I got up, I found a bathroom, and relieved myself. I saw there were showers, with a sign saying “10 MINS MAX” by them. I used to take showers twice that long all the time, but now being able to take a shower for 10 whole minutes sounded like unbelievable luxury. I wasn't sure whether the showers were for me or not, but I decided to just go for it instead of trying to find someone to ask. A clock started counting down from 10 minutes when I turned the water on. I used up almost the whole 10 minutes, and when I got dressed again, it struck me how much my uniform stank. I'd already known we'd been filthy, but I guess I'd adjusted to it and it was only apparent again now that it contrasted with my clean body.

I left the building and immediately ran into a Californian official who asked my name, told me to fill out some paperwork (which fortunately wasn't too long), took a picture of me, and printed out and handed me an ID hard.

“Your duffle bag's right over there,” he said, pointing to my belongings (sans gun) in a pile of luggage, “You can take it now or leave it and come back for it whenever. You can stay in this building again tonight. Tomorrow morning, some of those buses'll be heading back to Yuba City, and others'll be going to Sacramento and the bay. There will be job and housing fairs in all those locations, and also one in Placerville in case you decide to just stay here. We've got some pamphlets here summarizing what the available options will be in case that helps you decide where you want to go. And if you want to go somewhere else in the Peoples' State, let me know, and it is likely we will be able to help you out. Any questions?”

“Not right now. Thanks,” I said. I took a pamphlet, folded it up and put it in my pocket without reading it, and walked back to the bus charging station. The outdoor kitchen was still in operation, or perhaps in operation again. But the people there were different. Most of them were older than the women who had been there when I'd first arrived, and there were also some children present. Their genders were much more balanced, though still majority female. There were also some recently former Free States soldiers like me hanging around.

A woman waved at me as I approached, and it took me a second to realize it was Trisha; apparently not all the women who had been here earlier were gone. She was dressed much more conservatively than she had been earlier.

“This is how you dress when you're not trying to manipulate enemy soldiers?” I guessed.

“Um, not really. There's actually another bus coming soon with the families from your company,” she said.

“Oh, so you're dressing to manipulate a different demographic of enemy soldier.”

“Yes, exactly.”

“Hm, I didn't exactly sign up to stand here so my fellow soldiers on their way to Reno could stare at me on their way by and judge me for abandoning them.”

“Well, you better scram quick, then. The bus'll be here any moment now. Some of your comrades who joined us are hanging out over that hill,” she said, pointing.

I thought about it. “Actually, you know what, my buddy Kyle and his wife Ashley would probably be on that bus, and as awful as telling them I defected to their faces sounds, letting them find out later in Reno without the chance to say goodbye sounds worse. I'll stick around.”

“Alright.”

“I bet you don't eat like this every day,” I said, gesturing at the kitchen.

“Not quite,” she said, “Though there haven't really been food shortages, so we've been eating pretty well. The main limitation is that anything that takes a lot of water to grow is a bit expensive, since we can't grow it in California, and the war hasn't been great for trade. If you're wondering about shortages causing problems, the main thing is water. There's enough to drink, of course, but the water rationing is tight enough that I usually don't get to shower and wash my clothes as much as I'd like. We got a water bonus for participating in these greeting parties, so everyone around here is a bit cleaner than usual. Though we always get to shower and do laundry more than it looks like you guys have been, no offense, so I'm not sure if we really needed to bother cleaning up more than usual. Anyway, I usually shower for about 5 minutes every 3 days and use up my water ration.”

“So that 9-minute show I just took...?” I asked.

“Was part of your defection bonus. You didn’t just use up all your water for the week; don’t worry. I’m just warning you about what things will be like once you settle in. Though not for too long, hopefully. We expect the water crisis to end this year, so it's really only a short term problem.”

If they expected the water crisis to end this year, that meant that they expected to capture sources of water from Free States of America. Holding sources of water and denying them to California had been a deliberate strategy by the Free States of America to try to weaken California, which hadn't yet been terribly effective. We were all starting to figure California would recapture most of those water sources from us instead of collapsing, but if the Californians thought that was going to happen this year, then they were feeling even more optimistic than we thought they should.

A green bus pulled up. Trisha excused herself and ran off to take care of something. I watched people file out of the bus, looking for Kyle and Ashley. Instead, I saw Sergeant Tucker, Lieutenant Dan, their wives, and Sergeant Tucker's five-year-old daughter get off the bus and walk vaguely in my direction. I looked away and pretended I didn't see them. They kept walking closer.

“Private Carlson!” said Lieutenant Dan, “What the hell are y'all doing still in Placerville?” Dammit!

“Goddammit, Carlson! You defected, didn't you?!” said Sergeant Tucker, with a stern look. I nodded sheepishly. This was even more uncomfortable than telling that to Kyle and Ashley would have been. Fortunately, the tension was interrupted by a Californian boy, maybe about eleven or so, carrying boxes of food and handing them out to us. I took one even though I wasn't sure if it was only intended for the newcomers.

The kid stuck around and introduced himself as Ben, and we took turns introducing ourselves to him in between mouthfuls of food.

“They're making you do this?” Linda Tucker asked Ben.

“No, my teacher told us about it as a volunteer activity. No one had to be here,” said Ben.

“Wait, the schools are still running here?” asked Mrs. Tucker.

“Of course. Summer break doesn't start until May,” said Ben. Sergeant and Mrs. Tucker exchanged glances.

“Have things changed around here in the last few years, you know, with the war and all?” asked Mrs. Tucker.

“Yeah, my Dad's away on the front line near Redding. I haven't seen him in almost a year.” Ben looked sad.

“Uh, I guess so. Mom and my siblings and I are doing fine,” said Ben.

Sergeant and Mrs. Tucker exchanged some more glances. No one said anything, but it looked like they were having a whole private conversation with their eyes.

Sergeant Tucker looked away and made some awkward eye contact with Lieutenant Dan just as Mrs. Tucker said, “We're not going to Reno.”

Lieutenant Dan looked exasperated. “Linda, you goddamn hippy! You know everything's gonna be fine when we get back to the Free States, right?”

“We're not going to Reno,” Mrs. Tucker repeated. Sergeant Tucker nodded. Lieutenant Dan rolled his eyes and let out a disgusted grunt.

While this was going on, I overheard a conversation behind me between a Californian and a Free States soldier's wife who had apparently overheard Mrs. Tucker's exchange with Ben, and asked, “Is school mandatory? I wouldn't want to send my child to a public school.”

“No Ma'am, lots of people homeschool their kids,” said the Californian.

“But people who homeschool their kids still have to pay taxes for other people to go to public schools, right?”

“Well yes, they do pay taxes, Ma'am. Although, actually a fairly small amount of that has been going to schools lately.” The lady seemed reassured by this somehow, even though it was really just a diplomatic way of saying that funds had been diverted from schools to the war effort. I wasn't a big fan of taxpayer-funded public education myself, but I had trouble imagining why anyone would think that was any better.

I heard Ashley's voice, “Hey, there's Cole!” I turned my head and saw Kyle and Ashley running towards me. I left the gaggle I was in, ran towards them, and hugged both.

“So, you defected?” asked Kyle, sounding surprisingly not that disappointed.

“Yeah.”

“We were thinking of doing the same, honestly,” said Ashley, “Just don't tell the brass over there.”

“Actually the Tuckers are also defecting,” I said.

“You're shitting me!” said Kyle.

“No, they're really doing it.”

“I guess that would explain the tension you can see between the Tuckers and the Dans right now,” said Ashley.

Word about the Tuckers defecting went around pretty fast, and seemed to start a sort of domino effect. When the bus continued on its way to Reno, there were maybe a handful of people still on it. Not including the Dans, amazingly enough.

Soon after, another bus pulled up from the other direction. Private Jones and a few other soldiers disembarked. I saw Jones look through the crowd until he found Private Johnson, who was turned away and hadn't seen. Jones ran towards him and called out to him. They hugged each other. Kyle, Ashley, and I wandered over to see what was going on.

“I thought you were going to Reno,” said Johnson.

“The bus pulled over in South Lake Tahoe next to another green bus, and they told us that since they were both only half full, they'd be consolidating into just one bus for the rest of the trip. After I got out of the bus to board the other one, I asked if it was too late to change my mind, and ended up getting back on the same bus, turning around, and heading back here,” said Jones.

“Wait, I thought they said the bus would be staying in California-controlled territory all the way until Reno. Don't we control South Lake Tahoe?” asked Johnson. “We” might not have been the best pronoun to refer to the Free States by, now that we'd all defected to California, but no one pointed this out.

“Not anymore, we don't,” said Jones, “There were Californian soldiers all over the place. No Free States soldiers or signs of fighting to be seen.”

“Jesus. I wonder why they went that far before consolidating into fewer buses. They could have done that here in Placerville,” Johnson pointed out.

“I don't know, but my guess is they just wanted to flex on us by parking us in front of a California garrison in South Lake Tahoe,” said Jones, “One other thing. Get this: Just before we pulled in here, we passed another station just like this one. I saw a green bus pulling away from it in the other direction towards Reno. Then I looked at the people still milling around there on the ground, and there was Mrs. Fitch with her kids.”

“The Fitches defected?!” Johnson asked incredulously.

Jones shook his head, “Not Colonel Fitch. I only saw Mrs. Fitch and their kids.”

“Yeah, but you might have just not seen him. I mean, if his wife and kids were there,” said Johnson.

“Yeah, sure. I can't prove beyond all doubt that Colonel Fitch was on the bus. But if you seriously think that crazy son-of-a-bitch would stay with his wife and kids instead the Free States of America, then I've got a bridge to sell you.”

## Exact 2-cycles are degenerate isomorphisms

The situation in which you have vector spaces $V$ and $W$, and linear maps $f:V\rightarrow W$ and $g:W\rightarrow V$ such that $\ker\left(f\right)=\text{im}\left(g\right)$ and $\ker\left(g\right)=\text{im}\left(f\right)$ often arises in the situation in which you would have an isomorphism between $V$ and $W$ if you knew how to divide by $0$. Specifically, this happens when you'd need to divide by $0$ exactly once; in similar situations in which you'd need to know how to divide by $0$ multiple times in order to get an isomorphism, you get $f:V\rightarrow W$ and $g:W\rightarrow V$ such that $f\circ g=0$ and $g\circ f=0$ but whose kernels and images are not necessarily equal.

I'll call such a pair $\left(f,g\right)$ with $\ker\left(f\right)=\text{im}\left(g\right)$ and $\ker\left(g\right)=\text{im}\left(f\right)$ an exact 2-cycle of vector spaces. Note that the two vector spaces $V$ and $W$ in an exact 2-cycle are in fact isomorphic, as $\dim\left(V\right)=\dim\left(\ker\left(f\right)\right)+\dim\left(\text{im}\left(f\right)\right)=\dim\left(\text{im}\left(g\right)\right)+\dim\left(\ker\left(g\right)\right)=\dim\left(W\right)$.

Given a finite-dimensional vector space $V$ and an invertible linear map $f:V\rightarrow V$, its adjugate is almost its inverse; you just have to divide by $\det\left(f\right)$. If $f:V\rightarrow V$ is not invertible, then of course, $\det\left(f\right)=0$, so dividing by $\det\left(f\right)$ doesn't work. But if $f$ has nullity $1$, then $\ker\left(f\right)=\text{im}\left(\text{adj}\left(f\right)\right)$ and $\ker\left(\text{adj}\left(f\right)\right)=\text{im}\left(f\right)$. That is, $\left(f,\text{adj}\left(f\right)\right)$ is an exact 2-cycle. If $f$ has nullity $N\left(f\right)\geq2$, then $\det\left(f\right)=0^{N\left(f\right)}$, and hence inverting $f$ requires dividing by $0$ more than once, and $\text{adj}\left(f\right)=0$.

### Homogeneous polynomials and multilinear forms

Given a vector space $V$ over a field $k$, let $k\left[V\right]_{2}$ denote the space of quadratic forms on $V$ (that is, homogeneous quadratic polynomial maps $V\rightarrow k$), and let $Sym^{2}V^{*}$ denote the space of symmetric bilinear forms on $V$.

Given a symmetric bilinear form $B$ on $V$, we can construct a quadratic form $Q$ on $V$ by $Q\left(\vec{x}\right):=B\left(\vec{x},\vec{x}\right)$. This gives us a map $f:Sym^{2}V^{*}\rightarrow k\left[V\right]_{2}$ by $f\left(B\right)\left(\vec{x}\right)=B\left(\vec{x},\vec{x}\right)$.

$2B\left(\vec{x},\vec{y}\right)=B\left(\vec{x}+\vec{y},\vec{x}+\vec{y}\right)-B\left(\vec{x},\vec{x}\right)-B\left(\vec{y},\vec{y}\right)$, so we can recover $B$ from $f\left(B\right)$ by $B\left(\vec{x},\vec{y}\right)=\frac{1}{2}\left(f\left(B\right)\left(\vec{x}+\vec{y}\right)-f\left(B\right)\left(\vec{x}\right)-f\left(B\right)\left(\vec{y}\right)\right)$. That is, the map $g:k\left[V\right]_{2}\rightarrow Sym^{2}V^{*}$ given by $g\left(Q\right)\left(\vec{x},\vec{y}\right)=Q\left(\vec{x}+\vec{y}\right)-Q\left(\vec{x}\right)-Q\left(\vec{y}\right)$ is twice the inverse of $f$.

This doesn't quite work if $\text{char}\left(k\right)=2$, since we can't do the part where we divide by $2$. In fact, $f$ is not invertible in this case. But $g$ is still a well-defined map $k\left[V\right]_{2}\rightarrow Sym^{2}V^{*}$, and it's still true that $g\circ f=2 id_{Sym^{2}V^{*}}$ and $f\circ g=2 id_{k\left[V\right]_{2}}$; it's just that now that means $g\circ f=0$ and $f\circ g=0$. In fact, $\ker\left(f\right)=\text{im}\left(g\right)$ and $\ker\left(g\right)=\text{im}\left(f\right)$. $\ker\left(g\right)$ and $\text{im}\left(f\right)$ are the $\dim\left(V\right)$-dimensional space of diagonal quadratic forms (polynomials that are linear combinations of squares of linear functions $V\rightarrow k$), and $\ker\left(f\right)$ and $\text{im}\left(g\right)$ are the ${\dim\left(V\right) \choose 2}$-dimensional space of alternating symmetric bilinear forms. Thus $Sym^{d}V^{*}$ and $k\left[V\right]_{2}$ are both $\dim\left(V\right)+{\dim\left(V\right) \choose 2}$-dimensional.

Similar things happen with higher degree homogeneous polynomials and symmetric multilinear forms. Let $k\left[V\right]_{d}$ be the space of homogeneous degree-$d$ polynomials on $V$ and $Sym^{d}V^{*}$ the space of symmetric $d$-linear forms on $V$. We have functions $f:Sym^{d}V^{*}\rightarrow k\left[V\right]_{d}$ given by $f\left(\Phi\right)\left(\vec{x}\right):=\Phi\left(\vec{x},...,\vec{x}\right)$ and $g:k\left[V\right]_{d}\rightarrow Sym^{d}V^{*}$ given by $g\left(\phi\right)\left(\vec{x}^{1},...,\vec{x}^{d}\right)=\sum_{I\subseteq\left[d\right]}\left(-1\right)^{d-\left|I\right|}\phi\left(\sum_{i\in I}\vec{x}^{i}\right)$. $g\left(f\left(\Phi\right)\right)=d!\Phi$ and $f\left(g\left(\phi\right)\right)=d!\phi$, so if $\text{char}\left(k\right)=0$ or $\text{char}\left(k\right)>d$, then $f$ and $g$ are bijections, and $d!$ times each others' inverse. Otherwise, $g\circ f=0$ and $f\circ g=0$. If $\frac{d}{2}<\text{char}\left(k\right)\leq d$, then $\text{char}\left(k\right)$ divides $d!$ with multiplicity $1$, and $\ker\left(f\right)=\text{im}\left(g\right)$ and $\ker\left(g\right)=\text{im}\left(f\right)$. If $1<\text{char}\left(k\right)\leq\frac{d}{2}$, then $\text{char}\left(k\right)$ divides $d!$ with multiplicity $\geq2$, and all bets are off. Though $\dim\left(k\left[V\right]_{d}\right)=\dim\left(Sym^{d}V^{*}\right)={\dim\left(V\right)+d-1 \choose d}$, no matter what $\text{char}\left(k\right)$ is.

### Newtonian spacetime

In special relativity, we work with a 4-dimensional (3 for space and 1 for time) real vector space $T$, with a symmetric bilinear form $\left\langle \cdot,\cdot\right\rangle$, called the Minkowski inner product, of signature $\left(-+++\right)$; that is, the associated quadratic form can be given, in coordinates, by $-t^{2}+x^{2}+y^{2}+z^{2}$ ($t$ is the time coordinate and $x,y,z$ are spatial coordinates for some reference frame). If $\left\langle \vec{v},\vec{v}\right\rangle >0$, then $\vec{v}$ is spacelike, and $\sqrt{\left\langle \vec{v},\vec{v}\right\rangle }$ measures its distance (in the reference frame in which its temporal coordinate is $0$). If $\left\langle \vec{v},\vec{v}\right\rangle <0$, then $\vec{v}$ is timelike, and $\sqrt{-\left\langle \vec{v},\vec{v}\right\rangle }$ measures its duration (in the reference frame in which it is at rest). By currying, the Minkowski inner product can be seen as a linear map $T\rightarrow T^{*}$, where $T^{*}$ is the vector space of linear maps $T\rightarrow\mathbb{R}$. Since the Minkowski inner product is nondegenerate, this linear map $T\rightarrow T^{*}$ is an isomorphism.

In Newtonian physics, things are a little different. We can still work in 4-dimensional spacetime, but we don't have a single Minkowski inner product measuring both distance and duration. We do have a global notion of time; that is, there's a linear map $t:T\rightarrow\mathbb{R}$ that tells you what time it is at each point in spacetime. $\ker\left(t\right)$ is space in the present moment, so it should be Euclidean space; that is, it should be equipped with an ordinary inner product.

The time function $t$ induces a degenerate inner product on $T$ by $\left\langle \vec{v},\vec{w}\right\rangle :=t\left(\vec{v}\right)t\left(\vec{w}\right)$. As before, this can be seen as a linear map $T\rightarrow T^{*}$ (it sends $\vec{v}$ to $t\left(\vec{v}\right)t$), with 1-dimensional image $\text{span}\left(t\right)$ and 3-dimensional kernel $\ker\left(t\right)$.

The ordinary inner product on $\ker\left(t\right)$ gives us a degenerate inner product on $T^{*}$: since our inner product on $\ker\left(t\right)$ is non-degenerate, it induces an isomorphism between $\ker\left(t\right)$ and its dual, and hence induces an inner product on $\ker\left(t\right)^{*}$. There's a canonical map $T^{*}\rightarrow\ker\left(t\right)^{*}$ given by restriction: $\varphi\mapsto\varphi\restriction_{\ker\left(t\right)}$. So given $\varphi,\psi\in T^{*}$, we can define their inner product to be the spatial inner product of their restrictions to $\ker\left(t\right)$. This can be seen as a linear map $T^{*}\rightarrow T$ (given $\varphi:T\rightarrow\mathbb{R}$, restrict it to $\ker\left(t\right)$, and then find the element of $\ker\left(t\right)\subseteq T$ that corresponds to it via the spatial inner product) with image $\ker\left(t\right)$ and kernel $\text{span}\left(t\right)$. We have thus found canonical maps $T\rightarrow T^{*}$ and $T^{*}\rightarrow T$ such that the kernel of each is the image of the other.

### Why?

In the spacetime example, it is conventional in special relativity to normalize the speed of light to $1$. But another thing we can do is let the speed of light be the variable $c$. So $\left\langle \left[\begin{array}{c} t_{1}\\ x_{1}\\ y_{1}\\ z_{1} \end{array}\right],\left[\begin{array}{c} t_{2}\\ x_{2}\\ y_{2}\\ z_{2} \end{array}\right]\right\rangle =-c^{2}t_{1}t_{2}+x_{1}x_{2}+y_{1}y_{2}+z_{1}z_{1}$. As a map $T\rightarrow T^{*}$, this is $\left[\begin{array}{c} t\\ x\\ y\\ z \end{array}\right]\mapsto\left[\begin{array}{cccc} -c^{2}t & x & y & z\end{array}\right]$. The inverse map $T^{*}\rightarrow T$ is $\left[\begin{array}{cccc} \tau & \alpha & \beta & \gamma\end{array}\right]\mapsto\left[\begin{array}{c} -c^{-2}\tau\\ \alpha\\ \beta\\ \gamma \end{array}\right]$, or, as an inner product on $T^{*}$, $\left\langle \left[\begin{array}{cccc} \tau_{1} & \alpha_{1} & \beta_{1} & \gamma_{1}\end{array}\right],\left[\begin{array}{cccc} \tau_{2} & \alpha_{2} & \beta_{2} & \gamma_{2}\end{array}\right]\right\rangle =-c^{-2}\tau_{1}\tau_{2}+\alpha_{1}\alpha_{2}+\beta_{1}\beta_{2}+\gamma_{1}\gamma_{2}$. We're going to want to take a limit as $c\rightarrow\infty$ and get something finite, so we'll have to scale our inner product on $T$ down by a factor of $c^{2}$, giving us $\left\langle \left[\begin{array}{c} t_{1}\\ x_{1}\\ y_{1}\\ z_{1} \end{array}\right],\left[\begin{array}{c} t_{2}\\ x_{2}\\ y_{2}\\ z_{2} \end{array}\right]\right\rangle =-t_{1}t_{2}+c^{-2}x_{1}x_{2}+c^{-2}y_{1}y_{2}+c^{-2}z_{1}z_{1}$, or, as a map $T\rightarrow T^{*}$, $\left[\begin{array}{c} t\\ x\\ y\\ z \end{array}\right]\mapsto\left[\begin{array}{cccc} -t & c^{-2}x & c^{-2}y & c^{-2}z\end{array}\right]$. The limit $c\rightarrow\infty$ gives us our temporal inner product on Newtonian spacetime, $\left\langle \left[\begin{array}{c} t_{1}\\ x_{1}\\ y_{1}\\ z_{1} \end{array}\right],\left[\begin{array}{c} t_{2}\\ x_{2}\\ y_{2}\\ z_{2} \end{array}\right]\right\rangle =-t_{1}t_{2}$, and our spatial inner product on the dual space $\left\langle \left[\begin{array}{cccc} \tau_{1} & \alpha_{1} & \beta_{1} & \gamma_{1}\end{array}\right],\left[\begin{array}{cccc} \tau_{2} & \alpha_{2} & \beta_{2} & \gamma_{2}\end{array}\right]\right\rangle =\alpha_{1}\alpha_{2}+\beta_{1}\beta_{2}+\gamma_{1}\gamma_{2}$, giving us our exact 2-cycle of maps between $T$ and $T^{*}$, $\left[\begin{array}{c} t\\ x\\ y\\ z \end{array}\right]\mapsto\left[\begin{array}{cccc} -t & 0 & 0 & 0\end{array}\right]$ and $\left[\begin{array}{cccc} \tau & \alpha & \beta & \gamma\end{array}\right]\mapsto\left[\begin{array}{c} 0\\ \alpha\\ \beta\\ \gamma \end{array}\right]$. (I did say that this should only work if we have to divide by $0$ once, not if we must do so twice, and this involved $c^{2}$, but we never used $c$ on its own anywhere, so we can just say $c=\sqrt{\infty}$, and it's fine).

Let's go back to the first example. Given $f:V\rightarrow V$ of nullity $1$, perturb $f$ slightly to make it invertible by adding an infinitesimal $\varepsilon$ times some map $g:V\rightarrow V$. The only condition we need $g$ to satisfy is $g\left(\ker\left(f\right)\right)\nsubseteq\text{im}\left(f\right)$. That way $\det\left(f+\varepsilon g\right)$, which must be a multiple of $\varepsilon$, is not a multiple of $\varepsilon^{2}$. $\left(f+\varepsilon g\right)\circ\text{adj}\left(f+\varepsilon g\right)=\text{adj}\left(f+\varepsilon g\right)\circ\left(f+\varepsilon g\right)=\det\left(f+\varepsilon g\right)id_{V}$. Clearly $f\circ\text{adj}\left(f\right)=\text{adj}\left(f\right)\circ f=\det\left(f\right)id_{V}=0$. Given $\vec{x}\in\ker\left(f\right)$, $\left(f+\varepsilon g\right)\left(\vec{x}\right)=\varepsilon g\left(\vec{x}\right)$, so $\text{adj}\left(f+\varepsilon g\right)\left(\varepsilon g\left(\vec{x}\right)\right)=\det\left(f+\varepsilon g\right)\vec{x}$. Hence $\text{adj}\left(f+\varepsilon g\right)\left(\frac{\varepsilon}{\det\left(f+\varepsilon g\right)}g\left(\vec{x}\right)\right)=\vec{x}$. Since $\det\left(f+\varepsilon g\right)$ has $0$ constant term but nonzero coefficient of $\varepsilon$, $\frac{\varepsilon}{\det\left(f+\varepsilon g\right)}$ can be evaluated at $\varepsilon=0$, and has a nonzero, finite value. Then $\text{adj}\left(f\right)\left(\frac{\varepsilon}{\det\left(f+\varepsilon g\right)}\vert_{\varepsilon=0}g\left(\vec{x}\right)\right)=\vec{x}$. So $\left(f,\text{adj}\left(f\right)\right)$ forms an exact 2-cycle for reasons closely relating to the fact that perturbing each of them infinitesimally can make them inverses up to an infinitesimal scalar multiple.

Now, in the second example, where $V$ is a vector space over a field $k$ of positive characteristic, $\frac{d}{2}<\text{char}\left(k\right)\leq d$, and we have an exact $2$-cycle $f:Sym^{d}V^{*}\rightarrow k\left[V\right]_{d}$, $g:k\left[V\right]_{d}\rightarrow Sym^{d}V^{*}$, let ${\cal O}$ be an integral domain of characteristic $0$ with a unique maximal ideal $\mathfrak{m}$, such that ${\cal O}/\mathfrak{m}=k$ and $\text{char}\left(k\right)\notin\mathfrak{m}^2$ (for instance, if $k=\mathbb{F}_{p}$, we can use ${\cal O}=\mathbb{Z}_{p}$ and $\mathfrak{m}=p\mathbb{Z}_{p}$). Lift $V$ to a free ${\cal O}$-module $\tilde{V}$ with $\tilde{V}\otimes k=V$ (in coordinates, this means, instead of $V=k^{n}$, work with $\tilde{V}={\cal O}^{n}$, which carries a natural map to $k^{n}$ by reducing each coordinate mod $\mathfrak{m}$). Then there are natural maps $\tilde{f}:Sym^{d}\tilde{V}^{*}\rightarrow{\cal O}\left[\tilde{V}\right]_{d}$ and $\tilde{g}:{\cal O}\left[\tilde{V}\right]_{d}\rightarrow Sym^{d}\tilde{V}^{*}$ such that $\tilde{g}\circ\tilde{f}=d!id_{Sym^{d}\tilde{V}^{*}}$ and $\tilde{f}\circ\tilde{g}=d!id_{{\cal O}\left[\tilde{V}\right]_{d}}$, and $\tilde{f}$ and $\tilde{g}$ reduce mod $\mathfrak{m}$ to $f$ and $g$, respectively. Where $K$ is the field of fractions of ${\cal O}$ (so $K=\mathbb{Q}_p$ in our example with $k=\mathbb{F}_p$ and $\cal{O}=\mathbb{Z}_p$), $\tilde{f}\otimes K:Sym^{d}\left(\tilde{V}\otimes K\right)^{*}\rightarrow K\left[\tilde{V}\otimes K\right]_{d}$ and $\tilde{g}\otimes K:K\left[\tilde{V}\otimes K\right]_{d}\rightarrow Sym^{d}\left(\tilde{V}\otimes K\right)^{*}$ are bijections (in coordinates, $\tilde{V}\otimes K=K^{n}$, and tensoring a map with $K$ just means the same map extended over the field of fractions), as they are inverses of each other up to a multiple of $d!$, which is invertible in $K$. Since $\text{char}\left(k\right)\leq d$, $d!\in\mathfrak{m}$, and $g\circ f=0$ and $f\circ g=0$. Given $\phi\in\ker\left(g\right)$, if we lift $\phi$ to $\tilde{\phi}\in{\cal O}\left[\tilde{V}\right]_{d}$, $\tilde{g}\left(\tilde{\phi}\right)\in\mathfrak{m}Sym^{d}\tilde{V}^{*}$. Since $\text{char}\left(k\right)>\frac{d}{2}$, $d!\notin\mathfrak{m}^{2}$, and hence $\frac{\tilde{g}\left(\tilde{\phi}\right)}{d!}\in Sym^{d}\tilde{V}^{*}$, and of course, $\tilde{f}\left(\frac{\tilde{g}\left(\tilde{\phi}\right)}{d!}\right)=\tilde{\phi}$. Reducing mod $\mathfrak{m}$, we get $f\left(\frac{\tilde{g}\left(\tilde{\phi}\right)}{d!}\mod\mathfrak{m}\right)=\phi$. Thus $\ker\left(g\right)=\text{im}\left(f\right)$. Similarly, given $\Phi\in\ker\left(f\right)$, lift $\Phi$ to $\tilde{\Phi}\in Sym^{d}\tilde{V}^{*}$. $\tilde{f}\left(\tilde{\Phi}\right)\in\mathfrak{m}{\cal O}\left[\tilde{V}\right]_{d}$. $\frac{\tilde{f}\left(\tilde{\Phi}\right)}{d!}\in{\cal O}\left[\tilde{V}\right]_{d}$, and $\tilde{g}\left(\frac{\tilde{f}\left(\tilde{\Phi}\right)}{d!}\right)=\tilde{\Phi}$. Reducing mod $\mathfrak{m}$, we get $g\left(\frac{\tilde{f}\left(\tilde{\Phi}\right)}{d!}\mod\mathfrak{m}\right)=\Phi$. Thus $\ker\left(f\right)=\text{im}\left(g\right)$. So $\left(f,g\right)$ forms an exact cycle because, in $K$, they are inverses up to a factor of $d!$, which we can divide by, and which is $0$ with multiplicity $1$ in $k$, since $d!\in\mathfrak{m}\setminus\mathfrak{m}^{2}$.

### The general story

All three arguments from the previous section took the following form: Let ${\cal O}$ be a discrete valuation ring with residue field $k$, field of fractions $K$, and valuation $\nu:K^{\times}\rightarrow\mathbb{Z}$. Let $V$ and $W$ be free ${\cal O}$-modules, and let $f:V\rightarrow W$ and $g:W\rightarrow V$ be such that $f\circ g=\varepsilon id_{W}$ and $g\circ f=\varepsilon id_{V}$, for some $\varepsilon\in{\cal O}$ with $\nu\left(\varepsilon\right)=1$. Then $f\otimes K:V\otimes K\rightarrow W\otimes K$ and $g\otimes K:W\otimes K\rightarrow V\otimes K$ are isomorphisms, and each is $\varepsilon$ times the inverse of the other. $f\otimes k:V\otimes k\rightarrow W\otimes k$ and $g\otimes k:W\otimes k\rightarrow V\otimes k$ form an exact 2-cycle: they compose to $0$ because $f$ and $g$ compose to $\varepsilon id$, which goes to $0$ in $k$, and given $\vec{x}\in V\otimes k$ such that $\left(f\otimes k\right)\left(\vec{x}\right)=0$, we can lift $\vec{x}$ to $\tilde{x}\in V$. $f\left(\tilde{x}\right)\in\varepsilon W$, so $\varepsilon^{-1}f\left(\tilde{x}\right)\in W$, and $g\left(\varepsilon^{-1}f\left(\tilde{x}\right)\right)=\tilde{x}$, so tensoring with $k$ sends $\varepsilon^{-1}f\left(\tilde{x}\right)$ to some $\vec{y}\in W$ such that $\left(g\otimes k\right)\left(\vec{y}\right)=\vec{x}$. Thus $\ker\left(f\right)=\text{im}\left(g\right)$. The same argument with $f$ and $g$ switched shows $\ker\left(g\right)=\text{im}\left(f\right)$. The exact 2-cycle $\left(f\otimes k,g\otimes k\right)$ is a sort of shadow of the isomorphisms $f\otimes K,g\otimes K$.

In the spacetime example, $K=\mathbb{R}\left(\left(c^{-2}\right)\right)$, ${\cal O}=\mathbb{R}\left[\left[c^{-2}\right]\right]$, $k=\mathbb{R}$, and $\varepsilon=c^{-2}$. In the adjugates example, $K=k\left(\left(\varepsilon\right)\right)$, ${\cal O}=k\left[\left[\varepsilon\right]\right]$, and the $\varepsilon$ in the general story is $\det\left(f+\varepsilon g\right)$. In the homogeneous polynomials and symmetric multilinear forms example, $K$ is a discretely valued field of characteristic $0$ with residue field $k$, ${\cal O}$ is its valuation ring, and $\varepsilon=d!$.

All exact 2-cycles of vector spaces can be fit into this general story. Given any exact 2-cycle $f:V\rightarrow W$, $g:W\rightarrow V$ ($V$, $W$ vector spaces over $k$), we can take a discretely valued field $K$ with residue field $k$, and then lift $f,g$ to $\tilde{f},\tilde{g}$ with $\tilde{g}\circ\tilde{f}=\varepsilon id$ for some $\varepsilon\in K$ with $\nu\left(\varepsilon\right)=1$, exactly the conditions in the above argument.

### What more?

What about exact 2-cycles in abelian categories other than vector spaces? In general, the two objects in an exact 2-cycle need not be isomorphic. For instance, with abelian groups, there's an exact 2-cycle between the 4-element cyclic group and the Klein four-group. Though two objects in an exact 2-cycle must be isomorphic in any category in which every short exact sequence splits (this is the gist of the dimension-counting argument from the beginning showing that two vector spaces in an exact 2-cycle must be isomorphic). Is there still some way of seeing exact 2-cycles as degenerate isomorphisms even in contexts in which there need not be actual isomorphisms?

Also, what about exact $n$-cycles? That is, a cycle of $n$ functions such that the image of each is the kernel of the next. If an exact 2-cycle is a degenerate form of an isomorphism, and an isomorphism is an exact sequence of length 2, then perhaps an exact 3-cycle should be a degenerate form of an exact sequence of length 3 (i.e. a short exact sequence). This is hard to picture, as a short exact sequence is not symmetric between its objects. However, for reasons not understood by me, algebraic topologists care about exact 3-cycles in which two of the three objects involved are the same (these are called exact couples), and this apparently has something to do with short exact sequences in which the first two objects are isomorphic, which provides some support for the idea that exact 3-cycles should have something to do with short exact sequences. An exact sequence of length 1 just consists of the $0$ object, so this suggests an exact 1-cycle (i.e. an endomorphism of an object whose kernel and image are the same) should be considered a degenerate form of the $0$ object, which is also hard to picture.

## It's a big world out there

“Mom,” asked Kimi, “could there actually be a light that bright?”

“I don't know,” she answered, “but according to ancient myth, there is such a thing. Or was, at least. I suppose there's no way of knowing whether it's still around. It's in a far away world with no rock ceiling too, so goes the myth. Sweet dreams.” Their mother left.

“No rock ceiling,” Kimi whispered, “That's even wilder than the thing about the light. Like, would it just be air all the way up forever? Surely there'd have to be an end somewhere, right?”

“Maybe there is a rock ceiling there, but it's so high up that you can't hear the echo,” Jerilyn suggested.

“Wow, that would be so disorienting, not being able to hear the echo off the rock ceiling,” said Kimi.

“Given what Mom said about the light, maybe you could see the rock ceiling even if you couldn't echolocate it,” said Jerilyn.

“But I guess if people thought there wasn't a rock ceiling at all, it must be high enough that you can't see it either,” said Kimi.

“I guess so,” Jerilyn agreed.

“Jerilyn,” said Kimi.

“What?”

“Do you think it's real?”

“No,” said Jerilyn.

“Are you sure?”

Jerilyn hesitated. “No,” she said. She really had no way of knowing for sure, however outlandish it may sound.

“Jerilyn.”

“What?”

“I'm not actually feeling all that tired. Are you?”

“Eh, somewhat, but not especially.”

“Let's go find the place Mom was talking about.”

Jerilyn thought about it. On the one hand, the mythical place probably didn't exist, and even if it did, there was no way they were going to find it. On the other hand, an adventure might be fun. “Let's do it,” she said.

They snuck off and made their way to their canoe. They avoided making sounds so as not to advertise their presence, so they had to rely on touch to find their way, but they knew the route well enough that that wasn't a huge impediment.

They set off, and as they knew the waters immediately surrounding the dock by heart, they were able to navigate away from the island silently, but once they were a ways out, Kimi started making clicking noises with her tongue so they could echolocate their surroundings. They aimed straight for the closest point where the rock ceiling met the water. They couldn't echolocate that far, of course, but Jerilyn remembered the way from her navigation lessons. Once they got too far from the archipelago, they had to rely on trying to keep going in a straight line, but soon after, they encountered the wall.

“I suppose we look around for a tunnel,” said Jerilyn.

They turned right and followed where the rock ceiling met the water, keeping it on their left, their casual conversation sufficing to provide enough noise for them to track their surroundings. They never found a tunnel. Eventually they got tired, pointed their canoe back in the direction they came from, and set off for home. When they first encountered an island, they weren't sure which one it was, and they went all the way around it in a circle so they could estimate its size and shape. It seemed unfamiliar, but Jerilyn thought back to her navigation lessons, and by the time they had completed their circle around the island, she came up with a guess as to which island it was. If she was right, they were significantly off course. She turned the canoe in the direction she thought home was, and when they passed the next island, she gained confidence that she was right, and indeed, their new path took them straight home, where they docked the canoe, dried themselves off, went straight to bed, and each fell asleep instantly.

*

Kimi and Jerilyn made several more expeditions to find tunnels to new worlds, taking off in different directions each time. On their fourth trip, they found an indentation well into the rock, which tapered out into a vein of air sticking just above the surface of the water. They got as far in as they could, until the rock ceiling got too low for them to stay under while they were sitting in the canoe. They stashed their paddles in the canoe, and carefully got out and swam farther in while towing the canoe. They soon reached a point where the canoe couldn’t go any farther even without them inside. They found a part of the rock ceiling that jutted down below the rest, and they pulled the end of their canoe downwards, pushed it under the jutting rock, and released it, so that the jutting rock extended into the canoe and would keep it from floating away.

They swam in further. But soon even the indentation they found sunk below the surface of the water. They each took a big breath of air, and kept swimming farther out underwater. They hadn’t brought sonar rods or a light, and couldn’t snap underwater, so they had no way of echolocating underwater, and had to rely on touching the rock ceiling above them to tell where it was. They didn't get very far before Jerilyn decided that that wasn't a great idea. She turned back, and pushed Kimi to turn back as well. Even with Jerilyn's caution, they were both somewhat short of breath by the time they could get their noses back into the air.

On their next trip, they brought a pair of sonar rods, and aimed for the same indentation they had found on their previous trip. When they arrived at where the rock ceiling met the water, they were in unfamiliar territory. On their previous trip, they had been keeping the line where the rock ceiling met the water to their left as they’d followed it until finding the indentation, and this time, they’d tried going a bit to the right of the course they’d taken on the previous trip in an attempt to go more directly to the indentation, so they figured that they’d overcorrected, and turned left. They soon found the indentation again.

Again, they went as far as they could while keeping their heads above water, Kimi carrying the sonar rods. Then they dove down into the water, much deeper than necessary just to stay below the rock, so that they would be able to echolocate as far as possible without the nearby part of the rock getting in the way, and Kimi rang the sonar rods.

The rock ceiling’s descent flattened out not long after the last of it passed below the surface of the water, and there was a small air pocket just a bit after the rock ceiling flattened out. About twice as far in as the air pocket, the rock ceiling started to pitch up sharply.

They swam up back towards the surface for air, Kimi ringing the sonar rods between armstrokes so they could keep track of where the air was instead of bumping into the rock ceiling.

“Let’s check out that air pocket,” Kimi suggested, after they surfaced.

“Not a good idea,” said Jerilyn.

“We can totally make it there,” said Kimi.

“Air pockets sometimes have bad air in them. We could get there, of course, but I’m not so sure we could make it back after coughing out nasty air,” Jerilyn explained.

Kimi reluctantly agreed not to explore the air pocket, and they turned back.

*

On their next trip, they brought buckets. They figured if they weren’t sure they’d have enough air in their lungs for the trip to the air pocket and back, they could bring some more air outside their lungs.

When they’d gotten as far as they could while keeping their heads above water, they quickly discovered that it was just about impossible to swim underwater while carrying a bucket full of air. After a long while trying, they figured out how to get themselves positioned upside-down in the water with their feet against the rock ceiling while holding a bucket full of air pulling them up against the rock ceiling, so they could walk along it. Both of them still had trouble carrying a bucket and a pair of sonar rods underwater at the same time, so they’d put their sonar rods back in the canoe. But they were able to make enough sound to echolocate their immediate surroundings by hitting the sides of their buckets.

They both needed a breath by the time they got to the air pocket, as walking upside-down underwater was much slower than swimming. So they found flat portions of the rock ceiling to put their buckets down on, then turned around, exhaled, stuck their heads in their respective buckets, and took a breath. Then they exited the buckets, and Kimi approached the air pocket. She stuck her hand in, and made contact with the rock almost instantly; it was, evidently, a very shallow pocket. She stuck her nose in, being careful not to rise high enough to hit the rock ceiling, and, heeding Jerilyn’s warning, cautiously took a small breath of air. It was rancid. She coughed it up and recoiled out of the pocket, then scrambled for her bucket while fighting the urge to inhale. She finally got her head in the bucket, took deep breaths and kept coughing, while Jerilyn held her up so she could focus on regaining her breath instead of swimming.

By the time Kimi got her breathing under control, the air in the bucket was quite stale and she was short of breath again, so she left the bucket for her big sister to deal with while she swam back to fresh air. Jerilyn took another breath from her own bucket, dumped the remaining air out of the buckets, and swam back while carrying them, which took her a lot longer than it took Kimi because of the drag caused by the buckets.

“You were right. That was nasty,” Kimi commented, once Jerilyn surfaced.

They decided to make another trip underwater to try to explore past the ridge where the rock ceiling pitched back up again. They retrieved their sonar rods and tied them to Kimi’s wrist to make them easier to carry at the same time as the buckets, and set off in the same direction as before.

They set their buckets down near the air pocket, each took a breath, and then swam out to the ridge. Another ring of the sonar rods revealed that the rock ceiling pitched straight up into a vertical cliff, and that there was a wide expanse of air about thirty feet above them.

They retreated to their buckets, each took a breath from them, and then dumped the remaining air out and swam back with their buckets.

“I don’t understand how the water went so high up. The surface is definitely much higher on the other side than it is here,” said Jerilyn, after they surfaced again.

“Yeah, weird, isn’t it? Also, how are we supposed to get there? We canoed around the edge for miles in each direction and didn’t find any tunnels or places where it bends around or anything that could lead to that place,” said Kimi.

“There probably isn’t any route there going over the surface. If there was, it would be even harder to understand why the water level is different there than here,” said Jerilyn.

“A completely separate world! Do you think it’s the place Mom told us about?”

“I don’t know.”

*

They left for home, and on their next trip, they brought four buckets, with the intention of going all the way to the surface on the other side of the rock. Then they repeated the previous expedition’s trick of walking upside-down underwater with buckets of air, this time each carrying a bucket in each hand, which was even harder to get into position for, but eventually they figured it out. This made it not only difficult to ring the sonar rods, but also difficult to hit the buckets, and they resorted to periodically letting their buckets hit the rock ceiling to make enough noise to navigate.

They stopped briefly near the air pocket to turn rightside-up and take a breath from their buckets, and then turned back upside-down and kept going, buckets still in hand, all the way until the point where the ridge pitched back up again. They set down the buckets in stable locations, turned rightside-up, exhaled, took deep breaths from their full buckets, and swam up towards the surface, leaving four half-filled buckets on the underside of the ridge behind them, Kimi periodically ringing the sonar rods on their way up so they wouldn’t collide with the rock.

They surfaced and each began to take a deep breath, then stopped in shock, and cautiously started to breath again. Something was off about the air. It smelled... not stale, exactly, but strange, not like any air they'd ever smelled before. It smelled overly fresh, in a way, as if all the air they’d beathed until that point had been a bit stale, and they hadn’t noticed.

Jerilyn raised a hand out of the water, shook some water off of it, and snapped. For the briefest of instants, they both thought that perhaps there wasn't a rock ceiling above them at all. But then they heard the echo, and realized that there was a rock ceiling above them at perhaps three times the height that they were accustomed to at home. And they couldn’t see any bright lights in the sky, or anything at all for that matter, so they couldn’t be in the place Mom had described in the myth. Aside from the rock on one side of them and bending into a ceiling far above them, there was nothing around them, just water for as far as they could hear.

“We gotta get the canoe in here so we can explore this place,” said Kimi.

“How in the world are we going to do that?” asked Jerilyn, realizing as she spoke that perhaps it should have been “how out of the world” rather than how in it.

“I don’t know,” said Kimi.

They swam around a bit, but didn’t find anything interesting, and decided to go home. They dove down under the ridge, retrieved their buckets and inhaled from them, surfaced on the other side, got in their canoe, and headed home.

*

Later, they did some experimenting at home, and discovered that their canoe was almost exactly the same density as water. Armed with this fortuitus fact, several buckets, and a lot of rope, they set off again for the other world.

A test run revealed that their rope wasn’t quite long enough to stretch from where they could park their canoe to the air on the other side of the rock. Finding this out resulted in Jerilyn dropping the rope on her way up after crossing the ridge so she could surface and breath, and then returning to their canoe, and they reeled the rope back in.

They set up three buckets of fresh air on the underside of the ridge, and one by the air pocket. Then Jerilyn took the sonar rods and swam out to the ridge and treaded water with her head in a bucket while Kimi filled the canoe with water, and pushed it underwater and forward, Jerilyn ringing the sonar rods in the water to help Kimi tell what she was doing as she swam under, and periodically ducking down into the water to keep herself updated on Kimi’s progress. Kimi wasn’t getting very good resolution from the sonar rods, but it helped that she remembered the path. Kimi, pushing the canoe ahead of her, reached the bucket by the air pocket and took a breath in it. Jerilyn took one last big breath from a fresh bucket and took off for the surface as Kimi continued forward pushing the canoe. When Jerilyn surfaced, she was able to help by reeling in the canoe, holding onto the rock cliff for leverage. Kimi went ahead of the canoe so she wouldn’t run out of air, and together they finished reeling in the canoe to the surface. With some difficulty, they emptied the water out of the canoe, righted it, and got back in.

Righting the canoe had been a lot of work, and they took a quick break to catch their breath. Then they set off in their canoe, keeping their old world to their left.

They heard sounds of civilization coming to them before they echolocated the island from their own snapping. They turned towards it and approached. They were noticed, and it seemed that they had caused a fair amount of consternation.

They got close, and a man was standing on the end of a peninsula near them holding a long, straight stick, facing them and snapping repeatedly. There were also boulders sticking above the water a ways to either side of them.

“Hello,” said Kimi, “Who are you? I’m Kimi.”

The man did not respond, but he did stop snapping and started clicking his tongue. The tongue-clicking wasn’t giving them good resolution on him, but they could tell he was moving in some way. Jerilyn snapped, revealing that the man had both hands on the stick, which was pointed at them, and he was leaning back as if about to throw it. Jerilyn dug her paddle into the water and swung them around, just as the man threw the stick. It narrowly missed Kimi.

“Hey, what was that for?” Kimi shouted.

“Kimi, paddle forward hard!” said Jerilyn, as she began to do so herself. They heard splashing sounds to either side of them, followed by the sounds of people swimming towards them. The man on the shore began clicking his tongue again, and seemed to be preparing for another throw. Jerilyn swung the canoe around again, and the stick just missed her. She resumed paddling forward, and the man on the shore dove into the water.

Someone grabbed the back of the canoe near Jerilyn and pulled himself up towards her. She moved her paddle between them just in time to block a thrust of a stick towards her. He grabbed her paddle with the hand that hand been on the canoe. Kimi lunged at him and hit him in the neck with her paddle with a surprising amount of force for someone her size. He dropped Jerilyn’s paddle as well as his own stick and fell back into the water. The recoil from Kimi’s lunge caused their canoe to collide with someone else as he pulled up towards the position Kimi had just left. Jerilyn hit him over the head with the edge of her paddle, and he too lost his grip on the boat. Jerilyn pushed him away from the boat with her paddle while he was too disoriented to grab it, and then Kimi and Jerilyn returned to their former positions and kept paddling hard. No one caught up to them, and they relaxed a bit once their pursuers had given up.

It took a while before they encountered the next sign of civilization. They approached much more cautiously this time, coming to rest at shouting distance. A small gaggle of people were gathered at the shore closest to them.

“Hello!” one of them shouted. They sounded funny.

“Hello!” Kimi shouted back.

“What is that thing?” asked the person on the shore. Their words were tricky to understand.

“The thing you’re sitting on floating in the water,” the stranger clarified.

“The what?”

“This is called a ‘canoe’,” said Kimi, slapping the side of the canoe.

“Okay, so, what is it?”

“You use it to cross the water,” said Kimi. She wasn’t sure what else to say about the concept of canoes.

“We’re exploring,” said Kimi. “The last people we encountered weren’t very nice,” she added.

“Uh, were they the <unrecognizable word>?”

“Did you encounter them over there?” asked the stranger, gesturing in the direction Kimi and Jerilyn had just come from, and snapping to give them good enough resolution to tell where he was pointing.

“Yes,” said Kimi.

“What the <unrecognizable> were you doing over there?”

“Uh, we didn’t know not to go there.”

“Uh, well now you know. Good thing you survived. Where are you from?” asked the stranger.

“Elsewhere,” said Kimi, knowing the name of their island wouldn’t mean anything to them.

“Uh-huh. Hey, do you guys need any supplies, like food or anything? We’d be happy to help out if you show us how the canoe works,” said the stranger.

“That would be gr-” Kimi started.

“Kimi, no,” Jerilyn interrupted, “They want to steal our canoe.”

They were both getting hungry, but they’d have trouble getting back home without their canoe. It wasn’t worth the risk. They kept going. They were not pursued.

*

It was a long time before they found land again. When they did find land, it wasn’t an island separated from the old world by water like the others had been, but instead, the rock wall separating them from the old world flattened out to become navigable by foot. They were ravenous, having serious regrets about having ventured so far without food, and on the verge of turning back. So they were quite gratified when they smelled vents. They pulled their canoe onto the shore, located the vents, and gorged themselves on ventmoss. Their hunger sated, they noticed they were getting quite tired, and they went to sleep.

When they awoke, they decided to explore the new land they’d found. They walked inland for quite some time without finding another shore; they’d never imagined a land so vast before. Eventually they became tired again, gave up on finding water on the other side, and turned back. They lost track of the exact route they had taken, and when they reached the shore again, it wasn’t familiar territory. A gust of wind carried a faint smell of vents towards them, and, guessing that it was from the same vents they had found earlier, they followed the shore in the direction the wind had come from. This guess turned out to be correct, and they found their canoe right where they’d left it. They ate some more ventmoss, drank from the water, and rested for a while.

Then they decided to venture uphill, in the direction of the old world; perhaps they would be able to walk on top of the rock ceiling of the old world. The ground gradually steepened, and they kept going long past the point where they had to crawl on all fours, and each step brought them more up than forward. At times, they had to rely on their voices for echolocating footholds when their hands were occupied clinging to the rock and they couldn’t snap. When they turned back, it was due to some combination of the steepness spooking them, and them getting quite tired. They downclimbed facing backwards until the ground had flattened out enough that they could walk upright without falling over, and then they walked their way back to their canoe and the vents, had another meal, and went to sleep.

When they woke up again, they decided to return home. They followed the route they had taken last time, but steered clear of any signs of civilization. When they neared the place where they’d met the people who’d attacked them, they stayed very close to the rock wall separating them from the old world, paddled slowly, and instead of snapping, frequently gently tapped the rock next to them for guidance, in hopes of minimizing noise and not advertising their presence.

When they reached approximately the place where they had first surfaced into the new world, they had some trouble figuring out exactly the right place. Kimi periodically dove into the water with the sonar rods, and in most places, it was easy to tell that they couldn’t be in the right place because the rock extended down too far vertically into the water. But eventually they found a point below them where the rock didn’t extend as far down, and theorized that that might be their route home.

Jerilyn dove below the edge of the rock, rang the sonar rods, and sure enough, there were their buckets of air on the underside of the ridge. She went back to the surface, and they tied their rope to their canoe, filled the canoe with water, and pushed it under while Jerilyn held the rope. They surfaced again after pushing the canoe down a ways underwater, took deep breaths, dove all the way under the ridge until they got their heads in air buckets, and pulled the canoe further down by reeling in the rope until the canoe was below the ridge. Then they dumped the air out of their buckets and carried buckets and rope back to the other side, with a quick stop at the bucket they’d placed midway to take breaths and pack up that bucket too. They were desperate for air by the time they finally surfaced on the other side.

After they finished panting for breath, they reeled in their canoe, and laboriously emptied the water out of it, righted it, and went home. Their parents were delighted to see them, cross at the prolonged absence, and skeptical of their tales of the new worlds they’d discovered.

*

Some time later, Kimi and Jerilyn decided to make another expedition to the new world and try to climb further up the steep cliffs they’d found. Realizing that it would take a long time, and they’d want water and food other than ventmoss, they packed some dried fish and plenty of buckets, and fashioned some seals for their buckets so that water could be stored in them without spilling when jostled around.

They set out along the same path as in their previous expeditions, although it took them some time to find again the indentation in the rock where they’d crossed over into the new world. Once they did, they repeated their usual procedure to get to the other side, after tying their extra buckets (two containing dried fish sealed inside) to the canoe, since carrying the extra buckets underwater themselves would have been too unwieldy.

Once they reached the air on the other side, reeled in their canoe, righted it, and emptied the water out, they took a break to catch their breath. They then continued roughly in the same direction as their previous journey, with a detour to steer clear of signs of civilization before they rejoined their original route, which they successfully stayed on from then on, making for a lengthy but uneventful trip to the place they had landed at on their previous trip.

Unlike on their previous trip, they were not hungry when they reached the land, as they had been snacking on dried fish the whole time. But they were quite tired, so they went to sleep before going any further.

*

When they awoke, they filled the remaining space in their buckets of dried fish with ventmoss, filled two other buckets with water, and took off uphill, each carrying a food bucket over one shoulder and a water bucket over the other. They kept going past where they had turned back the previous time, and not long after, had to backtrack a bit because the route seemed too precariously steep. But after a little exploring, they were able to find a more navigable route up.

After a long ascent and many quick breaks, they decided they needed some sleep. Unfortunately, they were on very steep ground. However, after a bit of exploring, they managed to find a crevice of flat ground big enough for both of them to lie down in, and they went to sleep.

They continued their ascent when they awoke. At one point, Jerilyn, who was in the lead, slipped and fell on a steep stretch. Fortunately, she did not hit Kimi on the way down, and was not far above some flatter terrain on which she managed to stop her fall. Miraculously, the seals on both of her buckets had held.

Kimi downclimbed to join Jerilyn, and asked if she was alright. Jerilyn reported that while she would probably develop some bruises from the fall, she was otherwise undamaged. They looked around for a safer way up, eventually found one, and continued on.

The hill eventually flattened out considerably, and they were able to consistently walk upright without their hands on the ground, though still uphill. The rock ceiling got progressively lower, to the point where it wasn’t far above their heads. In places, they even had to duck under it, though there were also places where the rock ceiling was much higher. At one such point where they rock ceiling was anomalously high, they saw a few small points of light above them at an angle, and in that particular direction, the rock ceiling was further away than they could echolocate, if it was there at all.

Eventually they grew tired, and went to sleep again. When they awoke, they noticed that the ground quite a ways behind them was glowing brightly. The air in a line connecting the rock ceiling to that patch of ground was also glowing faintly. They walked towards it, but the glowing patch narrowed and disappeared before they reached it. They turned back uphill and pressed on.

Later, they saw another glowing patch of ground, again with accompanying faintly glowing ray of air shooting up to the rock ceiling, well to their right. They headed towards it, but it too narrowed and disappeared before they reached it, and they turned back uphill.

The rock ceiling narrowed further, and they had to crawl to keep going. On multiple occasions, the rock ceiling come so close to the ground that they could not go further, or even merged with the ground, becoming a wall in front of them, and in such cases, they had to backtrack and find a different route up. At one point, the only route forward was so narrow between ground and ceiling that, in order to get through them, they had to take their buckets off their shoulders and push them ahead, and advance while lying flat. Kimi, being smaller, had an easier time of this, and at one point, Jerilyn got stuck, but Kimi was able to turn around in a slightly wider spot just ahead and give Jerilyn a hand, helping her get through.

The ceiling rose further above them again, eventually to the point where they could walk upright without ducking. They saw a patch of little points of light ahead of them, and they went in that direction, which required a steep climb. As they drew close, it became apparent that the points of light were coming from a hole in the rock wall, as echoes bounced off rock to every side of the patch of points of light, but not from the patch itself.

They passed through the hole. Though the ground continued to stretch out before them in all directions, there was no longer any wall to either side or in front of them, nor a ceiling above them, as far as they could tell from the echoes of their snaps. There was an almost-vertical wall behind them surrounding the hole, but rather than bending above them into a ceiling as it rose, it bent back in the other direction, as if to form high ground after flattening out further beyond their hearing range. There were many little points of light in every direction above them. There was one big source of bright light, almost a disk, but with one side blunted slightly inwards. There was faint light pervading through the air, so that they could see things in their immediate vicinity, including each other, clearly, despite the bright lights being far above them, and they could even see geological features much farther away than they could hear.

“We found it!” said Kimi, “The place from the legend! Look, there’s the bright light Mom told us about!” She pointed to the big almost-disk of bright light above them.

“Yeah,” said Jerilyn, “It doesn’t hurt to look at, though. And Mom didn’t mention all the other lights. Still, considering it was an ancient myth, it did turn out to be remarkably accurate. That sure is a lot of light.”

They explored the new wide open land, snapping as they went to echolocate the ground, even though they could see it just fine, since they were not accustomed to using light to find their footing. They quickly discovered that it was far larger even than it had first appeared. For instance, they set off in the direction of what appeared to be a patch of vegetation low to the ground, which they could see but not echolocate, but the vegetation seemed to grow larger but draw further away as they approached, not coming within echolocating range until well after they expected it to.

On their way, they heard a burbling sound, and investigating, they found a trail of fast-moving liquid flowing across the ground. Kimi tapped the surface of the liquid hesitantly, then cupped her hands, plunged them under the surface, and brought some of the liquid back up in her hands. It felt like water. She sipped it. It tasted like water. She reported her findings, and Jerilyn followed suit, and concurred. They had never come across such a wide stretch of such fast-flowing, shallow water before. It was a fortunate find, as they had been running low on water, and would have had a hard time on their way back if they hadn’t found more water. The water was fresher than the water in their buckets, so they refilled their buckets with it.

By the time they finally reached the patch of vegetation they’d been headed towards, it became apparent that the vegetation, which they had initially thought to be low to the ground, was actually enormous, with thick stalks extending far over their heads, high enough to extend well past the rock ceiling from home, and branching out, with vegetation covering the branches far above them.

A similar phenomenon occurred when they headed for some small hills in the distance. Again, the hills seems to draw further away as they approached. But unlike the vegetation, the hills did not also seem to grow as they approached. They pursued the hills longer than they had pursued the vegetation, but the hills still seemed no closer, and their size hadn’t changed. They speculated that perhaps the hills were simply illusions, or perhaps they were vastly further away than the vegetation had been. They were tired. They found a good spot to lie down, and went to sleep.

*

Something was wrong. Kimi opened her eyes and screamed, waking Jerilyn, who also screamed. There was light everywhere. So much light, as if an anglerfish’s lure was right in front of their eyeballs, except that it was coming from all directions.

They quickly identified the source of the light: an inconceivably bright light coming from the ground in the distance, which, true to the legend, it hurt to look at. They turned away from the light, held each others’ hands, and took deep breaths to calm themselves down while they got used to the incredible quantity of light all around them. Jerilyn speculated that, since this light was so bright it hurt to look at, and was located far away on the ground, and the light they’d seen before was merely bright and located far up above them, that perhaps the light that their Mom had spoken of in the myth, which was supposed to be painfully bright and high up above them, was a conflation of the two lights that they’d seen.

Once they’d calmed down a bit, they kept exploring. The bright light slowly climbed above the ground and into the air, which Jerilyn noted meant she was probably wrong in her earlier speculations. At the same time, the light kept gradually getting even brighter, to the point where it hurt to look in any direction at all, and, counterintuitively, it actually got harder to see as the intensity of light increased. The novelty of so much light flooding their surroundings wore off quickly, so they ended up spending a lot of time with their eyes closed, but their ability to see farther than they could echolocate was useful for navigating, so sometimes they would squint or partially cover their eyes instead.

Eventually, Kimi noticed that her skin hurt. She remarked on this, and Jerilyn noticed that her skin hurt as well. There was no obvious cause to their ailments. Jerilyn speculated that, since they’d been fine before the light got so bright, and the skin under their clothing didn’t hurt, that perhaps the light was hurting their skin. They decided to try getting out of the light.

They found some more of the tall vegetation, which was dense enough to block much of the light from coming under it. They took a break under it. It was generally more pleasant there, as it was cooler (it had been warm earlier), and the reduced level of light didn’t hurt their eyes as much.

Their skin kept getting worse, though. This gave them some doubt over whether it was the light that was hurting their skin, but it still seemed possible that it was because of the light, and their skin was continuing to hurt because of damage already done. And they didn’t have any better ideas than staying there; the hole in the rock that they had emerged from was far away, and they didn’t feel like making their way back to it in all the light, in case it was the light that was hurting their skin. They had no guarantee that the light would go away, but since it had been much dimmer earlier, that gave them some hope that it would dim again.

Kimi began to cry out of some combination of fear and the pain of her skin. Jerilyn tried to comfort her, though her skin also hurt, and she was also concerned. They waited there a long time without the light going away. They were exhausted, as they had been woken up by the light well before they would have woken up on their own, but they also couldn’t get to sleep because of the light, stress, and pain in their skin. Their skins were growing blisters, and they were losing hope that the light would go away any time soon, so they were considering making their way back to the hole, when they noticed that the source of the light was slowly making its way back towards the ground. They decided to wait for it to get there to see what would happen.

The light slowly dimmed as the bright light drew close to the ground, and Kimi and Jerilyn took off for the hole they’d emerged from. They’d gotten used to the way that their surroundings would seem to grow and draw away as they approached, and they were able to use landmarks they recognized by sight to navigate back to the hole. They refilled their water buckets again when they reached the fast-flowing vein of water. There was plenty of vegetation and wildlife around them, and they speculated that some of it might be edible, and they had gone through well over half their food, so it was tempting to attempt to restock on food for the return trip, but they didn’t know how to determine what was edible, as they didn’t recognize any of it. Jerilyn was concerned that, since light seemed to be toxic to their skin in high doses, perhaps consuming vegetation that had been exposed to that much light might also be toxic to them (she realized later that this could also be an issue with the water that they’d found, but it wasn’t like they could just not drink water, so it was a risk they’d have to take). They had to make do with the food that they’d already packed.

The bright light in the sky was long gone, and, following its departure, the ambient light continued to dim. By the time they reached the cliff, the ambient light had returned to the level it had been at when they’d emerged, and they could see the little points of light, and the one big light in the sky that had so impressed them when they’d first seen it, but no longer seemed so grand, in comparison to the much brighter light that had replaced it for a time.

They found the hole that they’d emerged from, walked in, and retreated inwards quite a ways from the hole before they collapsed on the ground, exhausted, and slept.

*

Their skins were still painful and sensitive when they awoke, and the hole they’d traveled through was glowing intensely.

They continued on their way back home, but when they got to approximately the point where they thought the narrow spot they’d crawled up through was, it took them a long time before they found it. Crawling back through it was quite painful, as it was impossible to climb through without scraping their sensitive skin. But after some painful struggle, they made it through.

Their progress down was much slower than their progress up had been, both because of their skin sensitivity slowing down their crawling, and because it was difficult to retrace their steps. Recognizing this, they rationed their food and water so that it would last long enough. During the phase of their journey where they had to crawl under a low ceiling, they seemed hopelessly lost for a long time before they finally made their way to an area where they had enough room to stand up, and in that more open area, they were eventually able to find what seemed to be their previous path. Satisfied that they were no longer lost, they went to sleep before continuing.

They had an easier time following their route up from then on. Despite their skin pain and weariness slowing them down, they actually exceeded their pace from the way up on the flatter portions of the trip, but they lost that extra time on the portions where they had to climb. Despite their efforts to conserve food, they ran out by the next time they stopped to sleep.

After descending further for quite a long time, they ran out of water, but they realized that they were getting close to the water, the vents, and their canoe. They desperately needed more sleep, but they needed food and water more, so they pressed on. They were quite relieved when they finally reached the bottom. They drank from the water, sated their hunger with ventmoss, and went to sleep. When they woke up, they got in their canoe and set off for home.

## The Knot

George and I were afraid we might be late for our meeting with the wizard, and not wishing to keep them waiting for us, we rushed there. Just before we arrived, I checked my phone and saw that we were two minutes early. I apprehensively prepared myself to knock on the door, but it swung open before I did so, revealing an impressively cluttered office. There was no one inside. George and I looked at each other.

“Do you suppose we should go in?” George asked.

“I don't think the door would have opened if we weren't supposed to,” I said. After some hesitation, I stepped inside, and George followed. The door closed behind us, causing us both to reflexively turn back towards the door. I tried the door handle, and found that the door offered no resistance to being opened again. Reassured that we weren't trapped, I closed the door again, to return it to the condition that I presumed the wizard preferred it in.

I noticed a loop of thin red rope hanging over the doorknob. There were no ends tied together in a knot. I picked it up to look for where the ends had been fused together, but could not find any joint; the rope appeared to have been constructed in a perfectly homogeneous circle.

I placed the loop of rope back around the doorknob, and turned my attention to the other objects filling the room. There was a perfectly spherical orb sitting on the wizard's desk. The orb had a cloudy appearance, and the clouds drifted aimlessly on the surface of the orb, despite the orb otherwise appearing to be solid. There was a shelf on a wall, holding an old oscilloscope, a set of the five platonic solids, each made out of smooth black material, and a beaker that held a liquid which was dancing around violently, but never spilling out of the beaker despite appearing to come close very frequently. There was a fireplace in the corner with a fire, but the only material in the fireplace was a bird sitting in the middle of the fire, but the bird wasn't burning, and it looked like it was sleeping. The bird raised its head to look at us quizzically, and then went back to sleep. I heard a faint popping sound, which I soon figured out had come from the liquid in the beaker. There was a bookshelf, completely packed with books, covering an entire wall, and there were also a few open books and many loose sheets of paper covering the wizard's desk, as well as a few sheets of paper that had fallen to the floor. Very few of the papers I saw were in English, and most weren't in any script I recognized. Some didn't appear to have any writing at all on them, consisting only of cryptic diagrams.

I noticed a strand of rope sticking out from under some papers on the wizard's desk. It appeared to be made out of very similar material to the loop of rope on the doorknob, except that it was green instead of red. I carefully moved the papers that were on top of it out of the way so I could see the rest of the rope. Like the rope hanging on the doorknob, it formed a closed loop. There were three points where the rope crossed over another part of the rope. The crossings alternated, in the sense that if you started at any crossing, and followed the strand on top around the loop, it would lead to the bottom strand of the next crossing it encounters, and then the top strand of the third crossing, and then the strand going under the point where you started, and so on. Only part of the rope was the green I had initially seen, another stretch of rope was red, matching the loop of rope on the doorknob, and part of it was blue. The rope was arranged so that the three points where the colors changed were hidden under the crossings.
I moved the portion of red rope that crossed over the boundary between green and blue, so that I'd be able to see the point where the color of the rope changed from green to blue. To my surprise, the piece of rope that I had just uncovered was solid blue all the way up to the new point that the red strand crossed over it. George asked me how it had done that, but I didn't know, and I ignored the question. I wiggled the red strand some more, but the portion of the rope it was moving over kept changing between blue and green so that the color switch always occurred exactly under the red strand. I tried holding the red strand in place and pulling the green strand under it, but again blue rope turned green just as it emerged out from under the crossing. I lifted the red strand into the air, and moved my head around to look under it from both directions. The color of the lower strand shifted in unison with my head, so that I never caught a glimpse of the boundary between the colors. I wiggled the strands going over the other two crossings to see if they would exhibit the same phenomenon, and they did. I paused for a moment to stare at the rope in confusion, and then picked up a piece of green rope and moved it over the blue portion of the rope, forming two additional crossings. Blue rope turned red as the green strand passed over it, forming an additional stretch of red rope in the middle of the blue part of the rope, again with the color change happening precisely under the crossings. Next I tried moving the green strand over the point where the blue strand crossed over the boundary between red and green. As I had anticipated, the stretch of rope going over the crossing turned from blue to red as the green strand passed over it, and an additional short stretch of blue rope had formed out of the red rope coming out from under the crossing, with all color boundaries being hidden behind other stretches of rope. I returned the loop of rope to its original configuration, and then tried twisting part of the blue portion of the rope, so that it crossed over itself. This did not cause any color changes, and I undid the twist.

“Hey George, I want to try something. Can you go around to the other side of the desk for a minute?” I said.

“Are you sure the wizard will be okay with us messing with his stuff like this?” George asked.

“I'm sure it'll be fine. Come on,” I said, pushing George in the intended direction. I actually had no idea whether or not the wizard would mind, but my curiosity had won out over my fear of offending the wizard. George walked around to the other side of the desk as I had requested.

“Okay, now look closely at this crossing,” I instructed, pointing to where the green stretch of rope passed over the boundary between the red and blue strands, which we were looking at from opposite sides. I crouched so that I was looking at the knot from a shallower angle, and George followed my example. I lifted the green strand going over the crossing up in the air. I was looking at the crossing from the side that the red strand was coming out from, and the blue stretch of rope coming out the other side appeared to turn red as the green rope passed in front of it in my field of vision.

“What's it look like?” I asked.

“The rope under the green strand is now blue up until the point where it crosses behind the strand,” he said. I put my finger on the red rope directly under the green part I had lifted.

“So this looks blue?” I asked.

“Yeah,” he said.

“So you can see my finger touching a blue stretch of rope?” I asked.

“Yeah, that's what I said,” George confirmed. I stood up and bent over to look at the rope from above, and pressed the green strand I was holding into my face running vertically between my eyes, so that I could see the piece of rope crossing under it from opposite sides of the green strand with each eye. It was a purple blur that could have been the result of red light reflecting off the rope into my right eye and blue light reflecting off the rope into my left eye. I unfocused my eyes so that the stretch of rope I was looking at would appear in different places in my field of vision in each eye, and indeed, it appeared as separate red and blue strands.

Suddenly remembering the loop of rope on the doorknob, I dropped the rope I was holding and went to go get it. George walked back around the desk to the side facing the door. I returned with the red loop of rope and held it over the rope on the table. The green and blue portions of the rope that I could see through the red loop had switched colors, while the red portion of the rope on the table was not changed in appearance by viewing it through the red loop. I lifted part of the rope on the table, and slid the loop of red rope under it. The loop was no longer red all the way around, with color changes whenever it passed under a strand of rope of a different color. I grabbed the formerly red loop of rope by a blue stretch in the center of the loop of rope on the table, and pulled it out. I was holding a solid blue loop of rope. I put the blue loop of rope aside, took out my phone from my pocket, and opened the camera. I lifted the green strand and put my phone under it to take a picture of the spot where the rope crossing under it switched from red to blue. The camera image on the screen showed the strand changing from red to blue right under the spot where the green strand crossed over the phone, so that the boundary between red and blue wasn't visible on the screen. I took a picture, and then moved the rope out of the way so that I could see the spot where the color changed. But the picture I saw on the phone screen was of a completely red strand of rope. I moved the phone back under the green strand, and saw that the still image of a strand of rope in my camera was changing from red to blue as I moved the green strand over it. I pulled the phone back out the other side of the green strand, and it bore an image of a completely blue strand of rope. I closed the picture so I could take another one. The image of the knot in the phone screen looked the same as the actual knot, except that the colors red and blue were switched. I put down the phone, and pulled a pen and small notebook out of my pocket. I tore off a page of the notebook, and wrote on it the current color of the loop of rope I had taken from the doorknob (blue). I folded up the piece of paper, slipped it under the multicolored loop of rope with the crossings, and pulled it out through the center. I unfolded it, and found the word “green” written on it, in my handwriting, instead of the “blue” that I had written. I picked up my phone and called a friend.

She picked up, and before she said anything, I said, “Hi Kelly. Pick a color. Red, green, or blue?”

“Blue. Why do you ask?” she said.

“I'll explain later. Thanks. Bye,” I said, and hung up. I wrote “blue” under the word “green” on the piece of paper, folded it back up, and slipped it under the knot and pulled it out through the center, as I had done before. I unfolded it, and saw that the word “green” that had been near the top of the paper had turned into “red”, while the word “blue” that I had written when Kelly picked it had remained unchanged. I also noticed that the pen I was using had blue ink, and the color of the ink on the page had never changed. There were a couple more things I wanted to try. I thought through what I was going to do, and then called Kelly back.

“Can you pick a color again? Same options,” I asked.

“Red,” said Kelly.

“Thanks,” I said, and hung up. I lifted part of the knot into the air and stuck my right hand under it, so that my hand was sticking out through the center part of the knot. The plan was to hand the phone from my left hand to my right hand, and then pull it with my right hand back from under the knot, except that if Kelly had named the current color of the loop of rope that had been on the doorknob, I would only go through the motions of this without actually holding the phone. The loop of rope from the doorknob was blue, and Kelly had said red, so I kept the phone in my left hand as I moved my left hand towards my right, and I attempted to grasp the phone with my right hand. But while I saw my right hand grab the phone, I felt my fingers pass through thin air where I saw the phone. I withdrew my right hand out from under the knot, and while the phone was definitely pulled out of my left hand, and I saw my right hand holding the phone as it receded, I felt my right hand in a fist closed around nothing. As my hand passed out from under the knot, the fist became visible and the phone seemed to disappear. George gasped, as this was the first sign visible to him that anything was amiss.

“I don't know. In retrospect, I probably shouldn't have used my phone for that. At least we've still got your phone if we want to try taking more pictures,” I said. I felt rather foolish, as I had actually identified this outcome in advance as consistent with previous observations, but somehow hadn't seriously considered the possibility that it would actually happen.

“You just managed to lose your phone in the magic rope. I'm not letting you touch mine,” said George. He had a point. I thought about how I might get the phone back, but couldn't think of anything, and besides, there was another experiment I'd been going to try. I reached for the red strand of rope (chosen because it was the color that Kelly had picked), but before I touched it, it started receding under the green strand, as if the blue strand on the other side was being pulled, but the blue strand itself was motionless, and rather than turning blue as it came out from under the green strand, the red rope would simply vanish as it passed under the green strand, leaving a significantly shortened stretch of red rope by the time this stopped. The point where the red strand disappeared under the green was no longer aligned with the point that the blue strand came out from under the green on the other side. I grabbed the red rope near where it crossed under the blue strand and pulled. More red rope came out of nowhere so that the red strand still continued all the way up to where it disappeared under the blue strand, even as I pulled it away, just as if the green strand on the other side were passing under the blue strand and turning red, but the green strand itself did not move. The point where the red strand passed under the blue strand and vanished also became misaligned with the point where the green strand emerged out the other side. When I stopped pulling on the red strand, there was about the same amount of red rope visible as there had been before some of it had vanished under the green strand.

“Hello, folks. Sorry I'm late,” came a voice from behind us in a heavy accent that I didn't recognize. George and I turned around and saw someone of unidentifiable gender in robes and a pointy hat, carrying a wooden staff with a hexagonal piece of metal attached to the side and a shiny truncated octahedron fastened to the top, and wearing a ring on each of their ten fingers, each in a different style. The door was closed behind the wizard. I hadn't heard it open or close. The wizard's eye caught the knot of rope on their desk.

“Oy, the bloody thing's out of sync again,” they said, and walked over to the desk, put the staff down leaning against the desk, pulled a wand out of their robes, and jabbed their wand at the knot. They put their wand back in their robes, picked up the knot of rope, and threw it up in the air. When it landed back on the desk, the strands were perfectly aligned with each other again.

“There we go,” said the wizard. They picked up their staff and gestured with it towards a wall, out of which lept two folding chairs, which positioned themselves in front of the wizard's desk and unfolded into chairs that did not look the least bit like folding chairs.

“Have a seat,” said the wizard, indicating the chairs. I put my phone back in my pocket, and sat down.

## Why The Apple Falls

Like many children do, when my son Isaac was a little boy, he once asked me “How do people on the other side of the world stay up there? Wouldn't they fall down here?” So of course I explained that everything always falls toward the ground, even though that's the opposite direction on the other side of the world.

Isaac thought about this for a moment, and then asked, “What if you were halfway between here and the other side of the world? Which way would you fall?”

“As I said, you always fall towards the ground,” I told him, “So if you went East until you were halfway to the other side of the world, you would fall towards the ground there, which is the direction you'd call down once you were there, even though in a way, it's the direction we'd call East here. And if you went West, you'd still fall towards the ground there, which-”

“No!” he said, cutting me off, “What if you went that way,” he pointed straight up, “until you were halfway to the other side of the world?”

“You mean where the sun is?” I asked.

“Yes. What if you were holding onto the sun, and you lost your grip? Which way would you fall?”

There were so many things wrong with that question, I wasn't sure where to begin. “How would you get there?” I asked.

“I don't know, maybe you could use a ladder,” said Isaac.

“You can't balance a ladder going that high.”

“You could make a ladder long enough to go all the way to the other side of the world, prop both ends against the ground, and climb to the middle,” he said. It took me a while to figure out what he was getting at.

“Ok, but you can't even build a ladder that big.”

“But what if you did, Daddy? If you built the ladder and grabbed onto the sun and let go, which way would you fall?”

“You'd probably go blind from getting too close to the sun first,” I said.

“You could do it at night.” I had to admit he got me there. I was going to object that you probably still couldn't actually hold onto the sun, but I decided against it. I could see what I was supposed to say.

“Well, Isaac, I guess you'd fall towards the ground.”

“You mean away from the sun?” he asked.

“Yeah, away from the sun.” I was surprised he needed that clarification.

It's cute when little kids do it, but Isaac never gave up his habit of asking stupid questions. He's a great hunter, and he'll make a fine warrior too, but I can't say he's smart.

One day, when he was a teenager, he asked me, “Dad, why is it easier to throw a spear West than to throw it East?”

“Probably because you were throwing it on the Western face of a hill,” I said.

“I thought of that,” he said, “but it's like that everywhere. I even tried some target practice when we went to Brythsville, and that's almost on the other side of the world. East can't be uphill all the way around.”

“Maybe the wind was blowing West each time you tried?”

“No, the wind shouldn't affect a spear much. Besides, I've noticed the same thing in calm wind, and when the wind isn't blowing West.”

“Then you're probably imagining things. It can't be easier to throw a spear West than East on flat ground. That doesn't make sense,” I told him.

“I know it doesn't make sense, but I'm not imagining it. It's very consistent,” he insisted. We kept arguing about it for a while, and he kept rejecting all of my proposed explanations, but wouldn't let go of the idea that it was still easier to throw a spear West than East.

A few days later, as I was walking home through the village square, I head Isaac's voice shout “Dad!” from the top of the clock tower. I looked up, and saw him perched on top of the clock tower with his friend Emmy and a bucket. Emmy waved.

“How'd you get up there?” I asked.

“Watch closely,” Isaac said, ignoring my question, and he poured a bunch of pebbles out of the bucket.

“You didn't see them curve?” he asked.

“Curve? No.”

Isaac and Emmy climbed down the clock tower. “If you look closely, the pebbles curve a bit to the West as they fall,” Isaac said.

“It's probably just the wind,” I said, as Isaac and Emmy started picking up pebbles and putting them back in the bucket.

“The wind's pretty calm right now,” said Isaac. He was right. “Besides, exactly the same thing happened when we poured the pebbles inside the clock tower. Let me show you.” Isaac started climbing back up the clock tower with the bucket slung over his shoulder. Emmy led me inside the clock tower, and started explaining what was going to happen. The clock tower had no roof, so there was plenty of light. She pointed out that there were visible vertical lines on the walls formed by the edges of every other brick, and explained that since not all the pebbles were going to fall from the bucket at the same time, you could compare the positions of the highest pebbles to the positions of the lowest pebbles to see a line tracing out the path formed by the pebbles, and that it was going to curve slightly to the West, enough to be visible once the pebbles got near the ground.

Isaac reached the top, and started slowly pouring the pebbles out of the bucket. Sure enough, they followed exactly the path Emmy had said they would, curving just a tad to the West.

“Huh, you're right,” I said. Isaac started descending the tower.

“You see?” Isaac said, “There must be a small force pulling everything just a little to the West all the time, and it's usually too small to be noticeable unless something is in the air for long enough. That's why it's easier to throw a spear West than East-”

“Oh, not this again.”

“Because the spear is being pulled West. So in a way, it is kind of like East is uphill all the way around.”

“That's ridiculous!” I said, “A mysterious force pulling everything everything West?” I jokingly pretended I was being pulled involuntarily to the West, and screamed, “Aaaaaaahhh!” before ending the act and laughing.

“No, it's just so small that it can't pull you over when you're standing, and you don't usually notice it,” he insisted.

“Still, it makes no sense for everything to move the West mysteriously for no reason,” I said.

Isaac started grinning. “You're wrong,” he said. He picked up a pebble, dangled it out in front of himself, and dropped it. He paused for dramatic effect while I wondered what he was getting at. “I didn't push that pebble down,” he said, “It just mysteriously moved downwards for no reason.”

“Yeah, it fell. Things fall down.”

“Exactly! If things can be pulled downwards without anything touching them, why can't they be pulled a little bit West without anything touching them?”

“Ha! Well, if everything's getting pulled West, do spears veer off to the West when you throw them North or South?” I asked. I saw the confidence disappear from his face.

“Also, if there's a force pulling everything West that's just like the force pulling everything down except weaker, why would the pebbles curve to the West? Wouldn't they just move in a straight line that's sloped a little bit to the West?” Emmy added. Isaac looked like he was about to answer this, but then stopped, like it took him a moment to realize that he didn't have an answer.

“Nothing's getting mysteriously pulled West,” I said, “it's probably just that this tower is skewed a bit, so it looks like the pebbles move West when they actually fall straight down, just like everything always does. That's all.”

“Why do they curve, then?” Emmy asked.

“They probably don't,” I said, “They just went by fast enough that it was hard to tell exactly what the path looks like, and we tricked ourselves into thinking it was curved.” I couldn't believe I'd briefly bought into that nonsense about the pebbles falling in a curved path.

“Nope, definitely curved. We all saw it,” Emmy insisted. I argued about it with her for a bit, while Isaac just stood around looking confused.

“The pebbles do go in a curved path,” Isaac said a couple days later.

“Huh?” I said. I hadn't been thinking about the events a couple days prior, so it took me a moment to figure out what Isaac was talking about.

“The pebbles that you said must fall in a straight line from the top of the clock tower,” Isaac said, “Emmy and I tested it more precisely by dangling a rope off the top of the tower, and comparing the path the pebbles fell to the rope. The pebbles landed West of the end of the rope when we poured them directly in front of the rope. And you had almost convinced me that we were imagining the curving earlier, but when you compare it to the rope, it's harder to deny. The path was definitely curved. Which is pretty weird, when you think about it. Like, why would the rope just dangle straight down while the pebbles curve to the West? I think it might be that things only get pulled West when they're moving. The rope is just hanging there, not moving, so it doesn't get pulled West. But the pebbles are falling, so they get pulled West, and once they've fallen farther, they've picked up more speed, so they get pulled West harder, which would explain why they're curved. This could still explain why it's easier to throw a spear West than East, since the spear is moving, and why we don't feel ourselves getting pulled West, since we don't move very fast. But you had a good point about throwing spears North or South. They don't curve to the West at all. So maybe the direction it's moving matters. Things moving down, East, or West get pulled West, but things moving North or South don't. This seems pretty strange. Why would it work that way? I'm curious what happens to things that are moving up, but I can't figure out how to find out. It's hard to throw something straight up, and also hard to see what it's doing once you do. I did think of one thing we could try which would be really cool, but I don't think we could get enough rope. If we could stretch a piece of rope all the way around the world, and then pull both ends of the rope, we'd lift the whole rope up into the air. Then, if moving up also makes things get pulled West, we'd see the whole rope rotate to the West.”

I found basically everything he'd just said pretty implausible. “There definitely isn't enough rope to do that,” I said.

“Yeah, I know. I was just saying it would be really awesome if there was. And I'd be able to find out what happens to things that move straight up,” he said.

“Things get pushed to the East when they move up,” Isaac told me the next day.

“What convinced you of that?” I asked.

“Emmy and I cut a hole in a piece of wood to thread a rope through, tied a brick to one end of a long rope, and dropped the brick off the top of the clock tower while the rope was threaded through the hole in the wood. The other end of the rope moved East of the hole by the time it got pulled all the way up to the piece of wood. It actually took us a while to figure that out, since the wood was blocking our view of the end of the rope from the top of the tower. When one person watches from the ground, it's hard to see what the end of the rope is doing all the way up there from the ground; each of us took a turn watching from the ground, and neither of us could tell whether the end of the rope moved. So we dipped the end of the rope in paint and tried it again. The paint all splattered to the East of the hole. How weird is that? Things moving East, West, or down get pushed to the West, things moving up get pushed East, and things moving North or South don't seem to get pushed at all. Why? What's the pattern there? It makes no sense!” Isaac seemed oddly incensed about this.

“You're right about one thing, which is that that doesn't make any sense,” I said, “It was probably just from the rope randomly fluttering around. You don't need to postulate some sort of mysterious force that notices when things are moving and pushes them off in some other direction in order to explain a simple paint splatter.”

“It wasn't random,” he said, “We repeated it several times, sometimes changing details like what direction we held the brick away from the wood before dropping it. The paint always splatters East.”

A few days later, Isaac was out doing some target practice with a spear, and when he came back, he said, “You know, it occurred to me, if moving East or West causes things to get pushed West, then if you throw a spear in a diagonal direction, it's partially moving East or West as well as partially moving North or South, so it should veer off to the West. Like, if you throw it Southwest, it should veer off to the right. But that's not what actually happens. It just goes straight.”

“Ha! I told you it was all a bunch of nonsense,” I said.

“I thought of a better explanation, though,” he said, “I think when things move West, they get pushed a little bit up, and when things move East, they get pushed a little bit down. That still explains why it's easier to throw a spear West than East, because it gets a little boost upwards when thrown West and gets pushed a bit harder downwards when thrown East. This also means that it should be a little bit easier to throw it Southwest or Northwest, but not as easy as throwing it West, since it's partially moving West, and thus should get a smaller boost, and similarly, it should be a little bit harder to throw it Southeast or Northeast, but hot as hard as throwing it East. I think this is what actually happens, but it's hard to tell, since the effect is pretty subtle. And it makes so much more sense this way. Anything moving gets pushed in the direction that's 90 degrees clockwise from the direction it's moving, from the perspective of someone facing North. It's a clear pattern. I just still don't get why, though.”

I told him that that didn't make any more sense than what he'd been saying earlier, and he was probably imagining things. But he seemed pretty convinced that his new version of the story was better, somehow, and he kept trying to get me to help him come up with an explanation for it.

He never let go of this idea that things get pushed clockwise from the direction they're moving from the perspective of someone facing North. Every few months or so, I'd think he'd finally forgotten about the crazy idea, and he'd suddenly bring it up again, usually asking my opinion on some inane question like whether it had something to do with why things fall down, or if something moving West fast enough would fall up, or whether something could keep moving around in a big circle, moving down fast enough that it gets pushed West enough that its Westward movement makes it get pulled up, and its upward movement then making it get pulled East, and its Eastward movement making it move back down again. (I answered “no, of course not” to all three of those questions.)

Years later, Isaac and some of the other young men were having a contest to see which of them could throw a large rock the farthest. They were taking turns spinning around with the rock to gain speed and then throwing it forward, at which point others would mark where it hit the ground, and then the next person would bring it back and throw it again. When it was Isaac's turn, the rock landed right next the marker for where the rock had fallen from the best previous throw. They decided that Isaac's throw was a little bit shorter. I told them it looked like a tie to me, but they ignored me, and Isaac came in second place.

I tried to comfort Isaac about his loss afterwards, repeating that it looked like a tie to me, and saying he made a really great throw.

“It's only partially a throw,” he said, “It's also largely just spinning around and letting go. Once you're spinning around, the only thing stopping the rock from flying away is the fact that you're holding onto it, so you just have to let go to send it flying.” There was a pause before he continued, “I've got a riddle for you.” (that's what he says when he's about to ask a stupid question that he says he knows the answer to.) “When you're spinning around while holding a rock, if you pull the rock towards yourself, you'll spin faster. Does that mean it's a good idea to pull the rock towards yourself before letting go, to give it some extra speed?”

“No, of course not,” I said.

“Right, but why not?”

“Because you're trying to throw the rock away from you, not towards you. You'd be pulling it the wrong way,” I said.

“Sure, but then why would you spin faster when you pull the rock towards you?”

“I don't see what that has to do with throwing the rock forward.”

“The rock gets thrown forwards because of the speed it built up from spinning around, so if you spin faster, you should be able to throw it faster. But a big part of the reason pulling the rock inwards makes you spin faster is that, since the rock is closer to you, a full circle around you is shorter, so the speed the rock already had would take it all the way around you in less time. So the fact that you're spinning faster doesn't necessarily mean the rock is moving faster. It isn't clear to me what the effect of pulling the rock inwards on its speed is. The rock could speed up anyway, because if the rock is moving inwards while you're pulling it farther inwards, then you're pulling it in the direction of motion, which should make it speed up. Put differently, when you're a quarter-turn before the point where you let go of the rock, then pulling the rock inwards is actually the right direction. On the other hand, pulling the rock inwards could also slow it down, because once you pull it inwards, the rock is spinning around faster than you are, so it ends up pulling you forwards, which means you're also pulling the rock backwards, slowing it down. I think the second effect is probably bigger, so pulling the rock inward slows it down overall. In any case, the appearance that the rock moves forward faster when you pull it inwards is largely illusory, for the same reason it feels like the rock is pushing outwards in the first place; the rock's just trying to keep moving in the same way it's been going, but that isn't maintaining the same position relative to you.”

“Hm,” I said.

Isaac stopped talking for a bit, and I tried making small talk, but he seemed distracted and would only give at most two-syllable replies, so I gave up. I thought it was because he was sad over losing the contest, and forgot that he always acts like that right before saying something really inane. Isaac looked up for a bit.

“I figured it out,” said Isaac.

“Figured what out?”

“We're spinning.”

“Spinning?”

“Yeah. The world is spinning. You know how if you're spinning while holding a rock and you let go, it moves away from you just because it keeps moving the same way it was before? Well check this out,” he said, and jumped up to grab an apple from a tree that we were passing under. He held the apple in front of himself and dropped it. “Let go of the apple, and it seems to move away from the sun, the center of the world. Because the world is spinning. The apple just kept moving the same way it had already been moving, and the ground rose to meet it, because the ground is constantly being pulled towards the center of the world. That's why it's easier to throw a spear West than East. The world is spinning to the East, so anything moving East is spinning faster, so it moves away from the center of the world faster; that is, it falls faster. And anything moving West is spinning slower, so it falls slower. If something is very high, it is closer to the center of the world, so it's moving slower to keep up with the rotation of the world. When it falls, it gets farther away from the center of the world, so the same speed isn't enough to keep up with the world's rotation anymore, and part of its speed is directed down instead of East. Both of these effects make it look like it's moving West, opposite the direction of rotation. And if something moves up, the same speed makes it rotate around the center of the world faster than everything else, and the speed that was added to make it go up ends up pointing East. Both of these effects make it appear to move East.”

I laughed, and pointed out that I couldn't see anything spinning, but Isaac just said that's because I'm spinning the same way everything else was, so nothing would look out of place. I countered that if I jump, I land right back where I started instead of West of where I started, so the ground couldn't be rotating to the East under me. He had an answer to that, believe it or not. He said I keep moving East too, so I stay right over the same point on the ground.

And that's the story of how my son Isaac became convinced that the world is constantly spinning around in a circle. I think that's nonsense. The world doesn't look like it's spinning, and I don't think we need to suppose it is to explain the simple fact that things fall down and some mysterious forces acting on moving objects in Isaac's imagination.

## Metamathematics and probability

Content warning: mathematical logic.

Note: This write-up consists mainly of open questions rather than results, but may contain errors anyway.

### Setup

I'd like to describe a logic for talking about probabilities of logical sentences. Fix some first-order language ${\cal L}$. This logic deals with pairs $\left(\varphi,p\right)$, which I'm calling assertions, where $\varphi\in{\cal L}$ is a formula and $p\in\left[0,1\right]$. Such a pair is to be interpreted as a claim that $\varphi$ has probability at least $p$.

A theory consists of a set of assertions. A model of a theory $T$ consists of a probability space $\left(X,P\right)$ whose points are ${\cal L}$-structures, such that for every assertion $\left(\varphi,p\right)\in T$, $P_{*}\left(\left\{ {\cal M}\in X\mid{\cal M}\models\varphi\right\} \right)\geq p$, where $P_{*}$ is inner probability. I'll write $T\vdash_{p}\varphi$ for $\left(\varphi,p\right)$ can be proved from $T$, and $T\models_{p}\varphi$ for all models of $T$ are also models of $\left\{ \left(\varphi,p\right)\right\}$.

The rules of inference are all rules $\Gamma\vdash_{p}\varphi$ where $\Gamma$ is a finite set of assertions, and $\left(\varphi,p\right)$ is an assertion such that $P_{*}\left(\left\{ {\cal M}\in X\mid{\cal M}\models\varphi\right\} \right)\geq p$ in all models of $\Gamma$. Can we make an explicit finite list of inference rules that generate this logic? If not, is the set of inference rules at least recursively enumerable? (For recursive enumerability to make sense here, we need to restrict attention to probabilities in some countable dense subset of $\left[0,1\right]$ that has a natural explicit bijection with $\mathbb{N}$, such as $\mathbb{Q}\cap\left[0,1\right]$.) I'm going to assume later that the set of inference rules is recursively enumerable; if it isn't, everything should still work if we use some recursively enumerable subset of the inference rules that includes all of the ones that I use.

Note that the compactness theorem fails for this logic; for example, $\left\{ \left(\varphi,p\right)\mid p<1\right\} \models_{1}\varphi$, but no finite subset of $\left\{ \left(\varphi,p\right)\mid p<1\right\}$ implies $\left(\varphi,1\right)$, and hence $\left\{ \left(\varphi,p\right)\mid p<1\right\} \nvdash_{1}\varphi$.

Any classical first-order theory $T$ can be converted into a theory in this logic as $\left\{ \left(\varphi,1\right)\mid T\vdash\varphi\right\}$.

### Löb's Theorem

Let $T$ be a consistent, recursively axiomatizable extension of Peano Arithmetic. By the usual sort of construction, there is a $\Sigma_{1}^{0}$ binary predicate $\square_{y}\left(x\right)$ such that $T\vdash_{p}\varphi\iff\mathbb{N}\models\square_{p}\left(\ulcorner\varphi\urcorner\right)$ for any sentence $\varphi$ and $p\in\left[0,1\right]\cap\mathbb{Q}$, where $\ulcorner\urcorner$ is a coding of sentences with natural numbers. We have a probabilistic analog of Löb's theorem: if $T\vdash_{p}\square_{p}\left(\ulcorner\varphi\urcorner\right)\rightarrow\varphi$, then $T\vdash_{p}\varphi$. Peano arithmetic can prove this theorem, in the sense that $PA\vdash_{1}\square_{p}\left(\ulcorner\square_{p}\left(\ulcorner\varphi\urcorner\right)\rightarrow\varphi\urcorner\right)\rightarrow\square_{p}\left(\ulcorner\varphi\urcorner\right)$.

Proof: Assume $T\vdash_{p}\square_{p}\left(\ulcorner\varphi\urcorner\right)\rightarrow\varphi$. By the diagonal lemma, there is a sentence $\psi$ such that $T\vdash_{1}\psi\leftrightarrow\left(\square_{p}\left(\ulcorner\psi\urcorner\right)\rightarrow\varphi\right)$. If $\square_{p}\left(\ulcorner\psi\urcorner\right)$, then $\square_{1}\left(\ulcorner\square_{p}\left(\ulcorner\psi\urcorner\right)\urcorner\right)$ and $\square_{p}\left(\ulcorner\square_{p}\left(\ulcorner\psi\urcorner\right)\rightarrow\varphi\urcorner\right)$, so $\square_{p}\left(\ulcorner\varphi\urcorner\right)$. This shows that $T\cup\left\{ \left(\square_{p}\left(\ulcorner\psi\urcorner\right),1\right)\right\} \vdash_{1}\square_{p}\left(\ulcorner\varphi\urcorner\right)$. By the assumption that $T\vdash_{p}\square_{p}\left(\ulcorner\varphi\urcorner\right)\rightarrow\varphi$, this implies that $T\cup\left\{ \left(\square_{p}\left(\ulcorner\psi\urcorner\right),1\right)\right\} \vdash_{p}\varphi$. By a probabilistic version of the deduction theorem, $T\vdash_{p}\square_{p}\left(\ulcorner\psi\urcorner\right)\rightarrow\varphi$. That is, $T\vdash_{p}\psi$. Going back around through all that again, we get $T\vdash_{p}\varphi$.

If we change the assumption to be that $T\vdash_{q}\square_{p}\left(\ulcorner\varphi\urcorner\right)\rightarrow\varphi$ for some $q, then the above proof does not go through (if $q>p$, then it does, because $\left(\theta,q\right)\vdash_{p}\theta$). Is there a consistent theory extending Peano Arithmetic that proves a soundness schema about itself, $\left\{ \left(\square_{p}\left(\ulcorner\varphi\urcorner\right)\rightarrow\varphi,q\right)\mid q, or can this be used to derive a contradiction some other way? If there is no such consistent theory, then can the soundness schema be modified so that it is consistent, while still being nontrivial? If there is such a consistent theory with a soundness schema, can the theory also be sound? That is actually several questions, because there are multiple things I could mean by "sound". The possible syntactic things "sound" could mean, in decreasing order of strictness, are: 1) The theory does not assert a positive probability to any sentence that is false in $\mathbb{N}$. 2) There is an upper bound below $1$ for all probabilities asserted of sentences that are false in $\mathbb{N}$. 3) The theory does not assert probability $1$ to any sentence that is false in $\mathbb{N}$.

There are also semantic versions of the above questions, which are at least as strict as their syntactic analogs, but probably aren't equivalent to them, since the compactness theorem does not hold. The semantic version of asking if the soundness schema is consistent is asking if it has a model. The first two soundness notions also have semantic analogs. 1') $\left\{ \mathbb{N}\right\}$ is a model of the theory. 2') There is a model of the theory that assigns positive probability to $\mathbb{N}$. I don't have a semantic version of 3, but metaphorically speaking, a semantic version of 3 should mean that there is a model that assigns nonzero probability density at $\mathbb{N}$, even though it might not have a point mass at $\mathbb{N}$.

### Motivation

This is somewhat similar to Definability of Truth in Probabilistic Logic. But in place of adding a probability predicate to the language, I'm only changing the metalanguage to refer to probabilities, and using this to express statements about probability in the language through conventional metamathematics. An advantage of this approach is that it's constructive. Theories with the properties described by the Christiano et al paper are unsound, so if some reasonably strong notion of soundness applies to an extension of Peano Arithmetic with the soundness schema I described, that would be another advantage of my approach.

A type of situation that this might be useful for is that when an agent is reasoning about what actions it will take in the future, it should be able to trust its future self's reasoning. An agent with the soundness schema can assume that its future self's beliefs are accurate, up to arbitrarily small loss in precision. A related type of situation is if an agent reaches some conclusion, and then writes it to external storage instead of its own memory, and later reads the claim it had written to external storage. With the soundness schema, if the agent has reason to believe that the external storage hasn't been tampered with, it can reason that since its past self had derived the claim, the claim is to be trusted arbitrarily close to as much as it would have been if the agent had remembered it internally.

### First Incompleteness Theorem

For a consistent theory $T$, say that a sentence $\varphi$ is $T$-measurable if there is some $p\in\left[0,1\right]$ such that $T\vdash_{q}\varphi$ for every $q and $T\vdash_{q}\neg\varphi$ for every $q<1-p$. So $T$-measurability essentially means that $T$ pins down the probability of the sentence. If $\varphi$ is not $T$-measurable, then you could say that $T$ has Knightian uncertainty about $\varphi$. Say that $T$ is complete if every sentence is $T$-measurable. Essentially, complete theories assign a probability to every sentence, while incomplete theories have Knightian uncertainty.

The first incompleteness theorem (that no recursively axiomatizable extension of PA is consistent and complete) holds in this setting. In fact, for every consistent recursively axiomatizable extension of PA, there must be sentences that are given neither a nontrivial upper bound nor a nontrivial lower bound on their probability. Otherwise, we would be able to recursively separate the theorems of PA from the negations of theorems of PA, by picking some recursive enumeration of assertions of the theory, and sorting sentences by whether they are first given a nontrivial lower bound or first given a nontrivial upper bound; theorems of PA will only be given a nontrivial lower bound, and their negations will only be given a nontrivial upper bound. [Thanks to Sam Eisenstat for pointing this out; I had somehow managed not to notice this on my own.]

For an explicit example of a sentence for which no nontrivial bounds on its probability can be established, use the diagonal lemma to construct a sentence $\varphi$ which is provably equivalent to "for every proof of $\left(\varphi,p\right)$ for any $p>0$, there is a proof of $\left(\neg\varphi,q\right)$ for some $q>0$ with smaller Gödel number."

Thus a considerable amount of Knightian uncertainty is inevitable in this framework. Dogmatic Bayesians such as myself might find this unsatisfying, but I suspect that any attempt to unify probability and first-order arithmetic will suffer similar problems.

### A side note on model theory and compactness

I'm a bit unnerved about the compactness theorem failing. It occurred to me that it might be possible to fix this by letting models use hyperreal probabilities. Problem is, the hyperreals aren't complete, so the countable additivity axiom for probability measures doesn't mean anything, and it's unclear what a hyperreal-valued probability measure is. One possible solution is to drop countable additivity, and allow finitely-additive hyperreal-valued probability measures, but I'm worried that the logic might not even be sound for such models.

A different possible solution to this is to take a countably complete ultrafilter $U$ on a set $\kappa$, and use probabilities valued in the ultrapower $\mathbb{R}^{\kappa}/U$. Despite $\mathbb{R}^{\kappa}/U$ not being Cauchy complete, it inherits a notion of convergence of sequences from $\mathbb{R}$, since a sequence $\left\{ \left[x_{i,j}\mid i\in\kappa\right]\mid j\in\omega\right\}$ can be said to converge to $\left[\lim_{j\rightarrow\infty}x_{i,j}\mid i\in\kappa\right]$, and this is well-defined (if $\lim_{j\rightarrow\infty}x_{i,j}$ is for a $U$-large set of indices $i$) by countable completeness. Thus the countable additivity axiom makes sense for $\mathbb{R}^{\kappa}/U$-valued probability measures. Allowing models to use $\mathbb{R}^{\kappa}/U$-valued probability measures might make the compactness theorem work. [Edit: This doesn't work, because $\mathbb{R}^{\kappa}/U\cong\mathbb{R}$. To see this, it is enough to show that $\mathbb{R}^{\kappa}/U$ is Archimedean, since $\mathbb{R}$ has no proper Archimedean extensions. Given $\left[x_i\mid i\in\kappa\right]\in\mathbb{R}^{\kappa}/U$, let $A_n:=\{i\in\kappa\mid| x_i| for $n\in\mathbb{N}$. $\bigcup_{n\in\mathbb{N}}A_n = \kappa$, so by countable completeness of $U$, there is some $n\in\mathbb{N}$ such that $A_n\in U$, and thus $\left[x_i\mid i\in\kappa\right].]

## Complexity classes of natural numbers (googology for ultrafinitists)

Ultrafinitists think common ways of defining extremely large numbers don't actually refer to numbers that exist. For example, most ultrafinitists would maintain that a googolplex isn't a number. But to a classical mathematician, while numbers like a googolplex are far larger than the numbers we deal with on a day-to-day basis like 10, both numbers have the same ontological status. In this post, I want to consider a compromise position, that any number we can define can be meaningfully reasoned about, but that a special status is afforded to the sorts of numbers that ultrafinitists can accept.

Specifically, define an “ultrafinite number” to be a natural number that it is physically possible to express in unary. This isn't very precise, since there are all sorts of things that “physically possible to express in unary” could mean, but let's just not worry about that too much. Also, many ultrafinitists would not insist that numbers must be expressible in such an austere language as unary, but I'm about to get to that.

Examples: $20$ is an ultrafinite number, because $20 = SSSSSSSSSSSSSSSSSSSS0$, where $S$ is the successor function. 80,000 is also an ultrafinite number, but it is a large one, and it isn't worth demonstrating its ultrafiniteness. A googol is not ultrafinite. The observable universe isn't even big enough to contain a googol written in unary.

Now, define a “polynomially finite number” to be a natural number that it is physically possible to express using addition and multiplication. Binary and decimal are basically just concise ways of expressing certain sequences of addition and multiplication operations. For instance, “$18,526$” means $(((1*10 + 8)*10 + 5)*10 + 2)*10 + 6$. Conversely, if you multiply an $n$-digit number with an $m$-digit number, you get an at most $n+m+1$-digit number, which is the same number of symbols it took write down “[the $n$-digit number] times [the $m$-digit number]” in the first place, so any number that can be written using addition and multiplication can be written in decimal. Thus, another way to define polynomially finite numbers is as the numbers that it is physically possible to express in binary or in decimal. I've been ignoring some small constant factors that might make these definitions not quite equivalent, but any plausible candidate for a counterexample would be an ambiguous edge case according to each definition anyway, so I'm not worried about that. Many ultrafinitists may see something more like polynomially finite number, rather than ultrafinite number, as a good description of what numbers exist.

Examples: A googol is polynomially finite, because a googol is 10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000. A googolplex is not polynomially finite, because it would require a googol digits to express in decimal, which is physically impossible.

Define an “elementarily finite number” to be a number that it is physically possible to express using addition, multiplication, subtraction, exponentiation, and the integer division function $\lfloor x/y\rfloor$. Elementarily finite is much broader than polynomially finite, so it might make sense to look at intermediate classes. Say a number is “exponentially finite” if it is physically possible to express using the above operations but without any nested exponentiation (e.g. $a^b c^d$ is okay, but $a^{(b^c)}$ is not). More generally, we can say that a number is “$k$-exponentially finite” if it can be expressed with exponentiation nested to depth at most $k$, so a polynomially finite number is a $0$-exponentially finite number, an exponentially finite number is a $1$-exponentially finite number, and an elementarily finite number is a number that is $k$-exponentially finite for some $k$ (or equivalently, for some ultrafinite $k$).

Examples: a googolplex is exponentially finite, because it is $10^{10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000}$. Thus a googolduplex, meaning $10^\text{googolplex}$, is $2$-exponentially finite, but it is not exponentially finite. For examples of non-elementarily finite numbers, and numbers that are only $k$-exponentially finite for fairly large $k$, I'll use up-arrow notation. $a\uparrow b$ just means $a^b$, $a\uparrow^{n+1} b$ means $a\uparrow^n a\uparrow^n a\uparrow^n ... a$, where $b$ is the number of copies of $a$, and using order of operations that starts on the right. So $3\uparrow\uparrow3 = 3^{(3^3)} = 3^{27} = 7,625,597,484,987$, which is certainly polynomially finite, and could also be ultrafinite depending on what is meant by “physically possible” (a human cannot possibly count that high, but a computer with a large enough hard drive can store $3\uparrow\uparrow3$ in unary). $3\uparrow\uparrow\uparrow3 = 3\uparrow\uparrow(3\uparrow\uparrow3) = 3\text{^}3\text{^}3\text{^}...\text{^}3$, where there are $3^{27}$ threes in that tower. Under the assumptions that imply $3^{27}$ is ultrafinite, $3\uparrow\uparrow\uparrow3$ is elementarily finite. Specifically, it is $3^{27}$-exponentially finite, but I'm pretty sure it's not $3^{26}$-exponentially finite, or even $7,625,597,484,000$-exponentially finite. $3\uparrow\uparrow\uparrow\uparrow3 = 3\uparrow\uparrow\uparrow(3\uparrow\uparrow\uparrow3)$, and is certainly not elementarily finite.

Interestingly, even though a googolplex is exponentially finite, there are numbers less than a googolplex that are not. There's an easy nonconstructive proof of this: in order to be able to represent every number less than a googolplex in any encoding scheme at all, there has to be some number less than a googolplex that requires at least a googol decimal digits of information to express. But it is physically impossible to store a googol decimal digits of information. Therefore for any encoding scheme for numbers, there is some number less than a googolplex that cannot physically be expressed in it. This is why the definition of elementarily finite is significantly more complicated than the definition of polynomially finite; in the polynomial case, if $n$ can be expressed using addition and multiplication and $m, then $m$ can also be expressed using addition and multiplication, so there's no need for additional operations to construct smaller numbers, but in the elementary case, the operations of subtraction and integer division are useful for expressing more numbers, and are simpler than exponentiation. For example, these let us express the number that you get from reading off the last googol digits, or the first googol digits, of $3\uparrow\uparrow100$, so these numbers are elementarily finite. However, it is exceptionally unlikely that the number you get from reading off the first googol decimal digits of $3\uparrow\uparrow\uparrow\uparrow3$ is elementarily finite. But for a difficult exercise, show that the number you get from reading off the last googol decimal digits of $3\uparrow\uparrow\uparrow\uparrow3$ is elementarily finite.

Why stop there instead of including more operations for getting smaller numbers, like $\lfloor\log\rfloor$, which I implicitly used when I told you that the number formed by the first googol digits of $3\uparrow\uparrow100$ is elementarily finite? We don't have to. The functions that you can get by composition from addition, multiplication, exponentiation, $\max(x-y,0)$, and $\lfloor x/y\rfloor$ coincide with the functions that can be computed in iterated exponential time (meaning $O(2^{2^{...^{2^n}}})$ time, for some height of that tower). So if you have any remotely close to efficient way to compute an operation, it can be expressed in terms of the operations I already specified.

We can go farther. Consider a programming language that has the basic arithmetic operations, if/else clauses, and loops, where the number of iterations of each loop must be fixed in advance. The programs that can be written in such a language are the primitive recursive functions. Say that a number is primitive recursively finite if it is physically possible to write a program (that does not take any input) in this language that outputs it. For each fixed $n$, the binary function $(x,y)\mapsto x\uparrow^n y$ is primitive recursive, so $3\uparrow\uparrow\uparrow\uparrow3$ is primitive recursively finite. But the ternary function $(x,y,z)\mapsto x\uparrow^y z$ is not primitive recursive, so $3\uparrow^\text{googol}3$ is not primitive recursively finite.

The primitive recursively finite numbers can be put in a hierarchy of subclasses based on the depth of nested loops that are needed to express them. If the only arithmetic operation available is the successor function (from which other operations can be defined using loops), then the elementarily finite numbers are those that can be expressed with loops nested to depth at most 2. The $k$-exponentially finite numbers should roughly correspond to the numbers that can be expressed with at most $k$ loops at depth 2.

Next comes the provably computably finite numbers. Say that a number is provably computably finite if it is physically possible to write a program in a Turing-complete language that outputs the number (taking no input), together with a proof in Peano Arithmetic that the program halts. The famous Graham's number is provably computably finite. Graham's number is defined in terms of a function $g$, defined recursively as $g(0):=4$ and $g(n+1):=3\uparrow^{g(n)}3$. Graham's number is $g(64)$. You could write a computer program to compute $g$, and prove that $g$ is total using Peano arithmetic. By replacing Peano arithmetic with other formal systems, you can get other variations on the notion of provably computably finite.

For an example of a number that is not provably computably finite, I'll use the hydra game, which is described here. There is no proof in Peano arithmetic (that can physically be written down) that it is possible to win the hydra game starting from the complete binary tree of depth a googol. So the number of turns it takes to win the hydra game on the complete binary tree of depth a googol is not provably computably finite. If you start with a reasonably small hydra (say, with 100 nodes), you could write a program to search for the shortest winning strategy, and prove in Peano arithmetic that it succeeds in finding one, if you're sufficiently clever and determined, and you use a computer to help you search for proofs. The proof you'd get out of this endeavor would be profoundly unenlightening, but the point is, the number of turns it takes to win the hydra game for a small hydra is provably computably finite (but not primitive recursively finite, except in certain trivial special cases).

Next we'll drop the provability requirement, and say that a number is computably finite if it is physically possible to write a computer program that computes it from no input. Of course, in order to describe a computably finite number, you need the program you use to actually halt, so you'd need some argument that it does halt in order to establish that you're describing a computably finite number. Thus this is arguably just a variation on provably computably finite, where Peano arithmetic is replaced by some unspecified strong theory encompassing the sort of reasoning that classical mathematicians tend to endorse. This is probably the point where even the most patient of ultrafinitists would roll their eyes in disgust, but oh well. Anyway, the number of steps that it takes to win the hydra game starting from the complete binary tree of depth a googol is a computably finite number, because there exists a shortest winning strategy, and you can write a computer program to exhaustively search for it.

The busy-beaver function $BB$ is defined so that $BB(n)$ is the longest any Turing machine with $n$ states runs before halting (among those that do halt). $BB(10^{100})$ is not computably finite, because Turing machines with a googol states cannot be explicitly described, and since the busy-beaver function is very fast-growing, no smaller Turing machine has comparable behavior. What about $BB(10,000)$? Turing machines with 10,000 states are not too big to describe explicitly, so it may be tempting to say that $BB(10,000)$ is computably finite. But on the other hand, it is not possible to search through all Turing machines with 10,000 states and find the one that runs the longest before halting. No matter how hard you search and no matter how clever your heuristics for finding Turing machines that run for exceptionally long and then halt, it is vanishingly unlikely that you will find the 10,000-state Turing machine that runs longest before halting, let alone realize that you have found it. And the idea is to use classical reasoning for large numbers themselves, but constructive reasoning for descriptions of large numbers. So since it is pretty much impossible to actually write a program that outputs $BB(10,000)$, it is not computably finite.

For a class that can handle busy-beaver numbers too, let's turn to the arithmetically finite numbers. These are the numbers that are defined by arithmetical formulas. These form a natural hierarchy, where the $\Sigma^0_n$-finite numbers are the numbers defined by arithmetical formulas with at most $n$ unbounded quantifiers starting with $\exists$, the $\Pi^0_n$-finite numbers are the numbers defined by arithmetical formulas with at most $n$ unbounded quantifiers starting with $\forall$, and the $\Delta^0_n$-finite numbers are those that are both $\Sigma^0_n$-finite and $\Pi^0_n$-finite. The $\Delta^0_1$-finite numbers are the same as the computably finite numbers. $BB(10^{100})$ is $\Pi^0_2$-finite, because it is defined by “$\forall n$ every Turing machine with $10^{100}$ states that halts in at most $n$ steps halts in at most $BB(10^{100})$ steps, and there is a Turing machine with $10^{100}$ states that halts in exactly $BB(10^{100})$ steps.” Everything after the first quantifier in that formula is computable. $BB(BB(10^{100}))$ is $\Delta^0_2$-finite, but no lower than that. To get a number that is not arithmetically finite, consider the function $f$ given by $f(n)$ is the largest number defined by an arithmetical formula with $n$ symbols. $f(10,000)$ is $\Delta^0_{5,000}$-finite, but $f(10^{100})$ is not arithmetically finite. I'll stop there.

## Existential risk from AI without an intelligence explosion

[x-post LessWrong]

In discussions of existential risk from AI, it is often assumed that the existential catastrophe would follow an intelligence explosion, in which an AI creates a more capable AI, which in turn creates a yet more capable AI, and so on, a feedback loop that eventually produces an AI whose cognitive power vastly surpasses that of humans, which would be able to obtain a decisive strategic advantage over humanity, allowing it to pursue its own goals without effective human interference. Victoria Krakovna points out that many arguments that AI could present an existential risk do not rely on an intelligence explosion. I want to look in sightly more detail at how that could happen. Kaj Sotala also discusses this.

An AI starts an intelligence explosion when its ability to create better AIs surpasses that of human AI researchers by a sufficient margin (provided the AI is motivated to do so). An AI attains a decisive strategic advantage when its ability to optimize the universe surpasses that of humanity by a sufficient margin. Which of these happens first depends on what skills AIs have the advantage at relative to humans. If AIs are better at programming AIs than they are at taking over the world, then an intelligence explosion will happen first, and it will then be able to get a decisive strategic advantage soon after. But if AIs are better at taking over the world than they are at programming AIs, then an AI would get a decisive strategic advantage without an intelligence explosion occurring first.

Since an intelligence explosion happening first is usually considered the default assumption, I'll just sketch a plausibility argument for the reverse. There's a lot of variation in how easy cognitive tasks are for AIs compared to humans. Since programming AIs is not yet a task that AIs can do well, it doesn't seem like it should be a priori surprising if programming AIs turned out to be an extremely difficult task for AIs to accomplish, relative to humans. Taking over the world is also plausibly especially difficult for AIs, but I don't see strong reasons for confidence that it would be harder for AIs than starting an intelligence explosion would be. It's possible that an AI with significantly but not vastly superhuman abilities in some domains could identify some vulnerability that it could exploit to gain power, which humans would never think of. Or an AI could be enough better than humans at forms of engineering other than AI programming (perhaps molecular manufacturing) that it could build physical machines that could out-compete humans, though this would require it to obtain the resources necessary to produce them.

Furthermore, an AI that is capable of producing a more capable AI may refrain from doing so if it is unable to solve the AI alignment problem for itself; that is, if it can create a more intelligent AI, but not one that shares its preferences. This seems unlikely if the AI has an explicit description of its preferences. But if the AI, like humans and most contemporary AI, lacks an explicit description of its preferences, then the difficulty of the AI alignment problem could be an obstacle to an intelligence explosion occurring.

It also seems worth thinking about the policy implications of the differences between existential catastrophes from AI that follow an intelligence explosion versus those that don't. For instance, AIs that attempt to attain a decisive strategic advantage without undergoing an intelligence explosion will exceed human cognitive capabilities by a smaller margin, and thus would likely attain strategic advantages that are less decisive, and would be more likely to fail. Thus containment strategies are probably more useful for addressing risks that don't involve an intelligence explosion, while attempts to contain a post-intelligence explosion AI are probably pretty much hopeless (although it may be worthwhile to find ways to interrupt an intelligence explosion while it is beginning). Risks not involving an intelligence explosion may be more predictable in advance, since they don't involve a rapid increase in the AI's abilities, and would thus be easier to deal with at the last minute, so it might make sense far in advance to focus disproportionately on risks that do involve an intelligence explosion.

It seems likely that AI alignment would be easier for AIs that do not undergo an intelligence explosion, since it is more likely to be possible to monitor and do something about it if it goes wrong, and lower optimization power means lower ability to exploit the difference between the goals the AI was given and the goals that were intended, if we are only able to specify our goals approximately. The first of those reasons applies to any AI that attempts to attain a decisive strategic advantage without first undergoing an intelligence explosion, whereas the second only applies to AIs that do not undergo an intelligence explosion ever. Because of these, it might make sense to attempt to decrease the chance that the first AI to attain a decisive strategic advantage undergoes an intelligence explosion beforehand, as well as the chance that it undergoes an intelligence explosion ever, though preventing the latter may be much more difficult. However, some strategies to achieve this may have undesirable side-effects; for instance, as mentioned earlier, AIs whose preferences are not explicitly described seem more likely to attain a decisive strategic advantage without first undergoing an intelligence explosion, but such AIs are probably more difficult to align with human values.

If AIs get a decisive strategic advantage over humans without an intelligence explosion, then since this would likely involve the decisive strategic advantage being obtained much more slowly, it would be much more likely for multiple, and possibly many, AIs to gain decisive strategic advantages over humans, though not necessarily over each other, resulting in a multipolar outcome. Thus considerations about multipolar versus singleton scenarios also apply to decisive strategic advantage-first versus intelligence explosion-first scenarios.

## Principal Component Analysis in Theory and Practice

Prerequisites for this post are linear algebra, including tensors, and basic probability theory. Already knowing how PCA works will also be helpful. In section 1, I'll summarize the technique of principal component analysis (PCA), stubbornly doing so in a coordinate-free manner, partly because I am an asshole but mostly because it is rhetorically useful for emphasizing my point in section 2. In section 2, I'll gripe about how PCA is often used in ways that shouldn't be expected to work, but works just fine anyway. In section 3, I'll discuss some useless but potentially amusing ways that PCA could be modified. Thanks to Laurens Gunnarsen for inspiring this post by talking to me about the problem that I discuss in section 2.

### A brief introduction to Principal Component Analysis

You start with a finite-dimensional real inner product space $V$ and a probability distribution $\mu$ on $V$. Actually, you probably just started with a large finite number of elements of $V$, and you've inferred a probability distribution that you're supposing they came from, but that difference is not important here. The goal is to find the $n$-dimensional (for some $n\leq\dim V$) affine subspace $W_{n}\subseteq V$ minimizing the expected squared distance between a vector (distributed according to $\mu$) and its orthogonal projection onto $W_{n}$. We can assume without loss of generality that the mean of $\mu$ is $0$, because we can just shift any probability distribution by its mean and get a probability distribution with mean $0$. This is useful because then $W_{n}$ will be a linear subspace of $V$. In fact, we will solve this problem for all $n\leq\dim V$ simultaneously by finding an ordered orthonormal basis such that $W_{n}$ is the span of the first $n$ basis elements.

First you take $\text{Cov}_{\mu}\in V\otimes V$, called the covariance of $\mu$, defined as the bilinear form on $V^{*}$ given by $\text{Cov}_{\mu}\left(\varphi,\psi\right)=\int_{V}\varphi\left(x\right)\psi\left(x\right)d\mu\left(x\right)$. From this, we get the covariance operator $C_{\mu}\in V^{*}\otimes V$ by raising the first index, which means starting with $\left\langle \cdot,\cdot\right\rangle \otimes\text{Cov}_{\mu}\in V^{*}\otimes V^{*}\otimes V\otimes V$ and performing a tensor contraction (in other words, $C_{\mu}$ is obtained from $\text{Cov}_{\mu}$ by applying the map $V\rightarrow V^{*}$ given by the inner product to the first index). $\text{Cov}_{\mu}$ is symmetric and positive semi-definite, so $C_{\mu}$ is self-adjoint and positive semi-definite, and hence $V$ has an orthonormal basis of eigenvectors of $C_{\mu}$, with non-negative real eigenvalues. This gives an orthonormal basis in which $\text{Cov}_{\mu}$ is diagonal, where the diagonal entries are the eigenvalues. Ordering the eigenvectors in decreasing order of the corresponding eigenvalues gives us the desired ordered orthonormal basis.

### The problem

There's no problem with principal component analysis as I described it above. It works just fine, and in fact is quite beautiful. But often people apply principal component analysis to probability distributions on finite-dimensional real vector spaces that don't have a natural inner product structure. There are two closely related problems with this: First, the goal is underdefined. We want to find a projection onto an $n$-dimensional subspace that minimizes the expected squared distance from a vector to its projection, but we don't have a measure of distance. Second, the procedure is underdefined. $\text{Cov}_{\mu}$ is a bilinear form, not a linear operator, so it doesn't have eigenvectors or eigenvalues, and we don't have a way of raising an index to produce something that does. It should come as no surprise that these two problems arise together. After all, you shouldn't be able to find a fully specified solution to an underspecified problem.

People will apply principal component analysis in such cases by picking an inner product. This solves the second problem, since it allows you to carry out the procedure. But it does not solve the first problem. If you wanted to find a projection onto an $n$-dimensional subspace such that the distance from a vector to its projection tends to be small, then you must have already had some notion of distance in mind by which to judge success. Haphazardly picking an inner product gives you a new notion of distance, and then allows you to find an optimal solution with respect to your new notion of distance, and it is not clear to me why you should expect this solution to be reasonable with respect to the notion of distance that you actually care about.

In fact, it's worse than that. Of course, principal component analysis can't given you literally any ordered basis at all, but it is almost as bad. The thing that you use PCA for is the projection onto the span of the first $n$ basis elements along the span of the rest. These projections only depend on the sequence of $1$-dimensional subspaces spanned by the basis elements, and not the basis elements themselves. That is, we might as well only pay attention to the principal components up to scale, rather than making sure that are all unit length. Let a "coordinate system" refer to an ordered basis up to two ordered bases being equivalent if they differ only by scaling the basis vectors, so that we're paying attention to the coordinate systems given to us by PCA. If the covariance of $\mu$ is nondegenerate, then the set of coordinate systems that can be obtained from principal component analysis by a suitable choice of inner product is dense in the space of coordinate systems. More generally, where $U$ is the smallest subspace of $V$ such that $\mu\left(U\right)=1$, then the space of coordinate systems that you can get from principal component analysis is dense in the space of all coordinate systems whose first $\dim U$ coordinates span $U$ ($\dim U$ will be the rank of the covariance of $\mu$). So in a sense, for suitably poor choices of inner product, principal component analysis can give you arbitrarily terrible results, subject only to the weak constraint that it will always notice if all of the vectors in your sample belong to a common subspace.

It is thus somewhat mysterious that machine learning people seem to be able to often get good results from principal component analysis apparently without being very careful about the inner product they choose. Vector spaces that arise in machine learning seem to almost always come with a set of preferred coordinate axes, so these axes are taken to be orthogonal, leaving only the question of how to scale them relative to each other. If these axes are all labeled with the same units, then this also gives you a way of scaling them relative to each other, and hence an inner product. If they are aren't, then I'm under the impression that the most popular method is to normalize them such that the pushforward of $\mu$ along each coordinate axis has the same variance. This is unsatisfying, since figuring out which axes $\mu$ has enough variance along to be worth paying attention to seems like the sort of thing that you would want principal component analysis to be able to tell you. Normalizing the axes in this way seems to me like an admission that you don't know exactly what question you're hoping to use principal component analysis to answer, so you just tell it not to answer that part of the question to minimize the risk of asking it to answer the wrong question, and let it focus on telling you how the axes, which you're pretty sure should be considered orthogonal, correlate with each other.

That conservatism is actually pretty understandable, because figuring out how to ask the right question seems hard. You implicitly have some metric $d$ on $V$ such that you want to find a projection $\pi$ onto an $n$-dimensional subspace such that $d\left(x,\pi\left(x\right)\right)$ is usually small when $x$ is distributed according to $\mu$. This metric is probably very difficult to describe explicitly, and might not be the metric induced by any inner product (for that matter, it might not even be a metric; $d\left(x,y\right)$ could be any way of quantifying how bad it is to be told the value $y$ when the correct value you wanted to know is $x$). Even if you somehow manage to explicitly describe your metric, coming up with a version of PCA with the inner product replaced with an arbitrary metric also sounds hard, so the next thing you would want to do is fit an inner product to the metric.
The usual approach is essentially to skip the step of attempting to explicitly describe the metric, and just find an inner product that roughly approximates your implicit metric based on some rough heuristics about what the implicit metric probably looks like. The fact that these heuristics usually work so well seems to indicate that the implicit metric tends to be fairly tame with respect to ways of describing the data that we find most natural. Perhaps this shouldn't be too surprising, but I still feel like this explanation does not make it obvious a priori that this should work so well in practice. It might be interesting to look into why these heuristics work as well as they do with more precision, and how to go about fitting a better inner product to implicit metrics. Perhaps this has been done, and I just haven't found it.

To take a concrete example, consider eigenfaces, the principal components that you get from a set of images of people's faces. Here, you start with the coordinates in which each coordinate axis represents a pixel in the image, and the value of that coordinate is the brightness of the corresponding pixel. By declaring that the coordinate axes are orthogonal, and measuring the brightness of each pixel on the same scale, we get our inner product, which is arguably a fairly natural one.

Presumably, the implicit metric we're using here is visual distance, by which I mean a measure of how similar two images look. It seems clear to me that visual distance is not very well approximated by our inner product, and in fact, there is no norm such that the visual distance between two images is approximately the norm of their difference. To see this, if you take an image and make it brighter, you haven't changed how it looks very much, so the visual distance between the image and its brighter version is small. But their difference is just a dimmer version of the same image, and if you add that difference to a completely different image, you will get the two images superimposed on top of each other, a fairly radical change. Thus the visual distance traversed by adding a vector depends on where you start from.

Despite this, producing eigenfaces by using PCA on images of faces, using the inner product described above, performs well with respect to visual distance, in the sense that you can project the images onto a relatively small number of principal components and leave them still recognizable. I think this can be explained on an intuitive level. In a human eye, each photoreceptor has a narrow receptive field that it detects light in, much like a pixel, so the representation of an image in the eye as patterns of photoreceptor activity is very similar to the representation of an image in a computer as a vector of pixel brightnesses, and the inner product metric is a reasonable measure of distance in this representation. When the visual cortex processes this information from the eye, it is difficult (and perhaps also not useful) for it to make vast distinctions between images that are close to each other according to the inner product metric, and thus result in similar patterns of photoreceptor activity in the eye. Thus the visual distance between two images cannot be too much greater than their inner product distance, and hence changing an image by a small amount according to the inner product metric can only change it by a small amount according to visual distance, even though the reverse is not true.

### Generalizations

The serious part of this post is now over. Let's have some fun. Some of the following ways of modifying principal component analysis could be combined, but I'll consider them one at a time for simplicity.

As hinted at above, you could start with an arbitrary metric on $V$ rather than an inner product, and try to find the rank-$n$ projection (for some $n\leq\dim V$) that minimizes the expected squared distance from a vector to its projection. This would probably be difficult, messy, and not that much like principal component analysis. If it can be done, it would be useful in practice if we were much better at fitting explicit metrics to our implicit metrics than at fitting inner products to our implicit metrics, but I'm under the impression that this is not currently the case. This also differs from the other proposals in this section in that it is a modification of the problem looking for a solution, rather than a modification of the solution looking for a problem.

$V$ could be a real Hilbert space that is not necessarily finite-dimensional. Here we can run into the problem that $C_{\mu}$ might not even have any eigenvectors. However, if $\mu$ (which hopefully was not inferred from a finite sample) is Gaussian (and possibly also under weaker conditions), then $C_{\mu}$ is a compact operator, so $V$ does have an orthonormal basis of eigenvectors of $C_{\mu}$, which still have non-negative eigenvalues. There probably aren't any guarantees you can get about the order-type of this orthonormal basis when you order the eigenvectors in decreasing order of their eigenvalues, and there probably isn't a sense in which the orthogonal projection onto the closure of the span of an initial segment of the basis accounts for the most variance of any closed subspace of the same "size" ("size" would have to refer to a refinement of the notion of dimension for this to be the case). However, a weaker statement is probably still true: namely that each orthonormal basis element maximizes the variance that it accounts for conditioned on values along the previous orthonormal basis elements. I guess considering infinite-dimensional vector spaces goes against the spirit of machine learning though.

$V$ could be a finite-dimensional complex inner product space. $\text{Cov}_{\mu}\in\overline{V}\otimes V$ would be the sesquilinear form on $V^{*}$ given by $\text{Cov}_{\mu}\left(\varphi,\psi\right)=\int_{V}\overline{\varphi\left(x\right)}\psi\left(x\right)d\mu\left(x\right)$. $\left\langle \cdot,\cdot\right\rangle \in\overline{V}^{*}\otimes V^{*}$, so $\left\langle \cdot,\cdot\right\rangle \otimes\text{Cov}_{\mu}\in\overline{V}^{*}\otimes V^{*}\otimes\overline{V}\otimes V$, and applying a tensor contraction to the conjugated indices gives us our covariance operator $C_{\mu}\in V^{*}\otimes V$ (in other words, the inner product gives us an isomorphism $\overline{V}\rightarrow V^{*}$, and applying this to the first index of $\text{Cov}_{\mu}$ gives us $C_{\mu}\in V^{*}\otimes V$). $C_{\mu}$ is still self-adjoint and positive semi-definite, so $V$ still has an orthonormal basis of eigenvectors with non-negative real eigenvalues, and we can order the basis in decreasing order of the eigenvalues. Analogously to the real case, projecting onto the span of the first $n$ basis vectors along the span of the rest is the complex rank-$n$ projection that minimizes the expected squared distance from a vector to its projection. As far as I know, machine learning tends to deal with real data, but if you have complex data and for some reason you want to project onto a lower-dimensional complex subspace without losing too much information, now you know what to do.

Suppose your sample consists of events, where you've labeled them with both their spatial location and the time at which they occurred. In this case, events are represented as points in Minkowski space, a four-dimensional vector space representing flat spacetime, which is equipped with a nondegenerate symmetric bilinear form called the Minkowski inner product, even though it is not an inner product because it is not positive-definite. Instead, the Minkowski inner product is such that $\left\langle x,x\right\rangle$ is positive if $x$ is a space-like vector, negative if $x$ is time-like, and zero if $x$ is light-like. We can still get $C_{\mu}\in V^{*}\otimes V$ out of $\text{Cov}_{\mu}\in V\otimes V$ and the Minkowski inner product in $V^{*}\otimes V^{*}$ in the same way, and $V$ has a basis of eigenvectors of $C_{\mu}$, and we can still order the basis in decreasing order of their eigenvalues. The first 3 eigenvectors will be space-like, with non-negative eigenvalues, and the last eigenvector will be time-like, with a non-positive eigenvalue. The eigenvectors are still orthogonal. Thus principal component analysis provides us with a reference frame in which the span of the first 3 eigenvectors is simultaneous, and the span of the last eigenvector is motionless. If $\mu$ is Gaussian, then this will be the reference frame in which the spatial position of an event and the time at which it occurs are mean independent of each other, meaning that conditioning on one of them doesn't change the expected value of the other one. For general $\mu$, there might not be a reference frame in which the space and time of an event are mean independent, but the reference frame given to you by by principal component analysis is still the unique reference frame with the property that the time coordinate is uncorrelated with any spatial coordinate.

More generally, we could consider $V$ equipped with any symmetric bilinear form $\left\langle \cdot,\cdot\right\rangle$ taking the role of the inner product. Without loss of generality, we can consider only nondegenerate symmetric bilinear forms, because in the general case, where $D:=\left\{ x\mid\forall y\,\left\langle x,y\right\rangle =0\right\}$, applying principal component analysis with $\left\langle \cdot,\cdot\right\rangle$ is equivalent to projecting the data onto $V/D$, applying principal component analysis there with the nondegenerate symmetric bilinear form on $V/D$ induced by $\left\langle \cdot,\cdot\right\rangle$, and then lifting back to $V$ and throwing in a basis for $D$ with eigenvalues $0$ at the end, essentially treating $D$ as the space of completely irrelevant distinctions between data points that we intend to immediately forget about. Anyway, nondegenerate symmetric bilinear forms are classified up to isomorphism by their signature $\left(n,m\right)$, which is such that any orthogonal basis contains exactly $n+m$ basis elements, $n$ of which are space-like and $m$ of which are time-like, using the convention that $x$ is space-like if $\left\langle x,x\right\rangle >0$, time-like if $\left\langle x,x\right\rangle <0$, and light-like if $\left\langle x,x\right\rangle =0$, as above. Using principal component analysis on probability distributions over points in spacetime (or rather, points in the tangent space to spacetime at a point, so that it is a vector space) in a universe with $n$ spatial dimensions and $m$ temporal dimensions still gives you a reference frame in which the span of the first $n$ basis vectors is simultaneous and the span of the last $m$ basis vectors is motionless, and this is again the unique reference frame in which each time coordinate is uncorrelated with each spatial coordinate. Incidentally, I've heard that much of physics still works with multiple temporal dimensions. I don't know what that would mean, except that I think it means there's something wrong with my intuitive understanding of time. But that's another story. Anyway, the spaces spanned by the first $n$ and by the last $m$ basis vectors could be used to establish a reference frame, and then the data might be projected onto the first few (at most $n$) and last few (at most $m$) coordinates to approximate the positions of the events in space and in time, respectively, in that reference frame.