What I Would Do With Immortality

I’ve previously discussed that I didn’t like reading fiction growing up, because I knew that if I thought about it too hard, it would break down. In real life, you can do experiments to answer your questions, instead of needing to rely on authority or source material, but in a story, this isn’t true. The real universe is modeled by quarks; stories are modeled by plots.

The fact that real life is based on universally consistent laws is a nearly endless source of intrigue, entertainment, and general fun, at least for me. Whenever I ask a question about reality, I know that it has an answer, somewhere. If I don’t know it, I can learn about it from someone who does, and if nobody knows it, I can find the answer myself. The existence of a consistent reality that I can do experiments on means that I am not limited in my ability to learn stuff by anything besides my willingness to do so.

The primary reason that I haven’t gone on a quest to rediscover every single insight ever made by the human race – which seems to approximate the Maximum Fun Plan – is because doing so would take orders of magnitude more years than I am presently expected to live, barring major advances in medical science. So, I’ve got to solve the pesky mortality problem first. But once I do… I certainly plan to spend a lot of time rediscovering things.

It might seem a bit odd, that I would want to spend decades and centuries rediscovering things that other people already know. A waste of effort, isn’t it? It would be more efficient to ask somebody who knows about the thing already.

More efficient it may be, indeed – which is why I don’t do it right now – but more fun it is not. I’m certain that Isaac Newton had way more fun inventing calculus than I had learning about it in school, and that isn’t just because our modern school system is a train-wreck. The joy of discovering something for myself is substantially greater than that of hearing the solution from somewhere else before I’ve even tried my hand at the problem. (I do prefer that the solution be printed somewhere, especially if the experiments to confirm my solution are difficult to create. It would be nice to hear somebody else’s solution to the AI-box problem, for example.)

Not only is the joy of the knowledge-acquisition inherently less, but the quality of the knowledge itself is also lower. When you discover something for yourself, you don’t have the problem of storing as “knowledge” what is actually just a referentless pointer (ie. a physicist tells you that “light is waves”, and you store the phrase “light is waves”, but you don’t have the background knowledge to really know what it means, and you couldn’t regenerate the “knowledge” if it were deleted from your brain). You also won’t have the potential pitfall of taking the solution for granted. People often don’t properly contextualize beliefs that they themselves didn’t generate: it feels to them like things which are now understood by Science, like rocks and stars and brains, have always been that way, instead of having been a mystery to the human species for the many millennia until they suddenly weren’t anymore.

There is a more abstract objection to the idea of reinventing old discoveries, coming from a less efficiency-focused mental place. The idea seems to be that, if somebody already knows, the problem is for some reason no longer interesting. It’s the position taken by everyone who is enraptured by the breaking new scientific controversies, but is not the slightest bit interested in the proved-correct equations of General Relativity.

But in my book, it doesn’t much matter what somebody knows, if I don’t. When I was young, I wanted to know how my body worked. Why did my hands move when I willed them to, but a glass of water wouldn’t slide across the table to me with a similar mental effort? Why did eating a whole bunch of candy make me feel ill, but eating a whole bunch of salad didn’t? I didn’t know, and I wanted to know; if you had told young-me that somebody knew, I would have replied “okay, can they tell me, please?”.

So far as I was, and am, concerned, if somebody else happens to know the answer to a question, that doesn’t cheapen the discovery for me. In fact, it wouldn’t make any sense for me to have a term in my utility function for being the first person in the universe to make a discovery, because for all I know, super-advanced aliens on the other side of the galaxy have already discovered everything I could possibly want to learn. If my choices are between never taking joy in a discovery, because somewhere else, someone else might already know the answer, and having fun, I’ll pick the fun.

It might be impossible for me to do anything about this until I (or somebody else) create a feasible solution to the imminent mortality problem, but once it happens, you can bet you’ll find me in a remote field, trying to find the optimal way to rub two sticks together.

The Problem With Statements of the Form “X is Y”

Cats are mammals.

Gravity is the curvature of spacetime.

Both of these sentences are true. And they both use conjugations of the verb “to be”. But the usage of that verb is very different.

The first sentence is true in the sense that we as humans have defined a category called “mammals”, members of which have certain characteristics, and cats have the required characteristics to put them into that category.

The second sentence is true in the sense that “the curvature of spacetime” describes a thing that exists in the universe, to which we have given the name “gravity”. It’s a literal definition, not a taxonomical categorization.

The problem with these two different uses of “to be” is that people mistake the one for the other. They think that when scientists reclassified Pluto as a dwarf planet, it was because they’d learned something new about Pluto, as opposed to having simply revised the definition of a “planet”. But this is incorrect. A planet is whatever humans say a planet is, because the entire purpose of the category is to allow us to talk about “planets” without having to say “celestial bodies that have assumed a roughly spherical shape, orbit a star, and have cleared the space around their orbits” every time.

This isn’t to say that categories are meaningless, or can be defined in any way we please. Because a category is a convenient way of referring to a set of traits, those traits have to be common enough as a complete set to be worth referring to. We shouldn’t define a category to mean “a person with black hair and green eyes”, because those two traits aren’t any more likely to occur together than any other hair/eye color combination, and that set of traits doesn’t imply anything else useful about the person.

It is to say, however, that a category is not a fact about the universe, and that a category can and should be changed to better suit our purposes. Take, for example, the categories “male” and “female”. Most of us are accustomed to using these categories to imply a common combination of gender presentation, chromosomes, and sex characteristics. However, a lot of people have recently been changing these categories. In doing so, they aren’t violating any universal law that says “all masculine-presenting humans must have XY chromosomes”, because there is no such universal law. They are redefining a category that was defined in the first place by English-speaking humans and can be redefined by those same humans.

(Another recent change to categories: birds are reptiles.)

This same problem occurs the other way around: people mistake observations about the universe for human-made categories. I learned to say the phrase “gravity is the curvature of spacetime” when I was 9. It successfully impressed a lot of grownups, but the image in my head was a stretched-out tarp with a ball in the middle. It took nearly a decade for me to actually understand enough physics and math to grasp its real meaning. And when I did, my thought was basically “oh shit, gravity literally is the curvature of spacetime”.

When we talk about facts we understand, we’re not saying riddles or passwords. If somebody who doesn’t have the relevant background knowledge to understand us overhears, they might not realize that what we say is a literal observable fact, which they too can observe if they know where to look, but that doesn’t change it. The unfortunate reality is simply that the conversion from thought to language is not lossless. And until we invent telepathy, it probably never will be. Still, simply knowing the distinction between these two definitions of “X is Y” type sentences was helpful to me, and so I’m passing it on.

Ditch Pros and Cons: Use a Utility Function

If you’ve ever met me in person, you know I talk a lot about utility. My close friends are used to answering questions like “does this have net positive utility to you?” and “is that a strongly weighted component of your utility function?”. Every time I make a decision – what to do with my evening, or what to do with my life – I think about it in terms of utility.

I didn’t always do this, but I’ve adopted this way of thinking because it forces me to clarify everything that’s going on in my head and weight all my thoughts appropriately before making a decision. When I decide things this way, I genuinely feel like I’ve made the best possible choice I could, given everything I knew at the time.

What on earth is utility?

Utility is just a fancy word for “value”. If you enjoy chocolate ice cream, then eating chocolate ice cream has positive utility. If you don’t enjoy carrot cake, then eating carrot cake has negative utility.

One action can have multiple kinds of utility. You can add together all the utility types to get the action’s net utility. For example, if I assign positive utility to eating ice cream but a negative utility to gaining weight, there will be a certain optimal point where I eat as much ice cream as I can without gaining weight. Maybe, if I assign eating ice cream +5 utility, not gaining weight +5 utility, and exercising -5 utility, then it would make sense for me to hit the gym more often so that I can eat more ice cream without gaining weight.

The set of utilities I assign to all outcomes also tells me the optimal possible outcomes, with the highest net utility. In this example, that would be either modifying ice cream so that it doesn’t make me gain weight, or modifying my body’s energy processing system to get it to process ice cream without storing any of it as fat.

Having numbers that are consistent is helpful sometimes, but isn’t strictly necessary. When I need to make quick, relatively straightforward decisions, I typically just make up some utility numbers. Utility calculations in a small isolated system are basically matters of ratios: it doesn’t matter exactly how much utility I assign to something, but if outcome X has 5 more utility points than outcome Y, X is preferable.

Forcing yourself to make up numbers and compare them to each other reveals what you care about. If you initially thought you didn’t care much about something, but then realize that if you calculated net utility with a low number assigned to that thing, you’d be unsatisfied with the result, then you care more than you thought you did.

It might be somewhat unclear, with my super-simple examples so far, what you can assign utility to. So, here are some examples of things that I assign positive utility to:

  • Reading good books
  • Doing new things
  • Increasing utility according to the utility functions of people I care about
  • Building neat software
  • Drawing and painting
  • Writing stories and blog posts
  • Improving/maintaining my mental and physical health
  • Having interesting conversations
  • Improving the quality of life of all sentient beings
  • Running
  • Riding my bike
  • Taking walks with my girlfriend
  • Eating ice cream

If you enjoy doing it, if you think you should do it, if it makes you happy, if it’s a goal you have for some reason, or anything else like that, you assign it some amount of positive utility.

If you’d like to figure out how much relative utility you assign to different options, compare them: if I had to either give up on improving the quality of life for all sentient beings, or give up on ice cream, the ice cream has gotta go.

You can even assign positive utility to things you don’t end up doing. That’s because the net utility, after accounting for circumstances or mutually exclusive alternate actions. Knowing that you would do something, barring XYZ condition, is a useful thing to know in order to dissect your own thoughts, feelings, goals, and motivations. The converse is true, too: you can assign negative utility to things you end up doing anyway, because the net utility is positive.

So if that’s utility, what’s a utility function?

A utility function is a comprehensive set of everything that an agent assigns any utility to. “An agent” is anything capable of making decisions: a human, an AI, a sentient nonhuman alien, etc. Your utility function is the set of everything you care about.

The inputs to a utility function are quantities of certain outcomes, each of which are multiplied by their respectively-assigned utility value and then added together to get the total expected utility of a given course of action. In an equation, this is:

Ax + By + Cz + ...

Where A, B, C, and so on are individual facets of outcomes, and x, y, z, and so on are utilities.

Say I’m with my sister and we’re going to get food. I personally assign a strong net positive to getting burgers and a weak net negative for anything else. I also assign a positive utility to making my sister happy, regardless of where we go for food. If she has a strong net negative for getting burgers, and a weak net positive for sushi, I can evaluate that situation in my utility function and decide that my desire to make her happy overpowers the weak negative I have for anything besides burgers, so we go get sushi.

When evaluating more complex situations (such as moving to a job across the country, where positives include career advancement and increased income, and negatives include having to leave your home and make new friends), modeling your own utility function is an excellent way to parse out all the feelings that come from a choice like that. It’s better than a simple list of pros and cons because you have (numeric, if you like) weights for all the relevant actions.

How to use your utility function

I don’t keep my entire utility function in my head at one time. I’ve never even written it down. But I make sure I understand large swaths of it, compartmentalized to situations I often find myself in. However, if you decide to actually write down your utility values, proceed to make them all consistent, and actually calculate utility when you make decisions, there’s nothing stopping you.

In terms of the optimal way to think about utility calculations, I have one piece of advice. If you come out of a utility calculation thinking “gotcha, I can do this”, “alright, this seems reasonable”, or even “ugh, okay, I don’t like it but this is the best option”, then that’s good. That’s the utility function doing its job. But, if you come out of one thinking “hmmm… I guess, but what about XYZ contingency? I really don’t want to do ABC…”, or otherwise lingering on the point of decision, then you’ve forgotten something.

Go back and ask “what’s wrong with the ‘optimal’ outcome?”. It might be something you don’t want to admit to yourself, but you don’t gain anything by having an inaccurate perception of your own utility function. Remember that, in absence of a verbal reason, “I don’t wanna” is still a perfectly valid justification for assigning negative utility to an action or outcome. In order for this process to work, you need to parse out your desires/feelings/goals from your actions, without beating yourself up for it. Your utility function already is what it is, and owning up to it doesn’t make it worse.

Once you have a pretty good handle on your own utility function, you can go ahead and mentally model other peoples’. Humans are calculating utility all the time in the form of preferences and vague intuitions, so even if other people don’t know their utility functions, you can learn them by a combination of watching their actions and listening to their words.

The discrepancy between those two, by the way, is indicative of one of two things: either the person is choosing an action with suboptimal utility, or they don’t actually assign utility to the things they say they do aloud (perhaps for social reasons). You can point out this discrepancy politely, and perhaps help them to make better decisions in the future.

Once you begin to use utility functions for both yourself and others, you might be surprised at how much easier it is to make decisions. When considering possible courses of action for yourself, you’ll be able to choose the best option and know it was the best. And, in a group, having an accurate model of other peoples’ utility functions can let you account for their preferences, perhaps even better than they themselves do.

How to Be Happy

Growing up, my mom used to tell my siblings and I that when we were upset and didn’t want to be, we could choose to be happy instead. The whole concept seemed ridiculous to me. “I can’t just flip my emotions on and off like a light switch,” I remember telling her.

But the problem was, she was right. It’s entirely possible to “flip your emotions on and off like a light switch”. There’s a lot of research backing up that statement—not surprising, my mom graduated with a Master’s in psychology, I should have known she didn’t pull this idea out of nowhere. Further, though it took me longer than I would care to admit, I did personally realize the wisdom in her two-word advice, “choose happy”.

Many experiments show that if you smile, you’ll feel happier. It’s not even entirely about the conscious decision to feel happy – merely moving your facial muscles or even forcing a smile using chopsticks can do the trick. Your brain just has an association between smiles and happiness and so smiling can make you happy.

So the research says. But I doubted it. For years, until I realized the truth of it independently. Today, I’m going to dissect the reasons I doubted it, because I feel many people probably have the same doubts when reading articles like this one.

I had two reasons to doubt “choose happy”. The first was that I was afraid people would look at me weird if I went from crying to laughing in the span of less than two minutes. The unaltered procession of human emotions is a slow ebb and flow, and a drastic change would make people ask uncomfortable questions.

They probably would have done that. But I wish someone had told me that there are things much more important in life than seeming strange. Spending a majority of my time feeling depressed and anxious for no reason was dramatically worse than it would have been to have some people think I was odd. I should have weighed the pros and cons of feeling the emotion versus letting it go.

The second reason I doubted the wisdom of “choose happy” was that I thought all emotions were important. I thought that they were always there for a reason, even if I couldn’t find what that reason was. It was a gradual realization that led me to the simple fact that some emotions don’t make sense – they’re the result of hormonal imbalances, meaningless stressors, mental overstimulation, and many other things which don’t need to be dwelled on.

Nowadays, I think about emotions in the context of net utility. Is feeling this emotion useful to me? If I’m feeling embarrassed about a stupid mistake, that feeling can be useful, to prompt me to fix the mistake immediately. But after I’ve done everything I can to fix the mistake, including making the appropriate social reparations, I can let the emotion go, because it’s served its purpose. Continuing to feel embarrassed even when I can no longer do anything about the mistake, including learn from it, is pointless.

And if the emotion didn’t have any purpose to begin with – say, if I’m feeling angry because I’ve had a long difficult day at work, which is not even slightly connected to any particular problem that can be solved – I can analyze the cause, decide it’s pointless, and let go of the emotion.

How do you let go of emotions? After your brain stops intuitively holding on because it thinks they’re important, or that it would be weird to let go, it’s typically as simple as focusing on something else. If just passively thinking about something else doesn’t completely fix it, try smiling, putting on a fun or silly song, deliberately focusing on happy thoughts, or even closing your eyes and imagining a pleasant location to hang out for a while. (I’m deliberately giving advice that doesn’t require getting up, because I personally don’t like advice that says “get up! stretch! jog! sweat!” – it does genuinely work, but it’s always delivered in a very pushy way. That being said, if you haven’t already heard this advice from a hundred thousand people, being outside and/or exercising does in fact make you healthier and happier, so try it if you feel inclined.)

So the list of question to ask when you feel any emotion is:

  1. What emotion is it? Is that really what I’m feeling? Emotions are frequently very transparent, but they can become tangled. Further, some emotions can mask others: a lot of men have a tendency to express anger when they’re truly sad, for example. If your emotions are unclear, sort them out.
  2. What probably caused this emotion? Go over salient events in your mind and find the proximate cause. It doesn’t have to be anything major and it frequently isn’t. You’re looking for a cause, not a good reason.
  3. Does this emotion have net positive utility? Feeling negative emotions has inherent negative utility, but that may be outweighed by the positive utility of the action it makes you take: learning from a mistake, apologizing for a misdeed, fixing an internal or external problem, etc. Figure out if the emotion is prompting you to do anything useful, and if it isn’t, if you really need to keep it. Compute the net utility.
  4. An important note about these utility evaluations: A common trap I’ve seen many people fall into is where they keep a negative emotion around because they believe it prompts them to do something good which, in fact, they would do anyway. In particular, a lot of high-achievers end up with the misconception that being miserable is what prompts them to achieve things, when in fact, they would achieve more if they were happier. Therefore, strongly doubt any utility evaluation that leads you to the belief that you need to be miserable all the time in order to get things done.
  5. If you determined that the emotion has net positive utility, keep it around, but only as long as it continues to be useful. As soon as you’ve done everything useful that the emotion was prompting you to do, throw it away. There is no reason to be miserable longer than necessary.
  6. If you determined that the emotion has net negative utility, toss it immediately, using any of the tricks described above.

A final note about the utility of positive emotions: feeling good is a good thing. I’ve seen people be happy but wonder whether they really should be feeling happy. You can dissect the emotion and what actions it makes you take to figure this out, but don’t decide you need to be unhappy because it’s uncommon to see sane adults who visibly care about anything. Emotions are good to keep if they’re useful, and being happy uniformly makes your life better, so ceteris paribus, happiness is useful, and therefore, happiness is almost always good to keep.

In conclusion:
Choose Happy.

Places, Past and Future

We met in Baltimore
when the hot lights of the dance floor drove us out to the gardens
before the pouring of the rain drove us back in.

We got engaged in Pittsburgh
under the warm yellow glow of artificial lamplight
and I handed him the ring I’d bought with less ceremony than I’d like
though he seemed to love it anyway.

We’ll get married in San Francisco
surrounded by the warm California sun
by new and old friends
and by possibilities for our future spent together forever.

We’ll grow old among the stars
with the distant descendants of humanity at our side
accomplishing feats and forging friendships we can’t even dream of today.

And we’ll die
if in fact we must die
after impossible problems have been solved
after incomprehensible battles have been fought
after amazing spoils have been wrought:
we’ll die knowing that whatever else has come to pass
humanity has won.

Why Rationality?

I’ve identified as a rationalist for about five years now. The dictionary definitions are a bit off from what I mean, so here’s my definition.

Epistemic rationality: believing, and updating on evidence, so as to systematically improve the correspondence between your map and the territory.  The art of obtaining beliefs that correspond to reality as closely as possible.  This correspondence is commonly termed “truth” or “accuracy”, and we’re happy to call it that.

Instrumental rationality: achieving your values.  Not necessarily “your values” in the sense of being selfish values or unshared values: “your values” means anything you care about.  The art of choosing actions that steer the future toward outcomes ranked higher in your preferences.  On LW we sometimes refer to this as “winning”.

Eliezer Yudkowsky, “What Do We Mean By ‘Rationality’?”, LessWrong

Of course, these two definitions are really subsets of the same general concept, and they intertwine considerably. It’s somewhat difficult to achieve your values without believing true things, and similarly, it’s difficult (for a human, at least) to search for truth in absence of wanting to actually do anything with it. Still, it’s useful to distinguish the two subsets, since it helps to distinguish the clusters in concept-space.

So if that’s what I mean by rationality, then why am I a rationalist? Because I like believing true things and achieving my values. The better question here would be “why is everyone not a rationalist?”, and the answer is that, if it was both easy to do and widely known about, I think everyone would be.

Answering why it isn’t well-known is more complicated than answering why it isn’t easy, so, here are a handful of the reasons for the latter. (Written in the first person, because identifying as a rationalist doesn’t make me magically exempt from any of these things, it just means I know what they are and I do my best to fix them.)

  • I’m running on corrupted hardware. Looking at any list of cognitive biases will confirm this. And since I’m not a self-improving agent—I can’t reach into my brain and rearrange my neurons; I can’t rewrite my source code—I can only really make surface-level fixes to these extremely fundamental bugs. This is both difficult and frustrating, and to some extent scary, because it’s incredibly easy to break things irreparably if you go messing around without knowing what you’re doing, and you would be the thing you’re breaking.
  • I’m running on severely limited computing power. “One of the single greatest puzzles about the human brain,” Eliezer Yudkowsky wrote, “is how the damn thing works at all when most neurons fire 10-20 times per second, or 200Hz tops. […] Can you imagine having to program using 100Hz CPUs, no matter how many of them you had?  You’d also need a hundred billion processors just to get anything done in realtime. If you did need to write realtime programs for a hundred billion 100Hz processors, one trick you’d use as heavily as possible is caching. That’s when you store the results of previous operations and look them up next time, instead of recomputing them from scratch. […] It’s a good guess that the actual majority of human cognition consists of cache lookups.” Since most of my thoughts are cached, when I get new information, I need to resist my brain’s tendency to rely on those cached thoughts (which can end up in my head by accident and come from anywhere), and actually recompute my beliefs from scratch. Else, I end up with a lot of junk.
  • I can’t see the consequences of the things I believe. Now, on some level being able to do this (with infinite computing power) would be a superpower: in that circumstance all you’d need is a solid grasp of quantum physics and the rest would just follow from there. But humans don’t just lack the computing power; we can believe, or at least feel like we believe, two inherently contradictory things. This concept is, in psychology, called “cognitive dissonance”.
  • As a smart human starting from irrationality, knowing more information can easily hurt me. Smart humans naturally become very good at clever arguing—arguing for a predetermined position with propositions convoluted enough to confuse and confound any human arguer, even one who is right—and can thus use their intelligence to defeat itself with great efficiency. They argue against the truth convincingly, and can still feel like they’re winning while running away from the goal at top speed. Therefore, in any argument, I have to dissect my own position just as carefully, if not more carefully, than I dissect those of my opponents. Otherwise, I come away more secure in my potentially-faulty beliefs, and more able to argue those beliefs against the truth.

This is a short and incomplete list, of some of the problems that are easiest to explain. It’s by no means the entire list, or the list which would lend the most emotional weight to the statement “it’s incredibly difficult to believe true things”. But I do hope that it shed at least a little light on the problem.

If rationality is really so difficult, then, why bother?

In my case, I say “because my goal is important enough to be worth the hassle”. In general, I think that if you have a goal that’s worth spending thirty years on, that goal is also worth trying to be as rational as humanly possible about. However, I’d go a step further. Even if the goal is worth spending a few years or even months on, it’s still worth being rational about, because not being rational about it won’t just waste those years or months; it may waste your whole career.

Why? Because the universe rarely arrives at your doorstep to speak in grave tones, “this is an Important Decision, make it Wisely”. Instead, small decisions build to larger ones, and if those small decisions are made irrationally, you may never get the chance to make a big mistake; the small ones may have already sealed your doom. Here’s a personal example.

From a very young age, I wanted to go to Stanford. I learned that my parents had met there when I was about six, and I decided that I was going to go too. Like most decisions made by six-year-olds, this wasn’t based on any meaningful intelligence, let alone the full cost-benefit analysis that such a major life decision should have required. But I was young, and I let myself believe the very convenient thought that following the standard path would work for me. This was not, itself, the problem. The problem was that I kept on thinking this simplified six-year-old thought well into my young adulthood.

As I grew up, I piled all sorts of convincing arguments around that immature thought, rationalizing reasons I didn’t actually have to do anything difficult and change my beliefs. I would make all sorts of great connections with smart interesting people at Stanford, I thought, as if I couldn’t do the same in the workforce. I would get a prestigious degree that would open up many doors, I thought, as if working for Google isn’t just as prestigious but will pay you for the trouble. It will be worth the investment, the cached thoughts of society thought for me, and I didn’t question them.

I continued to fail at questioning them every year after, until the beginning of my senior year. At that point, I was pretty sick of school, so this wasn’t rationality, but a motivated search. But it was a search nonetheless, and I did reject the cached thoughts which I’d built up in my head for so long, and as I took the first step outside my bubble of predetermined cognition, I instantly saw a good number of arguments against attending Stanford. I realized that it had a huge opportunity cost, in both time and money. Four years and hundreds of thousands of dollars should not have been parted with that lightly.

And yet, even after I realized this, I was not done. It would have been incredibly easy to reject the conclusion I’d made because I didn’t want all that work to have been a waste. I was so close: I had a high SAT, I’d gotten good scores on 6 AP tests, including the only two computer science APs (the area I’d been intending to major in), and I’d gotten National Merit Commended Scholar status. All that would have been left was to complete my application, which I’m moderately confident I would have done well on, since I’m a good writer.

That bitterness could have cost me my life. Not in the sense that I would die for it immediately, but in the sense that everyone is dying for anything they spend significant time on, because everyone is dying. And it was here that rationality was my saving grace. I knew about the sunk cost fallacy. I knew that at this point I should scream “OOPS” and give up. I knew that at this point I should lose.

I bit my tongue, and lost.

I don’t know where I would end up if I hadn’t been able to lose here. The optimistic estimate is that I would have wasted four years, but gotten some form of financial aid or scholarship such that the financial cost was lower, and further, that in the process of attending college, I wouldn’t gain any more bad habits, I wouldn’t go stir-crazy from the practical inapplicability of the material (this was most of what had frustrated me about school before), and I would come out the other end with a degree but not too much debt and a non-zero number of gained skills and connections. That’s a very optimistic estimate, though, as you can probably tell given the way I wrote out the details. (Writing out all the details that make the optimistic scenario implausible is one of my favorite ways of combatting the planning fallacy.) There are a lot more pessimistic estimates, and it’s much more likely that one of those would happen.

Just by looking at the decision itself, you wouldn’t think of it as a particularly major one. Go to college, don’t go to college. How bad could it be, you may be tempted to ask. And my answer is, very bad. The universe is not fair. It’s not necessarily going to create a big cause for a big event: World War I was caused by some dude having a pity sandwich. Just because you feel like you’re making a minor life choice doesn’t mean you are, and just because you feel like you should be allowed to make an irrational choice just this once doesn’t mean the universe isn’t allowed to kill you anyway.

I don’t mean to make this excessively dramatic. It’s possible that being irrational here wouldn’t have messed me up. I don’t know, I didn’t live that outcome. But I highly doubt that this was the only opportunity I’ll get to be stupid. Actually, given my goals, I think it’s likely I’ll get a lot more, and that the next ones will have much higher stakes. In the near future, I can see people—possibly including me—making decisions where being stupid sounds like “oops” followed by the dull thuds of seven billion bodies hitting the floor.

This is genuinely the direction the future is headed. We are becoming more and more able to craft our destines, but we are flawed architects, and we must double and triple check our work, else the whole world collapses around us like a house on a poor foundation. If that scares you, irrationality should scare you. It sure terrifies the fuck out of me.

Language: A Cluster Analysis of Reality

Cluster analysis is the process of quantitatively grouping data in such a way that observations in the same group are more similar to each other than to those in other groups. This image should clear it up.

Whenever you do a cluster analysis, you do it on a specific set of variables: for example, I could cluster a set of customers against the two variables of satisfaction and brand loyalty. In that analysis, I might identify four clusters: (loyalty:high, satisfaction:low), (loyalty:low, satisfaction:low), (loyalty:high, satisfaction:high), and (loyalty:low, satisfaction:high). I might then label these four clusters to identify their characteristics for easy reference: “supporters”, “alienated”, “fans” and “roamers”, respectively.

What does that have to do with language?

Let’s take a word, “human”. If I define “human” as “featherless biped”, I’m effectively doing three things. One, I’m clustering an n-dimensional “reality-space”, which contains all the things in the universe graphed according to their properties, against the two variables ‘feathered’ and ‘bipedal’. Two, I’m pointing to the cluster of things which are (feathered:false, bipedal:true). Three, I’m labeling that cluster “human”.

This, the Aristotelian definition of “human”, isn’t very specific. It’s only clustering reality-space on two variables, so it ends up including some things that shouldn’t actually belong in the cluster, like apes and plucked chickens. Still, it’s good enough for most practical purposes, and assuming there aren’t any apes or plucked chickens around, it’ll help you to identify humans as separate from other things, like houses, vases, sandwiches, cats, colors, and mathematical theorems.

If we wanted to be more specific with our “human” definition, we could add a few more dimensions to our cluster analysis—add a few more attributes to our definition—and remove those outliers. For example, we might define “human” as “featherless bipedal mammals with red blood and 23 pairs of chromosomes, who reproduce sexually and use syntactical combinatorial language”. Now, we’re clustering reality-space against seven dimensions, instead of just two, and we get a more accurate analysis.

Despite this, we really can’t create a complete list of all the things that most real categories have in common. Our generalizations are leaky in some way, around the edges: our analyses aren’t perfect. (This is absolutely the case with every other cluster analysis, too.) There are always observations at the edges that might be in any number of clusters. Take a look at the graph above in this post. Those blue points at the top left edge, should they really be blue, or red or green instead? Are there really three clusters, or would it be more useful to say there are two, or four, or seven?

We make these decisions when we define words, too. Deciding which cluster to place an observation happens all the time with colors: is it red or orange, blue or green? Splitting one cluster into many happens when we need to split a word in order to convey more specific meaning: for example, “person” trisects into “human”, “alien”, and “AI”. Maybe you could split the “person” cluster even further than that. On the other end, you combine two categories into one when sub-cluster distinctions don’t matter for a certain purpose. The base-level category “table” substitutes more specific terms like “dining table” and “kotatsu” when the specifics don’t matter.

You can do a cluster analysis objectively wrong. There is math, and if the math says you’re wrong, you’re wrong. If your WCSS is so high that you have a cluster that you can’t label more distinctly than “everything else”, or if it’s so low you’ve segregated your clusters beyond the point of usefulness, then you’ve done it wrong.

Many people think “you can define a word any way you like”, but this doesn’t make sense. Words are cluster analyses of reality-space, and if cluster analyses can be wrong, words can also be wrong.


This post is a summary of / is based on Eliezer Yudkowsky’s essay sequence, “A Human’s Guide to Words“.

Dating: A Rational Approach

About four years ago, I decided I wanted to find a life partner. Primarily because I was socially oblivious and didn’t know any different, I took an analytical approach to do this—as we all do with other important areas of our lives. The entire process took me a matter of weeks and I have since been in a committed relationship for four years.

For a long time, primarily because of the fact that my method was so unorthodox and unheard-of, I largely imagined that the normal way of doing things was the best method, and I was a lucky fluke. But a number of recent conversations and some reading has led me to consider that maybe I’m one of the only people who does this right.

Before I dive into this, let’s establish a key point: if you’re going to get married, it is absolutely the most important decision of your life. While your choice of career dictates how you spend a good portion of your life, who you marry dictates how you spend all of your life, because it dictates who you spend your life with. The best recipe for misery is a bad marriage, and the best recipe for greatness is either no marriage or a great one.

John T. Reed, author of Succeeding, wrote about both the great importance of marrying the right person, and the haphazard way that many people take to get there. He writes, “The divorce rate is about 50% in the U.S. The median duration of marriages is seven years—just enough time to have some kids and acquire property so that the divorce really screws things up. […] Why are so many people screwing up the most important decision of their lives? Look at how they go about it. I read a book once that said most Americans feel the correct way to meet your spouse is ‘chance proximity’. […]  ‘Some enchanted evening, you will meet a stranger across a crowded room.’ Ask an old maid or old bachelor why they never married and they often tell you that the right person never ‘came along’. ‘Came along’! You gotta be kidding me! People make more effort to buy the right used car!”

And he’s right! Why, of all our important life decisions, do we fudge this one?

My best guess at a reason is this: habit is a powerful force, and societal habit is even more so. This way of coming at meeting spouses began when communities were small enough that putting forth much of a deliberate effort was unnecessary, and it’s perpetuated itself into modern society on our collective force of habit.

Nowadays, though, there are so many more people, and so many more ways to meet those people, that a systematic approach to dating is in order. Read Succeeding for John Reed’s method; below I’ll detail my own.

First, create a list, as comprehensive as you want to make it, of every important trait you want in a partner. This is physical traits (i.e., a beard), personality traits (i.e., wanderlust), or anything else you can think of. Once you’ve made this list, rank-order it, from most to least important.

Now, make a similar list of everything you don’t want in a partner. Be specific, but feel free to be obvious – while “emotionally manipulative” is an obvious anti-want, it might still be useful to put it on the list. Once you’ve made this list, rank-order it.

After you’ve done both of these, now it’s time to do some market research. What kind of dating pools exist? While answering this question, be sure to keep in mind which of these you’ll be willing to utilize. If you live in the U.S. and lack the budget or inclination to travel, it might be out of the question to try to find a date at a convention in London. If you’re considering internet-based dating pools, make sure you think about whether you’re willing to be long-distance for an extended period. Make a list of some potential dating pools and rank-order the list by feasibility.

Now it’s time to merge all these lists. Figure out what kind of person you’re mostly going to find at each of these dating pools and compare that to your lists of wants and anti-wants. Re-rank your list of dating pools against these criteria, then compare your list of dating pools ranked by plausibility of candidates against your list ranked by feasibility. Whatever dating pool is ranked highest in both (feel free to bias your ranking toward whichever you think is more important for you), make plans to go there.

Let’s go through my own story as an example. My list of wants included someone who is sensitive, who is okay with going against the grain, and who could adapt to my hectic lifestyle. My list of anti-wants included someone who is overly pompous or self-centered. I was young and very broke, so my options for dating pools were financially limited, but I also didn’t mind distance (I’d never really been taught that it was supposed to be hard, so I didn’t think it would be; and now, after having “suffered” two years of distance, I maintain that view). Based on my specific desires, dislikes, and difficulties, I was able to put at the top of my dating pool priority list a convention in Baltimore that ran three days in August.

This process is not over once you arrive at your dating pool: aimless drifting is still not a good plan (though it’s a better plan here than it would be elsewhere). No, now were going to systematically look for possible candidates.

To start with, walk around and scan crowds, finding people you find physically attractive. This is easy to judge from a distance just by a look, so do that first. Second, walk up to some attractive people and have conversations.

Think of this like going to a used car dealership. First, you look around the lot to see which cars are aesthetically pleasing. Then, you go around to some of the ones you think look nice enough, and you sit in the driver’s seat. You can find out a lot about a car just by sitting behind the wheel – just like you can find out a lot about a person just by having a conversation. And just like you don’t need to take every car in the lot on a test drive, you also don’t need to take every candidate on a date.

This variety of speed-dating has the benefit that you don’t mess with anyone’s heart—theirs or yours. You simply have a list of traits to compare this person against, and all you’re doing is comparing.

An important thing to do as you continue conversing with people is to take notes from your conversations and update your lists accordingly. If you started with a list item saying you want to meet people who do X, but when you actually met several people who did X they didn’t seem appealing to you, modify the list! If initially you thought that people who did Y were unbearable, but you met some people who did Y and they actually were fine, modify the list! Make sure to also modify the priority order of things if necessary.

These lists are not set in stone. In fact, it would be silly to have your actual experience with real people take second place to what you dreamed up about what real people might be like.

Once you’ve gone up and talked to a bunch of attractive people (my benchmark was 25, but you can do more or less, depending on what you think will work for you), you’ll likely have a small set (2-3) of people who you like the best. Ask them on formal dates. Once you’ve spent a few hours alone with each person, you’ll almost certainly have a winner.

Pretty good, eh? The only thing we still need is to account for feelings. It’s all well and good to meet a person you think would be perfect, but you both need to fall for each other. How do you account for that? Very simply, actually. If you start to feel something good for them as you’re talking, keep talking. And, as you usually do when you date the conventional way, look for signs that they like you back. If you’d like to improve your chances, you can try to ask some or all of the questions on this list. Humans are hardwired to fall in love: it doesn’t take too much of a push.

Now, if by the end of your first venture into a dating pool, you don’t have a life partner yet, don’t worry! Just go on back to your lists and find your second choice for a dating pool, then rinse and repeat. It may also be possible that your criteria are too broad, or too narrow, or you were wrong about what kinds of people frequent what places. If so, don’t sweat, just go back and revise your lists with your new knowledge. Then go on back into the world and keep at it! Having a systematic approach will work so much better than just waiting for someone to “come along”, and it will feel better, too. You’re being way more productive!

Obviously, this is a very different approach than the conventional one. But if you step back and think logically about how people should go about making this choice, it’s a much more reasonable approach. I’m sure there will be people saying it’s “not romantic”, but approaches like these have resulted in lasting relationships: my father took a similarly systematic approach to dating and my parents have been together for thirty years; John Reed followed a similar approach and was married for much longer. You don’t need to take my four-year relationship as your only data point.

Furthermore, “romantic” should mean “spontaneous”, not “stumbled into”. Too often, people confuse the two. Romance doesn’t have to be about random chance.