<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:media="http://search.yahoo.com/mrss/"><channel><title><![CDATA[This Is My True Name]]></title><description><![CDATA[This Is My True Name]]></description><link>https://www.georgeyw.com/</link><image><url>https://www.georgeyw.com/favicon.png</url><title>This Is My True Name</title><link>https://www.georgeyw.com/</link></image><generator>Ghost 5.87</generator><lastBuildDate>Mon, 04 May 2026 11:08:59 GMT</lastBuildDate><atom:link href="https://www.georgeyw.com/rss/" rel="self" type="application/rss+xml"/><ttl>60</ttl><item><title><![CDATA[Winning doesn't have to look dignified]]></title><description><![CDATA[(sometimes)]]></description><link>https://www.georgeyw.com/undignified/</link><guid isPermaLink="false">68cb2334a72d5eb24577e442</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Thu, 18 Sep 2025 01:33:28 GMT</pubDate><content:encoded><![CDATA[<p>Years ago, I was in Vegas watching a Smash Ultimate match at EVO, the biggest fighting game tournament in the world. There was nothing special about this match except that it was being shown. During qualifiers, random player pairs in the bracket would be brought to the main stage, and their match would be projected onto a screen so the audience would have something to watch. Sometimes they would be professional players, but neither of these players made it anywhere near the grand prize.</p><p>By spectator standards, it was a terrible match. The Pac-Man player spent the entire game at a distance, throwing out projectile attacks, and the other player failed to find a way to approach. There were no flashy combos, no tense moments. We watched Pac-Man throw fruit for a while and then it was over. The crowd <em>hated</em> this, the Pac-Man player was booed the whole time. He still advanced to the next match.</p><hr><p>I think about food a lot. A nontrivial amount of my daily willpower points are spent on not snacking too much, not overeating, and not being distracted by food in between for long enough to be productive. Some days this is harder than others.</p><p>One of the tricks that I&apos;ve learned works on me is to prepare a snack for myself, then set it next to my keyboard. I don&apos;t even eat it! It just sits there! This is somehow enough to quiet my brain.</p><p>I discovered this on accident after grabbing a snack a few times and then later realizing I hadn&apos;t eaten any of it after I sat back down. It sometimes feels a bit lame. It works anyways.</p><hr><p>In 2016, Mark Zuckerberg sent an email to some of his exec team about Snapchat analytics:</p><blockquote>Whenever someone asks a question about Snapchat, the answer is usually that because their traffic is encrypted we have no analytics about them.<br><br>Given how quickly they&apos;re growing, it seems important to figure out a new way to get reliable analytics about them. Perhaps we need to do panels or write custom software. You should figure out how to do this.</blockquote><p>About a week later, Project Ghostbuster allowed Facebook to install a man-in-the-middle attack that would let them intercept otherwise encrypted user data. They got the user analytics they were looking for.</p><hr><p>We automatically fabricate and attach rules to a goal we&apos;re chasing that aren&apos;t actually part of the goal, &quot;dignity&quot; being a common one. This is a low agency move, you can just do undignified things.</p><p>That&apos;s not to say you should ignore your honor <em>all</em> the time &#x2013; I don&apos;t endorse Project Ghostbuster for example &#x2013; but sometimes winning looks like deciding that you don&apos;t give a shit. Notice when the rules you attach are fake and don&apos;t hobble yourself when they are.</p>]]></content:encoded></item><item><title><![CDATA[Navigating burnout]]></title><description><![CDATA[Halt and catch fire]]></description><link>https://www.georgeyw.com/burnout/</link><guid isPermaLink="false">68169011a72d5eb24577e3fc</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sat, 03 May 2025 21:58:12 GMT</pubDate><content:encoded><![CDATA[<p>Burnout. Burn out? Whatever, it sucks.&#xA0;</p><p>Burnout is a pretty confusing thing made harder by our naive reactions being things like &#x201C;just try harder&#x201D; or &#x201C;grit your teeth and push through&#x201D;, which usually happen to be exactly the wrong things to do. Burnout also isn&#x2019;t really just one thing, it&#x2019;s more like a collection of distinct problems that are clustered by similar symptoms.</p><p>Something something intro, research context, this is what I&#x2019;ve learned / synthesized blah blah blah. Read on!</p><h1 id="models-of-burnout">Models of burnout</h1><p>These are models of burnout that I&#x2019;ve found particularly useful, with the reminder that these are just models with all the caveats that that comes with.</p><h2 id="burnout-as-a-mental-injury">Burnout as a mental injury</h2><p>Researchers can be thought of as &#x201C;mental athletes&#x201D; who get &#x201C;mental injuries&#x201D; (such as burnout) the way physical athletes get physical injuries, and we should orient towards these mental injuries in the same way we orient towards physical injuries.</p><p>I think this is by far the most useful model, because it makes it easier to actually generate a degree of self-compassion that makes you stop slamming your face into the brick wall. It is much easier to be hard on yourself while burnt out that you&#x2019;re not producing &#x201C;enough&#x201D; legible output than it is to be hard on yourself with a twisted ankle that you aren&#x2019;t keeping up with your weekly cardio goals. The way in which running on a twisted ankle is self-evidently Bad is more obvious than the way in which pushing through yet another deadline is Bad.</p><p>I&#x2019;m increasingly convinced over time of how deeply this physical analogy runs. Success as an athlete is as much about managing injuries and recovery as it is about optimizing your peak performance. Managing burnout is a core research skill as much as managing recovery is a core athletic skill.</p><p>(For rehab of a physical injury, you have &#x201C;active recovery&#x201D;, where you aren&#x2019;t back to normal but it benefits the healing process to start using the injured part again (mobility, physical therapy, light workouts). There is probably some way to triangulate active recovery for burnout here, but my guess is most people&#x2019;s first attempt at doing so will still end up way too &#x201C;active&#x201D; to be recovery.)</p><h2 id="burnout-as-a-deficit-of-activation-energy">Burnout as a deficit of activation energy</h2><p>I like this description of <a href="https://www.lesswrong.com/posts/xtuk9wkuSP6H7CcE2/ayn-rand-s-model-of-living-money-and-an-upside-of-burnout?ref=georgeyw.com"><u>willpower as &#x201C;living money&#x201D;</u></a> and burnout as being &#x201C;bankrupt&#x201D;. In general, tasks have a certain activation energy to them, and we have some capacity (that varies across time) to cross activation energy thresholds. When we&#x2019;re high capacity (not burnt out), moderate to high activation energy thresholds don&#x2019;t feel hard to cross, but when we&#x2019;re low capacity (burnt out), they feel ~impossible.</p><p>You can track progress on burnout, both in severity and recovery, by approximating how hard different tasks feel now vs. in the past. Progress on recovery should feel like <em>making things feel easy</em>, not about <em>figuring out how to do things that feel hard</em>. Sometimes life calls and you do just need to buckle up and do hard stuff, but meeting that deadline through blood, sweat, and tears is <em>not</em> the same as recovering from burnout. To borrow the physical analogy, you&#x2019;re successfully recovering from a broken leg if walking feels easy / normal again, not if you&#x2019;re managing to walk but your leg remains painful and fucked up each time you take a step.</p><h1 id="sources-of-burnout">Sources of burnout</h1><p>Burnout has many mechanisms, and it can be kind of hard to pinpoint what the root cause is because sometimes they spill over into each other. For example, maybe you&#x2019;re just burnt out because your sleep has been garbage and you can fix that directly. But it could also be the case that your sleep is bad because you&#x2019;re anxious about some other root cause, and just fixing your sleep won&#x2019;t make the other problems go away (it&#x2019;ll probably help a lot anyways, but might instead be resistant to improvement). I don&#x2019;t really know what to say to guide you here except that you should introspect, and if that doesn&#x2019;t work, read <a href="https://www.lesswrong.com/w/noticing?ref=georgeyw.com"><u>other articles</u></a> that have more to say about it and&#x2026; keep trying until it does?</p><p>Possible mechanisms can broadly be physiological, psychological, or work-related. For work-related burnout, Emmett Shear identifies <a href="https://x.com/eshear/status/1561120325584109574?ref=georgeyw.com"><u>three sources</u></a>: &#x201C;permanent on-call&#x201D;, &#x201C;broken steering&#x201D;, and &#x201C;mission doubt&#x201D;. I&#x2019;ll add a fourth one, which is how &#x201C;heavy&#x201D; your work feels.</p><h2 id="physiological-psychological">Physiological + Psychological</h2><p>Physiologically: if one of [sleep, diet, exercise] isn&#x2019;t solid, generally try to fix that first and see if the problem goes away.</p><p>Psychologically: are you like, generally happy? Do you get enough social interaction? Are there major sources of psychological stress (acute or not)? One of my minor spells of burnout turned out to be solved by just actually leaving the research dungeon and spending more time with friends.</p><h2 id="broken-steering-responsiveness">Broken steering / responsiveness</h2><p>People like being able to poke at the world and have the world react (see also <a href="https://sashachapin.substack.com/p/what-the-humans-like-is-responsiveness?ref=georgeyw.com"><u>this blog post</u></a> on responsiveness). It feels bad when this doesn&#x2019;t happen. It feels really bad when this doesn&#x2019;t happen for a long time. This is pretty closely related to the &#x201C;living money&#x201D; / willpower model; none of your bids to spend energy are sufficiently rewarded with some kind of response.&#xA0;</p><p>I think this is one of the hardest sources to deal with in research, because the steady state of challenging research work is to be stuck and muddling through things. That just kind of sucks! To some degree I think you just have to learn to be ok with this (but there is also a lot that you can do to improve the responsiveness of your work, and if it&#x2019;s really not responsive that can be a sign that you&#x2019;re chasing the wrong things or just pressing buttons hoping something happens).</p><p>As a sidebar, a friend was recently talking to me about the importance of being able to be bored. Several years ago, I gave TikTok a try. I lasted about a month before I became nearly incapable of focusing on work because it blew up my mental baseline for what a reasonable reward loop with the world was. I quickly returned to my previous baseline after deleting TikTok; I think this probably implies a <a href="https://www.lesswrong.com/posts/z8usYeKX7dtTWsEnk/more-dakka?ref=georgeyw.com"><u>More Dakka</u></a> thing exists here (like a digital detox), but I haven&#x2019;t tried it yet.</p><h2 id="permanent-on-call">Permanent on-call</h2><p>I&#x2019;m actually not exactly sure how to map this onto research; research often has stereotypically difficult work-life balance, but some of the ways this happens can be bad, while others are fine or at least instrumental. Crunch time before a research deadline is a form of this, as is being the single person that knows how to do X and having lots of people that need you to do X for them all the time.&#xA0;</p><p>From another angle, it might be when a research problem is really consuming you and it just follows you around everywhere. This is actually kind of ideal in some ways; a lot of conceptual research progress happens as background processing, but in order for background processing to happen, it has to occupy a lot of your background thoughts.</p><h2 id="mission-doubt">Mission doubt</h2><p>I would guess that this one is relatively less problematic in research work? It&#x2019;s probably less common for people to be forced into certain research roles through life circumstances than it is for other types of jobs, so it&#x2019;s more likely that you have some kind of investment in the mission you&#x2019;ve chosen. There&#x2019;s also comparatively more freedom to wiggle your work in a direction that you care about.&#xA0;</p><p>One specific way in which I could see this manifesting in research is believing in some grand vision, but not really believing in how some sub-mission connects to it (sort of like broken steering but at a different scale). You might just keep working on the sub-mission due to inertia and not really having a real-feeling outside option. If this is the case, I suggest that you <a href="https://www.benkuhn.net/abyss/?ref=georgeyw.com"><u>stare into the abyss</u></a>.</p><p>My coworkers sometimes talk about having radical, irrational optimism in your work &#x2013; sometimes you have to channel a little crazy to convince yourself to do the hard, uncertain work. It&#x2019;s not good to <em>always</em> maintain that state, but if you can&#x2019;t muster it up even a little bit, that seems Bad.</p><p>Anyways, if this is your problem with your work, stop working on it! As a researcher you almost certainly have outside options, including <em>spending your energy finding outside options</em>.</p><h2 id="lightness-and-heaviness">Lightness and heaviness</h2><p>Lightness roughly points in the direction of how fun or playful or generally low-stakes the work feels. This is like that whole attachment thing of holding things lightly.&#xA0;</p><ul><li>Hackathons / game jams have typically felt very light to me, and my recollection is feeling pretty good afterwards despite them being extremely concentrated periods of effort. Here, the stakes are fairly low and the work (and its impacts) are well-scoped. There&#x2019;s an atmosphere of (sometimes type 2) fun.</li><li>I&#x2019;ve also felt this when doing supporting work for some project that feels valuable but isn&#x2019;t addressing the crux. The project will survive without me, and someone competent that I trust is on the ball.&#xA0;</li><li>A former mentor of mine would save a certain kind of toy problem for her students to give to them when they seemed stressed out by the core work of their thesis. The toy problems weren&#x2019;t meant to be publishable, they were just little pieces of mathematical candy to remind her students that math could still be light and fun.</li></ul><p>Heaviness roughly points in the direction of high-stakes.&#xA0;</p><ul><li>You need this to work in order to get or keep your job.&#xA0;</li><li>It feels like you&#x2019;re singularly responsible for the crux of the project and other people are counting on you.&#xA0;</li><li>Your self-worth depends on producing legible outputs.</li><li>If you have to champion your work to other people, it can feel like you have to believe in it enough for several people, and every outcome carries the weight of needing to justify further investment.</li></ul><p>These things produce significant extra mental fatigue / friction, even though it might not feel like that when you&#x2019;re at high capacity. There is always some irreducible degree of difficulty, but to use the physical analogy, it&#x2019;s like running with a weighted vest vs. not.</p><h1 id="early-warning-signs">Early warning signs</h1><p>In sports, people say something like &#x201C;listen to your body&#x201D; in order to avoid injuries. I didn&#x2019;t find this particularly helpful until I got injured a few times and actually developed a sense of the limits of my body. I think if you pay close attention when that happens, you actually can learn a fairly fine-grained sense of what it feels like to push something close to its limit and how bad an injury really is.</p><p>I think getting better at noticing burnout is like this. That sucks a little because it sort of means that you just need to experience some burnout in order to have any kind of training data, but it&#x2019;s definitely possible to have high sample efficiency (if there is a way to make people good at this without them ever burning out, please let me know). I suspect each person&#x2019;s warning signs will look a bit different depending both on their personality and the type of burnout they&#x2019;re facing. My most common warnings are gradual increases in revenge sleep procrastination, playing more video games, letting chores pile up more, working out less, and eating worse.&#xA0;</p><p>It&#x2019;s tricky to know the correct response because these signs don&#x2019;t always indicate a serious problem, and when they do, sometimes you can just grind a little harder (but see the next section for a warning about this) and other times you can&#x2019;t. This is a skill that you can empirically make progress on though: in grad school I burned out horribly and didn&#x2019;t do any substantive research for over a year, and these days I&#x2019;ve been able to intervene quickly and (mostly) prevent long term work disruptions. Nearly all of that progress factors through &#x201C;listening to my mind&#x201D; (see also: <a href="https://www.lesswrong.com/w/noticing?ref=georgeyw.com"><u>Noticing</u></a>).</p><h1 id="coping-mechanisms-and-solutions">Coping mechanisms and solutions</h1><p>My biggest warning here by far is to be clear-headed about which things are coping mechanisms and which are robust, long-term solutions. As a general rule, things that feel like &#x201C;action scaffolding&#x201D; that push you in the direction of making short-term progress tend to be coping mechanisms. If you just do the action scaffolding stuff, you might get lucky and outrun your burnout (if, for example, you were grinding out job applications and you finally get an offer), but you might also just get more burnt out until those things don&#x2019;t work anymore either. This is a <em>very dangerous</em> fire to play with; extra marginal burnout can have extremely outsized impacts. Necessary recovery time can grow surprisingly quickly, and I know of at least one extreme case where someone had to quit an entire area of research permanently.</p><p>A (non-exhaustive) list of examples in your action space, (very) coarsely ordered from cope-y to solution-y (don&#x2019;t read too much into the ordering though, that will vary from person to person too):&#xA0;</p><ul><li>Take a break (a short vacation, reduced hours, etc).</li><li>Body doubling / coworking; this doesn&#x2019;t work for everyone, but I find that this robustly lowers the activation energy costs and reduces distractions.</li><li>Try to manufacture some fun in your work. Maybe set aside some time to chase down a sidequest or organize a casual hackathon. Do things that feel easy and don&#x2019;t do things that feel hard, try to follow the gradients for a while.</li><li>If you&#x2019;re stressed about other people&#x2019;s perceptions or of letting other people down, talk to them about it. Sometimes bearing the secret accounts for a large portion of the weight, and it suddenly becomes manageable once you say it out loud to someone.</li><li>In the case of heaviness-from-self-worth, referring to past evidence that you&#x2019;re worthy. It felt like a huge weight had been lifted from my shoulders during my PhD once I&#x2019;d finally published a paper I was proud of, and I really milked that as a persistent reminder that I could be a competent researcher. Ideally that wouldn&#x2019;t be necessary, but it did turn out to be a fairly robust coping mechanism and buys me time to have a healthier attachment to my work. If you&#x2019;re a very young researcher, maybe you don&#x2019;t have papers, but surely you&#x2019;ve ever done something you were proud of in life, let yourself feel a little good about it.</li><li>Fix your feedback cycles in your work; rearrange things or figure out how to actually feel like the world squishes appropriately when you poke it.</li><li>Focus on process: if you&#x2019;re persistently <a href="https://terrytao.wordpress.com/career-advice/continually-aim-just-beyond-your-current-range/?ref=georgeyw.com"><u>learning and improving</u></a>, the power of compounding investment makes it inevitable that impressive results will eventually just fall out.</li><li>Just like, generally being kinder to yourself? Research is a domain where there is always an ~infinite amount of useful work that can be done, and it is also easy to be ~infinitely hard on yourself for not accomplishing more or for not being good enough. This is particularly hard when you&#x2019;re really burnt out and need to scale down your productivity, because it feels like you&#x2019;re going in the wrong direction (this is where I find the physical injury analogy most useful personally).</li></ul><p>These are just generic examples that tend to work across broad subsets of root causes for burnout, and there&#x2019;s a decent chance you have to do something more bespoke to solve your problems. In any case, lots of people have had severe burnout and made it through to the other side. It is empirically a tractable problem even if it doesn&#x2019;t typically feel like it from the inside.</p><hr><p><em>Thanks for discussion and feedback: Andy, Cecilia, Maggie, Maggie</em></p>]]></content:encoded></item><item><title><![CDATA[Should you be worried about H5N1?]]></title><description><![CDATA[Ca-caw!]]></description><link>https://www.georgeyw.com/should-you-be-worried-about-h5n1/</link><guid isPermaLink="false">675219a6a72d5eb24577e2d7</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Thu, 05 Dec 2024 21:25:23 GMT</pubDate><content:encoded><![CDATA[<p><em>Epistemic status: a few people without any particular expertise in epidemiology spent an afternoon in a coffee shop discussing and reading about H5N1, with a focus on how an individual should orient towards this (as opposed to say, the government). This is a write-up of what I took away from that exercise, written from my perspective.&#xA0;</em></p><p><em>Some ideas were generated in collaboration with Claude but generally spot checked. This post was also sanity-checked by a friend who works in epidemiology. I feel ok about the ideas presented but would not be surprised if someone with more expertise has a significantly different conclusion. Any mistakes are mine.</em></p><p>I went into this exercise with a prior of &#x201C;I&#x2019;ve been hearing about bird flu for years, and it&#x2019;s always been nothing, it&#x2019;s probably nothing again this time.&#x201D; The main upshot is that I walked away from the exercise thinking &#x201C;I don&#x2019;t know if this is going to be something or not.&#x201D; As far as updates go, that seems directionally bad. My current orientation towards this is something like &#x201C;watch and wait, and spend appropriately more effort on this if / when certain milestones happen (and also make some trades).&#x201D;</p><h1 id="what%E2%80%99s-different-this-year">What&#x2019;s different this year?</h1><p>One main thing that seems to generate a heightened level of ongoing risk is sustained infections in dairy cow populations. This is already bad because it&#x2019;s mammal-to-mammal transmission in a population that hasn&#x2019;t historically had problems with bird flu, and it would be worse if it becomes endemic in farmed cows. As long as it is sustaining infections in farmed cows, H5N1 has a convenient breeding ground for mutations that is:</p><ul><li>In a high density population</li><li>Year round (bird flu is seasonal among birds)</li><li>In regular contact with humans</li></ul><p>The last point gets a bonus with the human flu season coming up. If a dairy worker gets sick with both bird flu and human flu simultaneously, the two strains might share their genetic information and mutate into a pandemic-worthy strain.</p><p>Another thing being talked about is the teen in Canada who got sick from an unknown source. I suspect that at least some of the attention here is because scary things generate ad revenue, but there is also evidence that a wild strain of H5N1&#xA0;<a href="https://bsky.app/profile/scottehensley.bsky.social/post/3lb36uy5a7k25?ref=georgeyw.com"><u>has a mutation</u></a> that improves binding to human receptors.</p><p>These seem to represent two distinct threat vectors to me; the Canadian teen seems to be very sick, whereas dairy workers have generally had mild symptoms (although in the case of the 1918 Spanish flu (also avian), younger people had higher fatality rates). A jump from birds directly to humans would be mostly independent of whatever is happening in dairy farms, and vice versa.</p><p>I don&#x2019;t have a concrete prediction here, but I notice that <a href="https://polymarket.com/event/bird-flu-pandemic-before-august-2025/bird-flu-pandemic-before-august-2025?tid=1732824718160&amp;ref=georgeyw.com" rel="noreferrer">Polymarket being at 14%</a> for a pandemic by August 2025 doesn&#x2019;t feel wild to me. I would tentatively assign more probability to something coming from the cow situation than directly from birds.</p><h1 id="what-to-watch-for">What to watch for</h1><p>As I said above, my plan is to watch and wait, but what to watch for? There are a few key things IMO:</p><ul><li>A (sustained) jump to pigs. Pigs are traditionally considered a mixing vessel for flu viruses, since they are apparently susceptible to them and are able to host avian, human, and swine flus. This is how the 2009 H1N1 pandemic mutations happened. There has been&#xA0;<a href="https://www.aphis.usda.gov/livestock-poultry-disease/avian/avian-influenza/hpai-detections/hpai-confirmed-cases-livestock?ref=georgeyw.com"><u>one confirmed case in a pig</u></a>, but that was over a month ago now with no more news that I&apos;ve seen.</li><li>A mutation to airborne transmission among cows. Seems self-explanatory, indicates that the virus is undergoing significant mutations in the cow population. Also airborne bad.</li><li>Whether it continues to infect cows / becomes endemic (see previous section on why this creates elevated risk).</li><li>A lot more cases like the teen in Canada, with evidence that the suspected mutation that makes it easier to bind to human receptors is common in wild H5N1.</li><li>Of course the big one is human to human transmission, with each of the previous points being plausible stepping stones.</li></ul><p>Something speculative is watching how pharma companies are reacting. As I understand it, to some crude approximation pharma companies take a hits-based approach to investing in drugs. In the venerated tradition of pulling numbers out of my ass, if O($100M) is an appropriate amount to spend on pre-empting a pandemic, and they expect to make O($1B) on a hit, then this indicates an O(10%) belief that some kind of pandemic happens in the near future?&#xA0;</p><p>The US government has&#xA0;<a href="https://www.cidrap.umn.edu/avian-influenza-bird-flu/hhs-awards-moderna-176-million-develop-mrna-h5-avian-flu-vaccine?ref=georgeyw.com"><u>paid Moderna $176M</u></a> to work on H5N1 vaccines (which impacts the above speculation, since this de-risks the investment for Moderna),&#xA0;<a href="https://www.pfizer.com/news/announcements/pfizer-reiterates-commitment-pandemic-preparedness?ref=georgeyw.com"><u>Pfizer is working on it as well</u></a> (though I didn&#x2019;t find anything indicating a federal grant), and the government has also&#xA0;<a href="https://abcnews.go.com/Health/us-72m-vaccine-manufacturers-advance-bird-flu-shot/story?id=114502971&amp;ref=georgeyw.com"><u>paid $72M to manufacturers</u></a> to make existing vaccines ready to use.</p><h1 id="vaccine-response">Vaccine response</h1><p>The US government has a strategic reserve of bird flu vaccines. Given that pharma companies are also preemptively working on a vaccine, what could we expect in terms of immediate response and an eventual bespoke vaccine for the pandemic strain?</p><p>The strategic reserve of bird flu vaccines is small relative to the population and seems basically intended for emergency / essential workers. The above link about&#xA0;<a href="https://abcnews.go.com/Health/us-72m-vaccine-manufacturers-advance-bird-flu-shot/story?id=114502971&amp;ref=georgeyw.com"><u>paying $72M to manufacturers</u></a> mentions that this would double the strategic reserve from 5 million to 10 million doses, which is already far from the total US population before accounting for multiple doses per person. Furthermore, these vaccines are based on known strains and may only provide limited protection against a new pandemic strain.</p><p>So we probably aren&#x2019;t going to get an immediate bail out of a pandemic. How long would it take to develop and ramp up a vaccine for the full population? For reference, widespread availability of the covid vaccine took around a year, while H1N1 took around 6 months. Why was H1N1 so much faster? Basically we&#x2019;re really good at making flu vaccines because we make them every year and there are mature platforms for this, whereas the coronavirus vaccine used fairly novel technology (mRNA vaccines).&#xA0;</p><p>So we could plausibly move as fast as with H1N1, but here are some serially bound steps that need to happen to make a pandemic-strain-specific vaccine, so much faster than H1N1 sounds hard. Something like 4-6 months seems plausible to me, not accounting for any political friction.</p><h1 id="sociopolitical-response">Sociopolitical response</h1><p>We didn&#x2019;t spend much time thinking about the political response and I don&#x2019;t super feel qualified to shoot from the hip here, but some potentially important factors to note:</p><ul><li>Pandemic fatigue seems real, and it might be hard to coordinate effectively a second time.</li><li>AFAIK RFK is pretty anti-vax (and is the nominee for HHS) and vaccines are super politicized now. Not really sure how to game this out &#x2013; if a hypothetical pandemic was bad enough, would it moderate everyone&#x2019;s views?</li><li>Individual reactions might be modeled by a bimodal distribution? Seems plausible to me that some people overreact while others underreact.</li></ul><h1 id="how-bad-would-a-pandemic-actually-be">How bad would a pandemic actually be?</h1><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh7-rt.googleusercontent.com/docsz/AD_4nXeQ5pk71-eT4oVy3Pwq4O_hSaaL7We0Y4mIuV9rTdz9-O9DQwsazPwGSvtB-PqGYt8H6cAeEgl3smIYSWQvEspp5f00bRmVpcn1rwjNbMST7GNTaUAck1xujltqkT6hJKQbm96b6w?key=CnOVWdhjAFUzPLqK1NjT2yGD" class="kg-image" alt loading="lazy" width="961" height="746"><figcaption><span style="white-space: pre-wrap;">From </span><a href="https://en.wikipedia.org/wiki/Influenza_pandemic?ref=georgeyw.com#Influenza_pandemics"><span style="white-space: pre-wrap;">Wikipedia</span></a></figcaption></figure><p>Wikipedia has&#xA0;<a href="https://en.wikipedia.org/wiki/Influenza_pandemic?ref=georgeyw.com#Influenza_pandemics"><u>a page on past influenza pandemics</u></a>. I was somewhat surprised to learn that the seasonal flu already kills hundreds of thousands of people every year (!), and is in fact typically more lethal each year than 2009 H1N1 was. I don&#x2019;t really understand the HXNY labeling, but it seems like H1N1-like strains have caused three pandemics, each time less bad than the last, which makes sense if you no longer have a naive population.</p><p>Our base case might look more like H2N2 or H3N2 then. The case fatality rate for these is very low compared to the observed fatality rate in some H5N1 strains (~50%?), but the H5N1 strain that is infecting dairy cows and workers hasn&#x2019;t killed anyone yet and seems much more mild than a coin-flip death sentence (and an effective pandemic strain is unlikely to have a 50% fatality rate &#x2013; being too lethal causes a virus to burn out before spreading widely, in addition to motivating a stronger response). H2N2 and H3N2 happened before we figured out how to surveil these diseases before a jump to humans, so it&#x2019;s not clear whether these looked like H5N1 does today or if they were much milder even before the jump.&#xA0;</p><p>For another point of reference,&#xA0;<a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC9950018/?ref=georgeyw.com#:~:text=based%20on%2012%20studies%20between,(IQR)%20%3D%201.16%5D."><u>covid had an original reproduction number of around 3</u></a> and a case fatality rate of around 1% (on top of other &#x201C;lucky&#x201D; features, like being contagious during a long asymptomatic incubation period).</p><p>We also already have antivirals for the flu like Tamiflu, in contrast to covid where it took a long time to develop things like Paxlovid. There&#x2019;s no guarantee that it would be super effective, but so far it seems like it has some effect and is currently being used to treat human H5N1 cases.</p><p>My best guess is that we should expect an H5N1 pandemic to look much more like a (possibly more lethal) H2N2 or H3N2 pandemic than a covid pandemic, though we shouldn&#x2019;t rule out something worse. This would be bad, but not&#xA0;<em>incredibly</em> bad.&#xA0;</p><h1 id="what-else">What else?</h1><p>I&#x2019;d be interested in hearing more takes and especially places where I might have gotten something wrong or missed something important. My guess is that there are also trades that are more profitable than buying Yes shares on Polymarket if they hit, though I&apos;m not sure about publicly speculating about this.</p><hr><p><em>Thanks for ideas, discussions, feedback: Maggie, James, Linda</em></p>]]></content:encoded></item><item><title><![CDATA[Reflections on the Metastrategies Workshop]]></title><description><![CDATA[A weekend of trying to be better]]></description><link>https://www.georgeyw.com/reflections-on-the-metastrategies-workshop/</link><guid isPermaLink="false">67526671a72d5eb24577e314</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Fri, 25 Oct 2024 01:50:00 GMT</pubDate><content:encoded><![CDATA[<p><a href="https://www.georgeyw.com/">I&apos;m</a> a research lead at <a href="https://timaeus.co/?ref=georgeyw.com">Timaeus</a> and attended a <a href="https://www.lesswrong.com/posts/JXcuXhtWoyYNWCt5b/interested-in-cognitive-bootcamp?ref=georgeyw.com">workshop</a> that <a href="https://www.lesswrong.com/users/raemon?mention=user&amp;ref=georgeyw.com">Raemon</a> ran from Oct 4-6 (it was shortened from 4 days to 2.5 days to fit into a weekend). I had prior interest + experience in deliberate practice and enjoyed lots of Ray&apos;s posts about it, so I was curious about the workshop on top of being in a position to actually make impactful plans.&#xA0;</p><p>This is a lightly edited write up that I initially made for myself and for the team at Timaeus about my experience and takeaways. It&apos;s not super polished, but seems better to not clean up and publish than to not clean up and not publish.</p><h1 id="what-was-the-workshop-about">What was the workshop about?</h1><p>Ray is interested in metacognitive improvements. A rough definition of these are: skills or strategies you can learn that make you <a href="https://www.lesswrong.com/posts/46qnWRSR7L2eyNbMA/the-lens-that-sees-its-flaws?ref=georgeyw.com">better able to understand your cognitive process</a> and to influence it in ways that make you more effective as a person. Some examples might include:</p><ul><li>Plan / decision making</li><li>Noticing</li><li>Prediction / calibration training</li></ul><p>This workshop was specifically about <strong>being better at making</strong> <strong>plans</strong>. The terminal goal was to help people who work in deeply confusing and long-feedback loop areas (particularly x-risk reduction). The target audience of this workshop was people who are in a position to be making meaningful plans in their job (i.e., they determine most of their own work and/or the work of others). This workshop was probably most helpful for people who are somewhat familiar with how to make good plans</p><p>Why plans?</p><ul><li>It&apos;s tractable: you can plausibly learn at least one concrete thing in a weekend that will make you better at making plans.</li><li>It&apos;s understandable: you can just follow the procedure of a concrete thing without needing to have inscrutable mental motions described to you.</li><li>It&apos;s impactful: the <em>absolute best</em> plan you can follow (which is might not be the best plan you can <em>make</em>) can often be&#xA0;&gt;&gt;10x better than your default plan, even if your current plan is good.<ul><li>I think this sounds a little crazy until you think about it? But someone could be doing a plan that&apos;s not even net positive. Someone could be working on entirely the wrong thing, and there could be a totally different research agenda or career that has 10x or 100x impact. Someone could just be chasing down the wrong medium-time horizon targets for their grand strategy that slows down the whole thing by massive OOM factors.</li><li>I personally have <em>tons</em> of examples where I retroactively saw the path to 10x better returns if I hadn&apos;t missed something that I could have figured out if I&apos;d thought way harder or more clearly to begin with.</li><li>These results compound: a plan might result in&#xA0;&gt;&gt;10x returns because it unlocks resources for even better plans.</li><li>It&apos;s unfortunately super hard to know ahead of time which plan is&#xA0;&gt;&gt;10x better, but the long-tailed nature of plans means that if you become only a few % better at plan making, you might actually get really outsized returns from it over a long time horizon.</li></ul></li><li>This basically turned out to be a bullet point outline of my theory of value for this type of work. FWIW I&#x2019;m extremely deliberate-practice-pilled, but I think I would judge the value of this workshop to be O($1-10k), and my personal speculation is that it&#x2019;s totally possible for this to have value on O($100k+) if one tried really hard to adopt the tools / mentality from the workshop (and had the leverage to use them, though I guess the sky&apos;s the limit if you have enough leverage).</li><li>I think due to time constraints, this workshop chose to take the&#xA0;<a href="https://www.lesswrong.com/posts/hvj9NGodhva9pKGTj/struggling-like-a-shadowmoth?ref=georgeyw.com"><u>shadowmoth</u></a> approach of &quot;throw people in the deep end and make them learn to swim or not&quot;. I don&#x2019;t think this was a crazy choice, but I think there are more gentle ways to introduce ideas before shadowmothing someone? <s>I&#x2019;ll speculate on this later.</s> (I did not end up speculating on it)</li></ul><h1 id="examples-from-the-workshop">Examples from the workshop</h1><p><strong>(Note: I would recommend&#xA0;<em>not</em> trying to solve or play around with these if you think there is any chance you&#x2019;d like to try out some of the workshop exercises in a formal setting)</strong></p><p>Here are some examples of exercises that we did (may not be in order):</p><ul><li>[Redacted]</li><li><a href="https://www.lesswrong.com/posts/jqb3prwGQjLriq7Lu/exercise-planmaking-surprise-anticipation-and-baba-is-you?ref=georgeyw.com"><u>One-shot Baba Is You</u></a><ul><li>This was my favorite exercise, and the second time that I&#x2019;d done it. I learned a few useful things from this and from listening to other people approach the problem, which I&#x2019;ll elaborate on later.</li><li>You are given a Baba Is You puzzle to solve, but you are not allowed to interact with the puzzle in any way. This is what makes it particularly hard, because Baba Is You is a notoriously unintuitive and difficult puzzle game, and usually the process of solving puzzles involves fiddling with bits and collecting little pieces of evidence and finding out how rules interact through empirics. In this exercise, you don&#x2019;t get any of that, you just get to Think Very Hard and try to make a plan that solves the puzzle on the first try (&quot;one-shotting&quot; it).</li><li>A motivating theme here is &quot;pretend like empirics / experiments are really hard &#x2013; every plan you make takes weeks or months of engineering time, but time spent planning up front is cheaper&quot;<ul><li>Unfortunately, this is kind of a hard mindset to embody, and it seems like a common takeaway is &quot;you can&#x2019;t think about anything successfully ever&quot; and &quot;empirics are incredibly OP&quot; (I think that&#xA0;<em>cheap</em> empirics are incredibly OP, but that&#x2019;s different)</li></ul></li><li>Another motivating theme is &quot;what does it feel like to feel surprised and what does it feel like to have gained an insight?&quot;</li></ul></li><li><a href="https://www.lesswrong.com/posts/PiPH4gkcMuvLALymK/exercise-solve-thinking-physics?ref=georgeyw.com"><u>Thinking Physics</u></a><ul><li>This is also a good exercise I think (and one that I&apos;ve done on my own before). The idea is that you get a puzzle from Thinking Physics and you just sit and try to solve it.</li><li>A motivating theme here is &quot;what does it feel like to have a fuzzy idea of why something is true vs have a very crisp idea of why something is true, in the sense that the world must be deeply fucked somehow if the crisp idea is true, but the world might be totally fine if the fuzzy thing is not true?&quot;</li><li>Another motivating theme is that of &quot;noticing insights&quot;, similar to with Baba Is You.</li></ul></li><li>Some time was spent on freewriting exercises about plans that you actually have IRL, to try and immediately apply lessons from the workshop to IRL problems. This was a significant portion of the time of the workshop, but I won&#x2019;t get into those much.</li></ul><h1 id="what-lessons-were-useful-to-take-away">What lessons were useful to take away?</h1><h2 id="general-skills-that-are-op-not-all-from-the-workshop-including-my-own-takes-in-here">General skills that are OP (not all from the workshop, including my own takes in here)</h2><p>(These are also skills that would make the workshop much more useful for you; having them doesn&apos;t mean the workshop won&apos;t help, it means you can use it to build on them)</p><ul><li>Being well calibrated</li><li>Noticing (I feel like metacognition is ~impossible if you have zero noticing (most people don&#x2019;t have literally zero noticing even if they&#x2019;ve never tried to train it))</li><li>Being unafraid of looking stupid / not having an ego attached to your performance<ul><li>This is a problem I have! I notice that I get emotionally stressed / anxious about not being competitive in solving the exercises compared to other participants. Noticing / acknowledging it seems to help a little.</li></ul></li><li>Patience / taking breaks</li><li>Taking naps / caring for physiological health</li><li>Embodying skills as a lifestyle.<ul><li>Being able to get into the mindset where it &quot;counts&quot; even in trivial situations. If you can&#x2019;t live the values when it doesn&#x2019;t matter, you won&#x2019;t do it when it does matter</li></ul></li></ul><h2 id="specific-skills-from-the-workshop-that-are-op">Specific skills from the workshop that are OP</h2><ul><li>Live logging (similar to having a research log): basically writing down all of the thoughts you&#x2019;re having while thinking.<ul><li>Related to white boxing (which is trying to un-black-box your thought process as much as possible). Live logging is trying to white box yourself, to yourself.</li><li>This is a clever ~shortcut to Noticing, for the purposes of this workshop (so the workshop doesn&apos;t have Noticing as a prereq)</li></ul></li><li>Metastrategic brainstorming: taking a step back and trying to think of radically different approaches to solving the problem<ul><li>Ray thinks this is like, 40-60% of the skill of making better plans (generating them in the first place)</li><li>Basically, you notice that you&#x2019;re stuck, and you see if there&#x2019;s a different type of thing you can try or some intervention that you can make that will shortcut your process. Some examples of different categories of strategies you might brainstorm:<ul><li>Think about the generating process of the problem<ul><li>In places where the problem has contacted reality elsewhere, what does that imply about it?</li></ul></li><li>Brute force + heuristics<ul><li>Is the space of options small enough that you can just do this?</li></ul></li><li>Babble<ul><li>Maximal explore, minimal exploit, or something like that</li></ul></li><li>Instead of building a wall, build a brick<ul><li>What&#x2019;s the smallest thing you could do that would make progress or tell you something new?</li></ul></li><li>Combinatorially examine your assumptions + action / option space<ul><li>Sometimes you will miss something that you were overlooking before (e.g. because your brain automatically pruned it as an option)</li></ul></li><li>Identify OODA loops or make some if you can&#x2019;t find any</li><li>Red team / construct an impossibility proof (of a strategy)<ul><li>Common approach in math, but if you try to prove that your idea is impossible, you might notice something interesting about it</li><li>A related personal takeaway that I noticed here is that this can be a shortcut way to mitigate rabbit-holing on a solution (you pretend like the solution is impossible even if you don&#x2019;t think it is &#x2013; if you lived in that world, what would your next guess be?). I found that at one point I thought that a solution was the only clear solution, then the game told me I was wrong, then I immediately found the right solution once I had let go of the previous one. So maybe if you could just really make yourself believe that your favorite solution is wrong, on the level of receiving hard evidence without having actually received the evidence?</li></ul></li><li>Take a nap / break / walk / eat something</li></ul></li><li>One participant was skeptical of this, got stuck on something, tried it, and then almost immediately became unstuck.</li></ul></li><li>Shortening feedback loops / &#x201C;<a href="https://www.lesswrong.com/posts/rYq6joCrZ8m62m7ej/how-could-i-have-thought-that-faster?ref=georgeyw.com">How could I have thought that faster?</a>&#x201D;<ul><li>This is tied to live logging, since having a log of your thoughts helps a lot, but going through your actual process for solving a problem and looking for the earliest points where you could have made an insight (instead of when you actually made the insight) and what you could have thought to have made that insight sooner</li><li>This is like, reinforcement training / conditioning for your brain</li></ul></li><li>Having 3 plans, 2 frames, and a crux<ul><li>I&#x2019;m not sure I&#x2019;m super sold on this, but the idea is something like &#x201C;can you try to generate multiple plans and multiple perspectives, and ideally a crux or two about what would make you choose one perspective or plan over another&#x201D; to avoid rabbit-holing (avoiding rabbit-holing is a big underlying theme, I&#x2019;m noticing)</li></ul></li><li>Some other misc things that are small or known elsewhere or weren&#x2019;t a big focus of the workshop<ul><li>Quantified intuitions about &#x201C;good&#x201D; &#x2013; can you figure out how to measure two plans in the same &#x201C;units&#x201D;? E.g. at one point Ray chose between &#x201C;should I help people in the workshop during the next (open working) session or should I spend some time working on iterations and reflections for improving the workshop?&#x201D;<ul><li>An example of a shared unit here is &#x201C;in expectation, the amount of (weighted?) improved researcher-hours on x-risk&#x201D; or something like that</li></ul></li><li>Figuring out your&#xA0;<a href="https://www.lesswrong.com/tag/crux?ref=georgeyw.com"><u>cruxes</u></a> and ideally internally&#xA0;<a href="https://www.lesswrong.com/tag/double-crux?ref=georgeyw.com"><u>double-cruxing</u></a> yourself (between two plans)</li><li>Framing things as ~<a href="https://en.wikipedia.org/wiki/OODA_loop?ref=georgeyw.com"><u>OODA loops</u></a><ul><li>I think this is a useful tool if you haven&#x2019;t run into it before</li><li>We did an exercise where we tried to frame things that we do IRL as OODA loops and identify places where we aren&#x2019;t using them but could be</li></ul></li></ul></li></ul><h1 id="what-did-i-personally-take-away">What did I personally take away?</h1><p>This is a bit more stream of consciousness / thinking out loud.</p><p>This phrase kept getting offhandedly repeated throughout the workshop when Ray was giving examples, and it&#x2019;s stuck in my head. Something like, &#x201C;I would just be in the middle of doing something and I would wake up and become sentient and look around me and be like, what am I doing?&#x201D;</p><p>Something about the &#x201C;wake up and become sentient&#x201D; thing feels like a really major core of the workshop to me. I&#x2019;ve spent a bunch of time (diffuse over many years) working on or thinking about deliberate practice and decision making and stuff, but I noticed that a lot of my patterns and intuitions here have kind of become a bit too subconscious, and I&#x2019;ve forgotten that they&#x2019;re a thing that I can just look at and continue to polish (I&#x2019;ve somehow gotten too busy to remember that deliberate practice is a thing that I should still be doing everywhere).</p><p>For example, in the Baba Is You exercise, I ended up doing 3 different levels (across two sessions), and it wasn&#x2019;t until I&#x2019;d thought about it a bunch afterwards that I realized I had the same type of blind spot in all three cases (my brain is a bit too eager to prune things that are above some threshold of &quot;I&#x2019;m sure the world works this way&quot;). I think this is a pretty generalizable pattern to notice, and it required &quot;waking up&quot; + live logging + reflecting about the process to notice it in the game setting.</p><p>Another big thing in the workshop for me was how important it is to</p><ol><li>Actually do these as exercises and not just read about them</li><li>Actually do these as exercises and not just&#xA0;<em>remember</em> that I&#x2019;d learned these lessons once upon a time (thinking of like, muscles that are not used or stretched in a long time become stiff or atrophy)</li><li>Actually think about calibration and practice noticing more often, and how much these things have helped me in the past and that I can just&#xA0;<a href="https://www.lesswrong.com/tag/more-dakka?ref=georgeyw.com"><u>keep improving these things</u></a> until I see diminishing returns</li></ol><p>This thing wasn&#x2019;t really part of the workshop (the workshop was not focused on execution of plans), but it is something the workshop reminded me of: I can in fact just spend time thinking about how to deliberately practice execution as a research lead. This is something that just seems obvious to do.</p><p>TLDR the thing that maybe had the biggest impact was the workshop as a catalyst for a meta-level &quot;waking up and becoming sentient&quot; about the fact that I can &quot;wake up and become sentient&quot; about object level things. (not that I have just literally been autopiloting, but there are degrees to getting out of your head about things / switching to manual control). Besides that, two object level patterns:</p><ul><li>I still get emotionally agitated sometimes when I feel like I&#x2019;m stupid because I can&#x2019;t solve something</li><li>I prune a little too much and maybe explore a bit too little. I probably trust my subconscious process a little too much and should more often have some doubts about the end result (but also don&#x2019;t want to overcorrect on this)</li></ul><p>And some skills that I&#x2019;m interested in practicing more and / or trying to use everywhere to internalize them:</p><ul><li>Noticing</li><li>Metastrategic Brainstorming</li><li>Live logging (way more than I already do)</li></ul><p>The latter two are the kind of thing where my theory of internalizing them is based on how I got myself to use ChatGPT originally, which was: shoehorn it into everything for a few weeks, then gradually prune use cases to what actually feels useful.</p><p>I also now have the idea of a &#x201C;Deliberate Practice Monastery&#x201D; tattooed on my brain and hear its beautiful siren call. I think there&#x2019;s a real argument to be made that this work is extremely important and in its idealized, full-potential form, can make our best x-risk researchers&#xA0;<em>significantly</em> better.</p><p>I also like a rule of thumb that Ray mentioned during the workshop, which was &quot;spend ~10% of your time on meta&quot;.</p><h1 id="other-misc-thoughts">Other misc thoughts</h1><ul><li>I found myself surprised at how not deliberate-practice-pilled a lot of people are. This type of practice and mindset feel like a natural companion to rationality, but it seems like very few people have spent significant time or attention on tuning these things.<ul><li>It feels like the fruit is hanging <em>so low</em> here, like I genuinely believe there are just Ideas that are not impossible tacit knowledge that can be turned into Words that you can just Say to people and if they Listen they can just immediately be more competent. Perhaps this workshop didn&apos;t yet have the perfect Words, but I think such incantations are definitely within reach.</li><li>It feels kind of insane to me that Ray is fighting to prove that this is useful enough to spend more effort on. I feel like this is just self evidently important or something, I don&apos;t know if I&apos;m missing something or <a href="https://www.lesswrong.com/posts/Zp6wG5eQFLGWwcG6j/focus-on-the-places-where-you-feel-shocked-everyone-s?ref=georgeyw.com">everyone else is</a>?</li></ul></li><li>It is hard to come up with good exercises for this type of workshop, and I&#x2019;m impressed at what Ray has done so far</li><li>I think case studies would be a really useful addition. My suggestion here was to include a few demonstrations (by Ray) of the proposed tools, but my wishlist would include examples in science or history where a plan making process was documented / legible and resulted in obviously better plans.</li><li>I still mostly think planning is essential and plans are meaningless or whatever the quote is.</li><li>Cheap empirics are OP as hell, you take things for granted until they&#x2019;re missing.</li><li>All else being equal, velocity is OP. If you can do things 10% faster, you learn 10% faster too, everything compounds and is more than 10% better.</li></ul><h1 id="concrete-value-in-my-day-to-day-research">Concrete value in my day-to-day research</h1><p>I&apos;m redacting most of this section b/c it&apos;s high context and also maybe private, I&apos;m not sure and don&apos;t want to think too hard about it. The thrust of this section was something like:</p><ul><li>Looking back, if I had really, with my whole ass, internalized and practiced some of these lessons at the start of this year (maybe more than what the workshop alone would give? but definitely something achievable with the workshop as a catalyst), I could easily imagine saving a few weeks here or there, adding up to potentially months of speed up in my productivity (in the original document, there were four concrete examples).</li><li>In practice, I would guess that over the past few weeks, trying hard to apply things like metastrategic brainstorming and other forms of deliberate / purposeful practice have already saved time measured in days (in expectation, over the next few months).</li></ul><h1 id="go-to-the-next-one">Go to the next one!</h1><p>Ray is hosting <a href="https://www.lesswrong.com/posts/ynsE7aB43bRJpHeNj/the-cognitive-bootcamp-agreement?ref=georgeyw.com">another workshop this weekend (Oct 25-27)</a>. Go do it. Actually doing the exercises is miles better than just reading about them (and going home afterwards and continuing to practice is miles better than that).</p>]]></content:encoded></item><item><title><![CDATA[So you want to work on technical AI safety]]></title><description><![CDATA[Advice for aspiring safety researchers]]></description><link>https://www.georgeyw.com/so-you-want-to-work-on-technical-ai-safety/</link><guid isPermaLink="false">6752671ea72d5eb24577e325</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Tue, 25 Jun 2024 01:53:00 GMT</pubDate><content:encoded><![CDATA[<p>I&#x2019;ve been to two EAGx events and one EAG, and the vast majority of my one on ones with junior people end up covering some subset of these questions. I&#x2019;m happy to have such conversations, but hopefully this is more efficient and wide-reaching (and more than I could fit into a 30 minute conversation).</p><p>I am specifically aiming to cover advice on getting a job in empirically-leaning technical research (interp, evals, red-teaming, oversight, etc) for new or aspiring researchers without being overly specific about the field of research &#x2013; I&#x2019;ll try to be more agnostic than something like Neel Nanda&#x2019;s <a href="https://www.neelnanda.io/mechanistic-interpretability/quickstart?ref=georgeyw.com">mechinterp quickstart guide</a>&#xA0;but more specific than the <a href="https://forum.effectivealtruism.org/posts/vLFec4srctCMvoPyS/advice-for-early-career-people-seeking-jobs-in-ea?ref=georgeyw.com">wealth of career advice</a>&#xA0;that already exists but that applies to ~any career. This also has some overlap with this excellent <a href="https://www.lesswrong.com/posts/dZFpEdKyb9Bf4xYn7/tips-for-empirical-alignment-research?ref=georgeyw.com">list of tips</a>&#xA0;from Ethan Perez but is aimed a bit earlier in the funnel.</p><p>This advice is of course only from my perspective and background, which is that I did a PhD in combinatorics, worked as a software engineer at startups for a couple of years, did the <a href="https://aifuturesfellowship.org/?ref=georgeyw.com">AI Futures Fellowship</a>, and now work at <a href="https://timaeus.co/?ref=georgeyw.com">Timaeus</a>&#xA0;as the research lead for our language model track. In particular, my experience is limited to smaller organizations, so &#x201C;researcher&#x201D; means some blend of research engineer and research scientist rather than strictly one or the other.</p><p>Views are my own and don&#x2019;t represent Timaeus and so on.</p><h2 id="requisite-skills">Requisite skills</h2><h3 id="what-kind-of-general-research-skills-do-i-need">What kind of general research skills do I need?</h3><p>There&#x2019;s a lot of tacit knowledge here, so most of what I can offer is more about the research process. Items on this list aren&#x2019;t necessarily things you&#x2019;re expected to just have all of or otherwise pick up immediately, but they&#x2019;re much easier to describe than e.g. research taste. These items are in no particular order:</p><ul><li><strong>Theory of change at all levels.</strong>&#xA0;Yes, yes, theories of change, they&#x2019;re great. But theories of change are most often explicitly spoken of at the highest levels: how is research agenda X going to fix all our problems? Really, it&#x2019;s theories of change <em>all the way down</em>. The experiment you&#x2019;re running today should have some theory of change for how you understand the project you&#x2019;re working on.&#xA0;Maybe it&#x2019;s really answering some question about a sub-problem that&#x2019;s blocking you. Your broader project should have some theory of change for your research agenda, even though it probably isn&#x2019;t solving it outright. If you can&#x2019;t trace up the stack why the thing you&#x2019;re doing day to day matters for your ultimate research ambitions, it&#x2019;s a warning flag that you&#x2019;re just spinning your wheels.</li><li><strong>Be ok with being stuck.</strong>&#xA0;From a coarse resolution, being stuck is a very common steady state to be in. This can be incredibly frustrating, especially if you feel external pressure from feeling that you&#x2019;re not meeting whatever expectations you think others have or if your time or money is running out (see also below, on managing burnout). Things that might help for a new researcher are to have a mentor&#xA0;(if you don&#x2019;t have access to a human, frontier LLMs are (un)surprisingly good!) that can reassure you that your rate of progress is fine and to be more fine-grained about what progress means. If your experiment failed but you learned something new, that&#x2019;s progress!</li><li><strong>Quickly prune bad ideas.</strong>&#xA0;Always look for cheap, fast ways to de-risk investing time (and compute) into ideas. If the thing you&#x2019;re doing is really involved, look for additional intermediates as you go that can disqualify it as a direction.</li><li><strong>Communication.</strong>&#xA0;If you&#x2019;re collaborating with others, they should have some idea of&#xA0;what you&#x2019;re doing and why you&#x2019;re doing it, and your results should be clearly and quickly communicated.&#xA0;Good communication habits are kind of talked about to death, so I won&#x2019;t get into them too much here.</li><li><strong>Write a lot.</strong>&#xA0;<a href="https://boz.com/articles/writing-thinking?ref=georgeyw.com">Writing is thinking</a>. I can&#x2019;t count the number of times that I felt confused about something and the answer came while writing it down as a question to my collaborators, or the number of new research threads that have come to mind while writing a note to myself or others.</li><li><strong>Be organized.</strong>&#xA0;Figure out some kind of system that works for you to organize your results. When you&#x2019;re immersed in a research problem, it can feel deceptively easy to just keep all the context of your work and the various scattered places information is stored in your head. I currently keep a personal research log<a href="#fny3yp69uvsmd"><sup>[1]</sup></a>&#xA0;in a Google doc (also visible to my collaborators) that I write entries into throughout the lifetime of a project. The level of detail I aim for is to be able to revisit an entry or a plot months later and to be able to recall the finer points from there &#x2013; on average this actually happens in practice about once a week and has saved me a great deal of headache.</li><li><a href="https://terrytao.wordpress.com/career-advice/continually-aim-just-beyond-your-current-range/?ref=georgeyw.com"><strong>Continually aim just beyond your range.</strong></a><strong>&#xA0;</strong>Terry Tao has a ton of great <a href="https://terrytao.wordpress.com/career-advice/?ref=georgeyw.com">career advice</a>, much of which is transferable to other fields of research beyond math. Research is a skill, and like many other skills, you don&#x2019;t grow by just doing the same things in your comfort zone over and over.</li><li><strong>Make your mental models legible. </strong>It&#x2019;s really hard to help someone who doesn&#x2019;t make it easy to help them! There&#x2019;s a ton of things that feel embarrassing to share or ask, and this is often a signal that you should talk about it! But it&#x2019;s also important to communicate things that you <em>don&#x2019;t</em>&#xA0;feel embarrassed about. You might be operating off of subtly bad heuristics, and someone with more experience can only correct you if you either say things in a way that reveals the heuristic or if you do a thing that reveals it instead (which is often more costly).</li><li><strong>Manage burnout.</strong>&#xA0;The framing that I find the most helpful with burnout is to think of it as a mental overuse injury, and the steps to recover look a lot like dealing with a physical overuse injury. Do all the usual healthy things (sleep enough, eat well, exercise) and ease into active recovery, which emphatically does not look like taking a few days off and then firing on all cylinders again. Much like physical overuse injuries, it&#x2019;s possible to notice signs ahead of time and to take preventative steps earlier. This is much easier after going through the process of burning out once or twice. For example, I notice that I start doing things like snacking more, procrastinating sleep, and playing more video games. These things happen well before there&#x2019;s any noticeable impact on my productivity. Finally, burning out is not a judgment of your research ability &#x2013; extremely competent researchers still have to manage burnout, just as professional athletes still have to manage physical injuries.</li></ul><h3 id="what-level-of-general-programming-skills-do-i-need">What level of general programming skills do I need?</h3><p>There is a meaningful difference between the programming skills that you typically need to be effective at your job and the skills that will let you <em>get</em>&#xA0;a job. I&#x2019;m sympathetic to the view that the job search is inefficient / unfair and that it doesn&#x2019;t really test you on the skills that you actually use day to day. It&#x2019;s still unlikely that things like <a href="https://leetcode.com/?ref=georgeyw.com">LeetCode</a>&#xA0;are going to <a href="https://www.lesswrong.com/posts/Z87fSrxQb4yLXKcTk/mats-winter-2023-24-retrospective?ref=georgeyw.com#Engineering_Tests">go away</a>. A core argument in their favor is that there&#x2019;s highly asymmetric information between the interviewer and interviewee and that the interviewee has to <a href="https://en.wikipedia.org/wiki/Signalling_(economics)?ref=georgeyw.com">credibly signal</a>&#xA0;their competence in a relatively low bandwidth way. False negatives are generally much less costly than false positives in the hiring process, and LeetCode style interview questions are skewed heavily towards false negatives.</p><p>Stepping down from the soapbox, the table stakes for passing technical screens are knowing basic data structures and algorithms and being able to answer interview-style coding questions. I personally used <a href="https://ocw.mit.edu/courses/6-006-introduction-to-algorithms-spring-2020/video_galleries/lecture-videos/?ref=georgeyw.com">MIT&#x2019;s free online lectures</a>, but there&#x2019;s an <a href="https://github.com/tayllan/awesome-algorithms?ref=georgeyw.com">embarrassment of riches</a>&#xA0;out there. I&#x2019;ve heard Aditya Bhargava&#x2019;s <em>Grokking Algorithms</em>&#xA0;independently recommended several times. Once you have the basic concepts, do LeetCode problems until you can reliably solve LeetCode mediums in under 30 minutes or so. It can be worth investing more time than this, but IME there are diminishing returns past this point.</p><p>You might also consider trying to create some small open source project that you can point to, which can be either AI safety related or not. A simple example would be a weekend hackathon project that you put on your CV and your personal GitHub page&#xA0;that prospective employers can skim through (which you should have, and which you should put some minimal level of effort into making look nice). If you don&#x2019;t have a personal GitHub page with lots of past work on it (I don&#x2019;t, all of my previous engineering work has been private IP, but do as I say, not as I do), at least try to have a personal website to help you stand out (mine is <a href="https://www.georgeyw.com/">here</a>, and I was later told that one of my blog posts was fairly influential in my hiring decision).</p><p>Once you&#x2019;re on the job, there&#x2019;s an enormous number of skills you need to eventually have. I won&#x2019;t try to list all of them here, and I think many lessons here are better internalized by making the mistake that teaches them. One theme that I&#x2019;ll emphasize though is to be fast if nothing else. If you&#x2019;re stuck, figure out how to get moving again &#x2013; read the documentation, read the source code, read the error messages. Don&#x2019;t let your eyes just gloss over when you run into a roadblock that doesn&#x2019;t have a quick solution on Stack Overflow. If you&#x2019;re already moving, think about ways to move faster (for the same amount of effort). All else being equal, if you&#x2019;re doing things 10% faster, you&#x2019;re 10% more productive. It also means you&#x2019;re making more mistakes, but that&#x2019;s an opportunity to learn 10% faster too :)</p><h3 id="what-level-of-aiml-experience-do-i-need">What level of AI/ML experience do I need?</h3><p>Most empirical work happens with LLMs these days, so this mostly means familiarity with them. <a href="https://course.aisafetyfundamentals.com/alignment?ref=georgeyw.com">AI Safety Fundamentals</a>&#xA0;is a good starting point for getting a high level sense of what kinds of technical research are done. If you want to get your hands dirty, then the aforementioned <a href="https://www.neelnanda.io/mechanistic-interpretability/quickstart?ref=georgeyw.com">mechinterp quickstart guide</a>&#xA0;is probably as good a starting point as any, and for non-interp roles you probably don&#x2019;t need to go through the whole thing. <a href="https://mango-ambulance-93a.notion.site/ARENA-Virtual-Resources-7934b3cbcfbf4f249acac8842f887a99?ref=georgeyw.com">ARENA</a>&#xA0;is also commonly recommended.</p><p>Beyond this, your area of interest probably has its own introductory materials (such as <a href="https://www.lesswrong.com/s/czrXjvCLsqGepybHC?ref=georgeyw.com">sequences</a>&#xA0;or <a href="https://www.lesswrong.com/posts/6g8cAftfQufLmFDYT/you-re-measuring-model-complexity-wrong?ref=georgeyw.com">articles</a>&#xA0;on LessWrong) that you can read, and there might be <a href="https://devinterp.com/projects?ref=georgeyw.com">lists of bite-sized open problems</a>&#xA0;that you can start working on.</p><h3 id="should-i-upskill">Should I upskill?</h3><p>I feel like people generally overestimate how much they should upskill. Sometimes it&#x2019;s necessary &#x2013; if you don&#x2019;t know how to program and you want to do technical research, you&#x2019;d better spend some time fixing that. But I think more often than not, spending 3-6 months just &#x201C;upskilling&#x201D; isn&#x2019;t too efficient.</p><p>If you want to do research, consider just taking the shortest path of actually working on a research project. There are tons of accessible problems out there that you can just start working on in like, the next 30 minutes. Of course you&#x2019;ll run into things you don&#x2019;t know, but then you&#x2019;ll know what you need to learn instead of spending months over-studying, plus you have a project to point to when you&#x2019;re asking someone for a job.</p><h3 id="should-i-do-a-phd">Should I do a PhD?</h3><p>Getting a PhD seems to me like a special case of upskilling. I used to feel more strongly that it was generally a bad idea for impact unless you also want to do a PhD for other reasons, but currently I think it&#x2019;s unclear and depends on many personal factors. Because the decision is so context-dependent, it&#x2019;s a bit out of scope for this post to dive into, but there are some more <a href="https://www.lesswrong.com/posts/yi7shfo6YfhDEYizA/more-people-getting-into-ai-safety-should-do-a-phd?ref=georgeyw.com">focused posts</a>&#xA0;with good discussion elsewhere. I think <a href="https://www.georgeyw.com/my-phd-costed-me-1-17m-so-far-but-id-do-it-again/">my own experience</a>&#xA0;was very positive for me (even if it wasn&#x2019;t clear that was the case at the time), but it also had an unusual amount of <a href="https://slatestarcodex.com/2020/05/12/studies-on-slack/?ref=georgeyw.com">slack</a>&#xA0;for a PhD.</p><h2 id="actually-getting-a-job">Actually getting a job</h2><h3 id="what-are-some-concrete-steps-i-can-take">What are some concrete steps I can take?</h3><p>Here&#x2019;s a list of incremental steps you can take to go from no experience to having a job. Depending on your background and how comfortable you feel, you might skip some of these or re-order them. As a general note, <strong>I don&apos;t recommend that you try to do everything listed in depth</strong>. I&apos;m trying not to leave huge gaps here, but you can and should try to skip forward aggressively, and you&apos;ll probably find that you&apos;re ready for later steps much sooner than you think you are (see also upskilling above).</p><ul><li>Learn to code (see above)</li><li>If your area of interest is low-level (involves digging into model internals), learn the basics of <a href="https://www.youtube.com/watch?v=fNk_zzaMoSs&amp;list=PLZHQObOWTQDPD3MizzM2xVFitgF8hE_ab&amp;ref=georgeyw.com">linear algebra</a>&#xA0;and run through some quick tutorials on implementing some model components from scratch (you can just search &#x201C;implement X from scratch&#x201D; and a billion relevant Medium articles will pop up)</li><li>Find some ML paper that has a walkthrough of implementing it. It doesn&#x2019;t have to be in your area of interest or even AI safety related at all. There are varying degrees of hand-holding (which isn&#x2019;t a bad thing at this stage). Here is a particularly <a href="https://course.fast.ai/?ref=georgeyw.com">in-depth guide</a>&#xA0;that eventually implements the Stable Diffusion algorithm in part 2, but it might be a bit overkill. You can probably find plenty of other resources online through a quick web search, e.g. this <a href="https://news.ycombinator.com/item?id=34503362&amp;ref=georgeyw.com">HN thread</a>.</li><li>Learn a little bit about your area of interest (or <em>some</em>&#xA0;area of interest; you don&#x2019;t have to know right this second what field you&#x2019;re going to contribute to forever!)</li><li>Find a paper in your area of interest that looks tractable to implement and try to implement it on your own. If you have trouble with this step, try finding one that has a paper walkthrough or an implementation somewhere on GitHub that you can peek at when you get stuck.</li><li>Find an open problem in your area of interest that looks like it could be done in a weekend. It&#x2019;ll probably take more than a weekend, but that&#x2019;s ok. Work on that open problem for a while.</li><li>If you get some interesting results, that&#x2019;s great! If you don&#x2019;t, it&#x2019;s also ok. You can shop around for a bit and try out other problems. Once you&#x2019;ve committed to thinking about a problem though, give it a real shot before moving on to something else. In the long run, it can be helpful to have a handful of bigger problems that you rotate through, but for now, just stick to one at a time.</li><li>Ideally you now have some experience with small open problems. This is where legibility becomes important &#x2013; try and write up your results and host your code in a public GitHub repo.</li><li>Now you have something to take to other people to show that you&#x2019;re capable of doing the work that you&#x2019;re asking someone to pay you to do, so go and take this to other people. Start applying for fellowships, grants, and jobs. While you&#x2019;re applying, continue working on things and building up your independent research experience.</li><li>It might take a while before you get a bite. A job would be nice at this point, but it might not be the first opportunity you get. Whatever it is, do a good job at it and use it to build legible accomplishments that you can add to your CV.</li><li>Continue applying and aiming for bigger opportunities as your accomplishments grow. Repeat until you have a job that you like.</li></ul><p>The general idea here is to do a small thing to show that you&#x2019;re a good candidate for a medium thing, then do a medium thing to show you can do a bigger thing, and so on. It&#x2019;s often a good idea to apply for a variety of things, including things that seem out of your reach, but it&#x2019;s also good to keep expectations in check when you don&#x2019;t have any legible reasons that you&#x2019;d be qualified for a role. Note that some technical research roles might involve some pretty unique work, and so there wouldn&#x2019;t be an expectation that you have legible accomplishments in the same research area. In those cases, &#x201C;qualified&#x201D; means that you have transferable skills and general competency.</p><h3 id="how-can-i-find-and-make-job-opportunities">How can I find (and make!) job opportunities?</h3><p>I used <a href="https://jobs.80000hours.org/?ref=georgeyw.com">80k Hours</a>&#x2019; job board and LessWrong (I found Timaeus here). If you find your way into Slacks from conferences or local EA groups, there will often be job postings shared in those as well. My impression is that the majority of public job opportunities in AI safety can be found this way. I started working at Timaeus before I attended my first EAG(x), so I can&#x2019;t comment on using those for job hunting.</p><p>Those are all ways of finding job opportunities that already exist. You can also be extremely proactive and make your own opportunities! The cheapest thing you can do here is just cold email people (but write <a href="https://sriramk.com/coldemail/?ref=georgeyw.com">good cold emails</a>). If you really want to work with a specific person / org, you can pick the small open problems you work on to match their research interests, then reach out to discuss your results and ask for feedback. Doing this at all would put you well above the average job candidate, and if you&#x2019;re particularly impressive, they might go out of their way to make a role for you (or at least have you in mind the next time a role opens up). At worst, you still have the project to take with you and talk about in the future.</p><h2 id="sensemaking-about-impact">Sensemaking about impact</h2><p>When I first decided to start working in AI safety, I had very little idea of what was going on &#x2013; who was working on what and why, which things seemed useful to do, what kinds of opportunities there were, and how to evaluate anything about anything. I didn&#x2019;t already know anyone that I could ask. I think I filled out a career coaching or consultation form at one point and was rejected. I felt stressed, confused, and lonely. It sucked! For months! I think this is a common experience. It gets better, but it took a while for me to feel anywhere close to oriented. These are some answers to questions that would have helped me at the time.</p><h3 id="what-is-the-most-impactful-work-in-ai-safety">What is the most impactful work in AI safety?</h3><p>I spent a lot of time trying to figure this out, and now I kind of think this is the wrong way to think about this question. My first attempt when I asked myself this was something like &#x201C;it must be something in AI governance, because if we really screw that up then it&#x2019;s already over.&#x201D; I still think it&#x2019;s true that if we screw up governance then we&#x2019;re screwed in general, but I don&#x2019;t think that it being a bottleneck is sufficient reason to work on it. I have doubts that an indefinite pause is possible &#x2013; in my world model, we can plausibly buy some years and some funding if policy work &#x201C;succeeds&#x201D; (whatever that means), but there still has to be something on the other side to buy time and funding for. Even if you think an indefinite pause <em>is</em>&#xA0;possible, it seems wise to have insurance in case that plan falls through.</p><p>In my model, the next things to come after AI governance buys some time are things like evals and control. These further extend the time that we can train advanced AI systems without major catastrophes, but those alone won&#x2019;t be enough either. So the time we gain with those can be used to make further advances in things like interpretability. Interpretability in turn might work long enough for other solutions to mature. This continues until hopefully, somewhere along the way, we&#x2019;ve &#x201C;solved alignment.&#x201D;</p><p>Maybe it doesn&#x2019;t look exactly like these pieces in exactly that order, but I don&#x2019;t think there&#x2019;s any one area of research that can be a complete solution and can also be done fast enough to not need to lean on progress from other agendas in the interim. If that&#x2019;s the case, how can any single research agenda be the &#x201C;most impactful?&#x201D;</p><p>Most research areas have people with sensible world models that justify why that research area is good to work in. Most research areas also have people with sensible world models that justify why that research area is bad to work in! You don&#x2019;t have to be able to divine The Truth within your first few months of thinking about AI safety.</p><p>What&#x2019;s far more important to worry about, especially for a first job, is just personal fit. Personal fit is the comparative advantage that makes you better than the median marginal person doing the thing. It&#x2019;s probably a bad idea to do work that you&#x2019;re not a good fit for, even if you think the work is super impactful &#x2013; this is a waste of your comparative advantage, and we need to invest good people in all kinds of different bets. Pick something that looks remotely sensible that you think you might enjoy and give it a shot. Do some work, get some experience, and keep thinking about it.</p><p>On the &#x201C;keep thinking&#x201D; part, there&#x2019;s also a sort of competitive exclusion principle at play here. If you follow your curiosity and keep making adjustments as your appraisal of research improves, you&#x2019;ll naturally gradient descent into more impactful work. In particular, it&#x2019;ll become clearer over time if the original reasons you wanted to work on your thing turned out to be robust or not. If they aren&#x2019;t, you can always move on to something else, which is easier after you&#x2019;ve already done a first&#xA0;thing.</p><h3 id="on-how-to-update-off-of-people-you-talk-to">On how to update off of people you talk to</h3><p>Ok this isn&#x2019;t a question, but it&#x2019;s really hard, as a non-expert, to tell whether to trust one expert&#x2019;s hot takes or another&#x2019;s. AI safety is a field full of hot takes and people that can make their hot takes sound really convincing. There&#x2019;s also a massive asymmetry in how much they&#x2019;ve thought about it and how much you&#x2019;ve thought about it &#x2013; for any objection you can come up with on the spot, they probably have a cached response from dozens of previous conversations that makes your objection sound naive. As a result, you should start off with (but not necessarily keep forever) a healthy amount of blanket skepticism about everything, <a href="https://slatestarcodex.com/2019/06/03/repost-epistemic-learned-helplessness/?ref=georgeyw.com">no matter how convincing it sounds</a>.</p><p>Some particularly common examples of ways this might manifest:</p><ul><li><strong>&#x201C;X research is too reckless / too dual-use&#x201D; or</strong>&#xA0;<strong>&#x201C;Y research is too slow / too cautious.&#x201D;</strong>&#xA0;We all have our own beliefs about how to balance the trade-offs of research between capabilities and safety, and we self-select into research areas based on a spectrum of those beliefs. Then from where we stand, we point in one direction and say that everyone over <em>there</em>&#xA0;is too careful and point in the other direction and say that everyone over <em>there</em>&#xA0;is too careless.</li><li><strong>&#x201C;Z research doesn&#x2019;t seem (clearly) net-positive.&#x201D;</strong>&#xA0;People have different thresholds for what makes something obvious and also have disagreements on how much not-obviously-net-positive work is optimal (I claim that the optimal amount of accidentally net-negative work is <a href="https://scottaaronson.blog/?p=40&amp;ref=georgeyw.com">not zero</a>, which is probably much less controversial than if I try to claim <em>exactly how much</em>&#xA0;is optimal).</li></ul><p>I emphatically do not mean to say that all positions on these spectrums are equally correct, and it&apos;s super important that we have truth-seeking discussions about these questions. But as someone new, you haven&apos;t yet learned how to evaluate different positions and it&#x2019;s easy to prematurely set the Overton window based on the first few takes you hear.</p><h2 id="some-encouragement">Some encouragement</h2><p>This isn&#x2019;t a question either, but dropping whatever you were doing before is hard, and so is finding a foothold in something new. Opportunities are competitive and rejections are common in this space, and interpreting those rejections as &#x201C;you&#x2019;re not good enough&#x201D; stings especially hard when it&#x2019;s a cause you care deeply about. Keep in mind that applications skew heavily towards false negatives and that to whatever degree that &#x201C;not meeting the bar&#x201D; can be true, it is a dynamic statement about your current situation, not a static statement about who you fundamentally are. Remember to take care of yourself, and good luck.</p><hr><p><em>Thanks for feedback: Jesse Hoogland, Zach Furman</em></p>]]></content:encoded></item><item><title><![CDATA[Calibrating your gut]]></title><description><![CDATA[Trusting your gut 101]]></description><link>https://www.georgeyw.com/calibrating-your-gut/</link><guid isPermaLink="false">66a51b54e68b640a7230caf8</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sat, 17 Feb 2024 03:20:18 GMT</pubDate><content:encoded><![CDATA[<p>By <em>gut</em> or <em>gut feeling</em> I mean our non-verbal physical and emotional responses to a decision that are pointed to by words like <em>intuition</em> or <em>instinct</em> and phrases like <em>follow your heart</em>. When it cooperates, our gut can be a source of deeply felt conviction, even in murky waters, or &#x2013; when it doesn&apos;t &#x2013; a source of uncertainty, even when we know all the information. Gut feelings act on both a short timescale (the milliseconds before conscious thought even starts) and on a long timescale (the hours, days, weeks of deliberation on a difficult life choice). Whether we notice or not, our gut feelings influence the way we move through the world.</p><p>Okay, and what about <em>calibrating</em>? By calibrating, I mean something like &quot;bringing into alignment with reality,&quot; i.e. making it so that your gut feelings seem to be good predictors of real-world outcomes. But words like intuition and instinct seem to imply a somewhat immutable relationship to your gut &#x2013; what is there to <em>do</em> about it? </p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://i.kym-cdn.com/photos/images/original/000/572/078/d6d.jpg" class="kg-image" alt loading="lazy" width="500" height="407"><figcaption><span style="white-space: pre-wrap;">&quot;just trust your gut&quot; -- unfortunately unhelpful if you already know how to </span><i><em class="italic" style="white-space: pre-wrap;">and</em></i><span style="white-space: pre-wrap;"> if you don&apos;t</span></figcaption></figure><p>My answer to this is good, old-fashioned, high-quality, educated speculation grounded in a sample size of one. As with many things, it starts with the <a href="https://agentyduck.blogspot.com/p/noticing.html?ref=georgeyw.com" rel="noreferrer">art of noticing</a>. That is, to &quot;listen to your gut,&quot; but not in the sense of do-what-your-gut-tells-you, more in the sense of try-to-notice-that-your-gut-is-saying-anything-at-all-even-if-you-have-no-idea-what-it&apos;s-going-on-about. In fact, you should be careful of the first sense, because at this stage your gut might still be wrong.</p><h2 id="noticing">Noticing</h2><p>There&apos;s a video game series I&apos;m fond of called Fire Emblem. It&apos;s a tactical RPG where you try to strategically hit things and not get hit back too badly. If you do this well, eventually you get to hit a dragon until it dies and you win. There are various game mechanics involved, but the most basic one is your hit chance. For this, Fire Emblem uses a system called True Hit. Naturally, it lies to you.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.escapistmagazine.com/wp-content/uploads/2023/06/fire-emblem-gba-nintendo-switch-online-june-23.jpg?resize=800%2C400" class="kg-image" alt="Fire Emblem Joins GBA Nintendo Switch Online on June 23" loading="lazy" width="800" height="400"><figcaption><span style="white-space: pre-wrap;">right: player character a 93% hit chance is displayed, but the actual hit chance is ??%</span></figcaption></figure><p>The details of how exactly it lies to you aren&apos;t that important. The important effect is that displayed hit chances that are above 50% are actually higher in practice, and those below 50% are actually lower in practice. Fire Emblem does this because most players, on a <em>gut level</em>, feel that a 95% chance event ought to happen much more often than 95% of the time, and that a 99% chance event is an <em>almost certainty</em> &#x2013; it feels like an injustice when it doesn&apos;t happen! In that sense, Fire Emblem calibrates the game mechanics to better fit a player&apos;s gut feelings to make the game feel more intuitive. But what this really does is just hide what your gut is trying to say, even if it&apos;s wrong.</p><p>Reality typically won&apos;t recalibrate itself to match your gut feelings, but that means we can <em>notice</em> what happens when our gut is saying something. The exact &quot;motions&quot; (mostly involuntary and mental rather than physical) happening aren&apos;t super important, only the noticing of them.  There might be a certain way that we plan <em>around</em> an event, taking it for granted when we shouldn&apos;t. We might mentally dwell too much on an irrelevant fact or gloss over something important. We might flinch away from an uncomfortable thought or fidget or feel various physical sensations. There might be a specific flavor of mental unease that pushes us to defer entirely to spreadsheets. Whatever it is, there&apos;s something there to attend to.</p><h2 id="calibrating-yourself-to-your-gut">Calibrating yourself to your gut</h2><p>Last year, I attended an ad hoc workshop on improving your calibration for prediction markets. There was surprisingly little quantitative analysis &#x2013; it was mostly about learning to pay conscious attention to the specific gears moving in our brains that led to them spitting out a probability. The point was to be able to correctly associate these gut feelings / gears / mental motions to reality. That is, suppose you think something is the right call 90% of the time and that comes with some kind of internal subjective experience composed of various feelings and motions. Actually though, suppose it turns out that&apos;s only true 70% of the time. If we&apos;re paying attention, we can learn to re-associate our mental connection with that cluster of emotions from being &quot;what we feel when something is right 90% of the time&quot; to &quot;what we feel when something is right 70% of the time.&quot;</p><p>That exercise was limited to prediction markets and assignment of probabilities, but I suspect this is much more broadly applicable. If we&apos;re listening to a new argument for something, we can pay attention to where we feel resistance, if our nervous system gets activated, or where our mind tries to pattern match, and later cross-check that against a more objective judgment. When presented with an unfamiliar situation, we can notice the things that happen internally as we move through the situation and make decisions and later review if our internal feelings seemed to correspond well with what turned out to be the right decision or not. </p><p>As a personal example, I&apos;ve historically struggled with writing the bottom line first and working backwards to justify it. I recently started noticing &#x2013; <em>feeling in my gut</em> &#x2013; what it feels like when I&apos;m rationalizing something, before I consciously realize I&apos;m doing so (and I then try to follow that up by explicitly acknowledging it and positively reinforcing myself for successfully noticing). This has lots of smaller, day-to-day consequences, but it was also a larger scale version of this noticing (that I was rationalizing my trajectory at the time) that became a catalyst for my recent career shift to <a href="https://www.georgeyw.com/moving-to-ai-safety/" rel="noreferrer">AI safety</a>. </p><h2 id="actually-trusting-your-gut">Actually trusting your gut</h2><p>So far we&apos;ve been talking about how to listen to and interpret what your gut is saying. That seems pretty valuable to me already, but sometimes you might have to do the <em>opposite</em> of what your gut is saying if you&apos;ve noticed it&apos;s consistently getting something wrong (such as being uneasy with hard conversations despite them typically turning out well). I claim that at this point, most of the difficult work is done, and you just have to give it time.</p><p>If you maintain this explicit mental attention / feedback loop between your internal motions and outcomes, then your gut feelings will gradually evolve. They&apos;ll tend to ease their discomfort in situations where you notice that such discomfort is unnecessary and to strengthen their conviction where you notice it is warranted. If you stay in tune with these changes, then better calibration &#x2013; and being able to <em>trust</em> your gut &#x2013; is something that just naturally follows.</p>]]></content:encoded></item><item><title><![CDATA[Alternatives to (math) academia and how to get there]]></title><description><![CDATA[This is a reference post for math PhDs]]></description><link>https://www.georgeyw.com/math-alternatives/</link><guid isPermaLink="false">66a51b54e68b640a7230caf6</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sun, 03 Dec 2023 07:08:55 GMT</pubDate><content:encoded><![CDATA[<p><em>[This is a reference post aimed at mathematicians (PhDs, postdocs, professors) who might be interested in options outside of academia. It outlines common career paths and how to find a foothold in those directions. This might be of some limited use to those who are interested in the careers but come from a different background. This is unlikely to be of use to people with neither the background nor the career interest.]</em></p><p>I&apos;ve had this conversation a bunch of times now, and I&apos;m happy to talk about it each time, but it seems like it would be more convenient for all parties if there were just a compiled list of my advice somewhere, so here&apos;s an attempt at such a compilation. Apologies in advance for any gaps, please feel free to email me with questions, and answers may get added to this reference, though I think I&apos;m approximately aiming for an 80/20 rule here. </p><h2 id="why-leave">Why leave?</h2><p>Whatever reasons there are will be highly personal. I can list some of my own reasons below, as well as some that I&apos;ve heard, but if you&apos;re reading this, you might have some of your own as well. Note that this is unashamedly biased towards leaving academia, but I don&apos;t think someone is wrong if they conclude that staying in academia is right for them, so long as the decision is well-informed. </p><p>But my first lukewarm take is that the discoverability of other options is generally terrible, and the friction of figuring out what options there are and how to actually achieve those options is enough to stop many people from seriously considering an exit.</p><p>I&apos;ll also note a personal mental obstacle that I had as a young grad student. For whatever reason, I had a sort of cached mental impression that not &quot;making it&quot; to an academic job was a failure. It&apos;s hard to pinpoint exactly where that impression came from. I&apos;m sure some of it was internal, but there&apos;s also a degree to which this view is baked in around us &#x2013; for example, it&apos;s common to refer to leaving academia as selling out, which has some negative connotation. I don&apos;t think holding the view that it&apos;s a failure is healthy, nor is it true. </p><p>In reality, I think <strong>most of the reasons why someone might remain in academia are superseded by the appropriate arrangement outside of academia</strong> (of which there are <em>many</em>), up to and including being able to do <em>more</em> research outside of academia. Read on for more details.</p><h3 id="reasons-to-leave">Reasons to leave</h3><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.georgeyw.com/content/images/2024/07/image-1.png" class="kg-image" alt loading="lazy" width="592" height="327"><figcaption><span style="white-space: pre-wrap;">Do you really need more convincing than this? Scroll down, I guess... (</span><a href="https://twitter.com/nearcyan/status/1681166975303598080?ref=georgeyw.com" rel="noreferrer"><span style="white-space: pre-wrap;">source</span></a><span style="white-space: pre-wrap;">)</span></figcaption></figure><ul><li>You make (a lot) more money outside of academia<ul><li>By the time you would get tenure in academia, you could counterfactually be a senior or staff level researcher/engineer (see later on for a note on &quot;levels&quot;) in tech and earn <a href="https://www.levels.fyi/?compare=Apple%2CFacebook%2CAmazon%2CMicrosoft%2CGoogle&amp;track=Software+Engineer&amp;ref=georgeyw.com" rel="noreferrer">300-500k+ total compensation</a> (combined salary + equity grant, which is as good as cash)</li><li>The <a href="https://www.georgeyw.com/my-phd-costed-me-1-17m-so-far-but-id-do-it-again/" rel="noreferrer">opportunity cost</a> of staying in academia is well into the six figures per year, just from the raw salary difference, not even accounting for compounding growth of investments. If you weren&apos;t already in academia would you <em>pay</em> $200k+ per year in order to <em>be</em> in academia? (if you look further down, my estimation of the opportunity cost at the tenured professor level is much higher)</li></ul></li><li>The job demands, stress, and time commitment are typically lower<ul><li>The demands are often low enough that you can hold a full time job and <em>still have more time for research than in academia</em></li><li>The Google employees in the tweet above are probably on the low end, but not <em>that </em>far from the true median.</li></ul></li><li>You have more flexibility in where you live (especially with remote work)</li><li>Remote work in tech specifically also comes with extremely flexible hours</li><li>Less administrative burden, zero course preps, no outreach/service expectations</li><li>If you learn certain practical skills, you are incredibly employable and don&apos;t <em>need</em> the job security guarantees of tenure (and make enough money anyways that you can absorb unlikely periods of unemployment, or just straight up retire after ~10-20 relatively chill years and just do whatever the hell you want after)</li><li>There may be a personal situation tying you to your job. Depending on that situation (e.g. with visa status), you may find a solution anyways outside of industry (many tech companies sponsor visas)</li></ul><h2 id="where-do-i-go">Where do I go?</h2><p>Ok, so let&apos;s say you&apos;re considering leaving. What are your career options? When I first Googled &quot;math major career options&quot; in undergrad, the top result was being an actuary. Nothing against actuaries, they&apos;re probably cool people, but I don&apos;t want to personally find out. Your career options are much more varied than this.</p><ul><li>Tech / Silicon Valley<ul><li>This can mean researcher (e.g. machine learning, data science, etc) or software engineer</li><li>This is broadly split into Big Tech (FAANG companies) and startups</li><li>Big Tech has the highest compensation between the two and the lowest job demand (which is paradoxical, but FAANG makes a <em>lot</em> of money, and it&apos;s hard to be ultra efficient at their scale)<ul><li>As an intuition pump: Elon caught a lot of flak for firing like, 80% of X (f.k.a. Twitter), but it&apos;s still running (albeit with some performance degradation here and there). I was among the doubters, but so far it seems like he was right on this point. Many big tech companies could probably survive &gt;50% cuts if they had the guts to do it.</li></ul></li><li>Startups have the most exciting work, but lower compensation (higher than academia) and higher job demand (comparable to academia). If you&apos;re an early employee at a startup that eventually IPOs, you&apos;re probably going to walk away very rich, but don&apos;t count on it. The attraction here is working on problems that feel like they&apos;re either building the future or really fixing existing problems in a meaningful way. Be careful not to choose a startup whose mission you don&apos;t care about. The <em>entire point</em> of working at a small startup is mission alignment. Working on a mission you <em>do </em>care about feels amazing. Some real examples of startups that you can really work on:<ul><li>Flying houses (mission: tiny houses that can be airlifted into disaster relief areas)</li><li>Lab grown meat (mission: end animal suffering due to factory farming)</li><li>Alternative energy (mission: solve world energy supply)</li><li>Wearable tech (biometrics: live healthier lives / AI assistants: extend our cognition with new tools)</li><li>Self-driving cars (lots of potential missions, e.g. solve car accident mortality)</li><li>Literal space travel (mission: go to SPACE and maybe COLONIZE MARS)</li></ul></li><li>Those are just the eye-catching ones. If you can think of a real problem in the world, even if it&apos;s incredibly unsexy, there&apos;s probably a startup somewhere working on it (lots of startups also suck though and are working on fake problems, but evaluating startups is an entire 5k word guide on its own)</li></ul></li><li>Quant finance<ul><li>Pay is similar to big tech</li><li>Great if you are specifically extremely competitive, resilient to stress, and can be happy with high variance positive EV decision making even if the realized consequence was negative (think activities like poker)</li><li>There is a particular type of person for whom this is a good fit, but most people are not, and that is fine</li></ul></li><li>NSA<ul><li>The NSA employs a lot of mathematicians, but I&apos;m not familiar with the precise kind of work they do. Presumably a lot of cryptography related work.</li><li>I hear the work-life balance is great at the NSA</li></ul></li><li>CCR campuses<ul><li>Apparently the NSA also has more &quot;relaxed&quot; research centers called Center(s) for Communications Research</li><li>Examples: <a href="https://idaccr.org/?ref=georgeyw.com" rel="noreferrer">Princeton</a>, <a href="https://www.ida.org/en/ida-ffrdcs/center-for-communications-and-computing/center-for-communications-research-la-jolla?ref=georgeyw.com" rel="noreferrer">La Jolla</a> (there are more if you Google, for some reason they don&apos;t seem to be aggregated anywhere)</li></ul></li><li>NASA<ul><li>You can do cool space stuff!</li></ul></li><li>Government labs<ul><li>There are a variety of national labs that fund fundamental research</li><li>If such a lab happens to fund the type of research you do, you can enjoy what is essentially an academic job but without the teaching</li></ul></li><li>Healthcare / biostats<ul><li>A close friend of mine does work in epidemiology using statistics and gets to do covid modeling</li></ul></li><li>Honestly, almost anything<ul><li>Your most powerful tool is being agentic. You can, in some sense, really do whatever you put your mind to &#x2013; most people are <a href="https://drmaciver.substack.com/p/learning-to-walk-through-walls?ref=georgeyw.com" rel="noreferrer">limited by artificial walls</a>. A PhD is <em>proof that you can do hard stuff on your own</em>. The degree of career freedom you have outside of academia is incredible. Not all artificial walls are worth tearing down, but before you think &quot;I could never do that,&quot; look with fresh eyes and ask if that&apos;s really true.</li><li>There are plenty of careers that there aren&apos;t degrees for. America employs ~160M people, and my undergrad institution had like ~120 majors. Lots of people have jobs that they didn&apos;t formally study for, and you can do that too. </li></ul></li><ul><li>Things that I did or almost did at various points (in the span of only 2.5 years!) that no one had to give me permission or a formal list of prereqs for: education tools, music tech, science communication, game design, sports betting (for American football &#x2013; and I <em>still don&apos;t even know what a tight end is</em>), algorithmic trading, software engineering, blockchain/web3, tech consulting/advising for early startups, data science, supply chain optimization, AI safety research, and probably a lot more that I&apos;ve forgotten already</li></ul><li>Other thoughts<ul><li>Perspectives that people have on PhDs are mixed, but there&apos;s more than enough people who look at it and say &quot;oh shit, they have a PhD, they&apos;re probably really smart&quot; rather than &quot;their thesis isn&apos;t relevant, let&apos;s not hire them&quot; that I think it&apos;s net positive (you only need one good hiring manager to think positively towards a PhD)</li><li>There is a ton of interesting work that happens outside of academia.</li></ul></li></ul><h2 id="how-do-i-actually-concretely-get-a-job">How do I actually, concretely get a job?</h2><p><em>[This is from the perspective of a student, but postdocs and professors can adapt the same general strategies.]</em></p><p>For most of the common career paths, the primary new skill you need to pick up is some degree of programming (coding combinatorics experiments is a good starting point, but not representative of the skills you need) and/or some degree of statistics/data science/machine learning (these three all blend together to different degrees at different jobs). Everything else you can learn on the job (and don&apos;t worry &#x2013; everyone does). If you&apos;re picking the &quot;choose-your-own-adventure&quot; option in the previous section, then I&apos;ll leave it up to you to determine how much of what skill you need to learn and how you learn it.</p><ul><li>You probably don&apos;t need to learn much programming, if any, for jobs like at the NSA or at government labs (or if you do, they might just teach you on the job)</li><li>You need to learn a little programming (probably leaning towards statistical languages, e.g. Python, R) but mostly data science concepts for things like biostats and data science</li><li>You need to learn a decent amount of programming if you want to do software engineering</li><li>You need both programming (less than for software engineering) and data science/ML for quant finance</li><li>You need both if you want to specialize to certain parts of software engineering, such as machine learning engineering</li><li>Lukewarm take, but I think basically any job in the realm of knowledge work will benefit from some proper programming experience, even if the job description doesn&apos;t explicitly call for it (and even if none of your coworkers know any)</li></ul><p><strong>From here on, the main focus will be on getting tech/quant/data science jobs. The reason for this is that these are 1) what I&apos;m most familiar with and 2) what I expect will be the most involved processes / require the most domain specific prep knowledge.</strong></p><h3 id="learning-data-science-and-machine-learning">Learning data science and machine learning</h3><p>Your core resource is <a href="https://www.kaggle.com/?ref=georgeyw.com" rel="noreferrer">Kaggle</a>. Start with the <a href="https://www.kaggle.com/competitions/titanic?ref=georgeyw.com" rel="noreferrer">Titanic</a> and <a href="https://www.kaggle.com/competitions/house-prices-advanced-regression-techniques?ref=georgeyw.com" rel="noreferrer">Housing Prices</a> intro competitions (in that order). Find some example notebooks, or if you&apos;re braver, try to analyze the dataset and generate predictions on your own without hints. These notebooks exist to familiarize you with the common tools, languages, and techniques that are used in data science. Pay attention to the questions asked, but also the heuristics and rigor. Notice the types of questions which are <em>not</em> asked. Data science is a different field from mathematics with different standards for what is or isn&apos;t a good answer to a question.</p><p>Work your way up to trying some <a href="https://www.kaggle.com/competitions?ref=georgeyw.com" rel="noreferrer">live competitions</a>, including competitions with real prize money. Pick one that seems interesting, maybe find a few friends who are also trying to learn data science and form a team. Give yourself somewhere from a few weeks to a couple months and dig really deep into that competition. There is no substitute to learning by doing.</p><p>As a reference point for the level of difficulty of questions you might get in interviews:</p><ul><li>One of my onsite questions at a quant hedge fund was essentially a series of questions about the housing prices dataset</li><li>Another ML/data science role asked me to code the k-nearest neighbors algorithm from scratch (I later learned that they didn&apos;t mean <em>literally</em> from scratch, it was meant to be a test of how well I knew the numpy library in Python)</li><li>Various questions on linear regression. These types of questions did not inherently require knowledge beyond the undergraduate level, but they <em>did</em> require a depth of understanding of those concepts that wouldn&apos;t normally be expected of an undergraduate. <ul><li>E.g. suppose you have a dataset on the 2D plane, where data is uniformly distributed in a rectangle. What happens to the linear regression line as you rotate that rectangle 90 degrees clockwise or counter-clockwise?</li><li>Similar questions about normal distributions are common</li></ul></li></ul><h3 id="learning-to-program">Learning to program</h3><p>Programming experience in the software industry is typically split into &quot;production&quot; and &quot;not production&quot; level experience. &quot;Production&quot; means work that was done in a commercial setting or otherwise at a commercial level. This typically implies things such as: real money is on the line, code is run live without supervision most of the time, and coding is done as a team in a shared repository with peer reviews, among other things. Unfortunately, within the software industry, things like personal projects or coding experiments for math research are not typically considered production-level coding experience. </p><p>At first glance, there is an apparent catch-22 that it&apos;s hard to find a job if you don&apos;t have production-level coding experience, but you can&apos;t get production-level coding experience if you never get a job. Internships can be one solution, but maybe you&apos;re applying in your last year and you don&apos;t have any summers left, or maybe you&apos;re a postdoc/professor already. In that case, here is a &quot;minimum viable solution&quot; to getting out of the catch-22:</p><ul><li>Watch the MIT OCW <a href="https://ocw.mit.edu/courses/6-006-introduction-to-algorithms-spring-2020/?ref=georgeyw.com" rel="noreferrer">Introduction to Algorithms</a> videos. Doing problem sets is optional. <ul><li>Become comfortable with the data structures and algorithms presented as well as in the analysis of time and space complexity of solutions.</li></ul></li><li>Doing problem sets in the previous step is optional because doing problems in this step is <em>not</em> optional. Make a <a href="https://leetcode.com/?ref=georgeyw.com" rel="noreferrer">LeetCode</a> account, consider if a premium subscription is useful to you (I bought one month of premium at one point and found it useful enough for one month but not more), and then do a ton of problems.<ul><li>Start with easy problems, then move up to medium once you feel good about easy problems.</li></ul></li><ul><li>There is a (now somewhat infamous maybe) list of <a href="https://gist.github.com/krishnadey30/88c4e2f601e96597974c00185e479532?ref=georgeyw.com" rel="noreferrer">75 LeetCode problems</a> that aim to be representative of the types of problems you&apos;ll run into in interviews. If you&apos;re pressed for time (e.g. graduating soon), this is the minimal set of LC problems you should aim to solve</li><li>LeetCode is seriously the core of the current <a href="https://en.wikipedia.org/wiki/Metagame?ref=georgeyw.com" rel="noreferrer">meta</a> of (entry level) tech interviews. It is a fake skill in the sense that the connection with ability to do real work is dubious (though I don&apos;t think it&apos;s totally indefensible), but it is a real skill in the sense that you can really study to the exam and get good at LC style problems and this is highly instrumental for getting a job in the first place.</li><li>A good benchmark is being able to consistently do LeetCode medium problems on your own, with no hints, in 15-30 minutes. The vast majority of interview problems are not harder than LeetCode medium. Past this point, you start to get diminishing returns on the time spent on LeetCode, but it is (IMO) quite hard to accidentally get to a point where marginal practice is pointless.</li><li>Quick plug for a fun alternative to LeetCode: <a href="https://adventofcode.com/?ref=georgeyw.com" rel="noreferrer">Advent of Code</a>. This is an annual advent calendar of coding problems that increases in difficulty as it approaches Christmas and has a silly story to go along with it.</li></ul><li>Finally, work on some actual personal projects. Find something interesting and just do it. <ul><li>You can sometimes get a 2 for 1 with data science/ML experience by doing some kind of coding-intensive competition on Kaggle.</li><li>The first project I ever worked on that made it onto my resume was an automated arbitrage bot for <a href="https://www.mtgo.com/en/mtgo?ref=georgeyw.com" rel="noreferrer">Magic The Gathering: Online</a>. The code was trash, it was a buggy mess (in our defense, this was partly because MTGO itself was a buggy mess), and we got banned from the platform shortly after. We made some actual money off of running the bot, but the real value was returns on career capital &#x2013; it has one of the highest return rates of any single week I&apos;ve ever spent coding. It is the most common talking point in my interviews and is still proudly displayed on my resume.</li><li>Check out this <a href="https://news.ycombinator.com/item?id=38511280&amp;ref=georgeyw.com" rel="noreferrer">Hacker News thread</a> for examples of long-tail outcomes of interesting projects</li></ul></li><li>Another avenue to show off some of your coding (and to actually get some real code review/feedback) is to make open source contributions on Github. I won&apos;t get into how to do this, but some employers view this very favorably and like to see actual code examples as a way to de-risk hiring you. </li><li>A longer path (which you can take if you decide well in advance that this is the path you&apos;d like to take) is to take some computer science classes, and maybe even get a masters in CS. This seems effective if you&apos;re willing to put in the time commitment and can plan ahead for it.</li></ul><h3 id="misc-quant-questions">Misc quant questions</h3><p>There is a small industry built around certain types of quantitative problems and brain teasers that get asked almost exclusively in quant interviews. Getting into these problems is probably out of scope for this document, but you can find plenty of examples online.</p><h3 id="general-interview-advice">General interview advice</h3><p>Interviewing is a skill, too. Your first few interviews will probably feel like garbage. If I haven&apos;t interviewed for a while, my interview skills get rusty, and it takes a few warm-up interviews to get comfortable with them again. Bias towards interviewing earlier in your interview prep process and just get the crappy first few interviews out of the way. I promise it gets better, and there are enough companies out there that flubbing a few interviews for middle-range targets isn&apos;t an issue.</p><p>For most of the quantitative jobs above, the vast, vast majority of your interviews will be technical interviews. Other interviews (such as with HR or recruiters) will be more like formalities in the process, mostly just checking that you&apos;re reasonably easy to talk to. I didn&apos;t have a single real behavioral interview until one point where I was speculatively interviewing for a totally different kind of job.</p><p>For technical interviews, a full guide is also out of scope for this reference, but here are some general pointers:</p><ul><li>Think out loud (seriously, it is sometimes a red flag for interviewers if you are too quiet)</li><li>Ask clarifying questions (green flag if you ask them, red flag if you don&apos;t + were wrong about interpreting the question)</li><li>Don&apos;t be afraid to start with naive solutions to programming problems where the time complexity sucks and is obviously not optimal. As long as you <em>say that out loud</em> and then iterate from there, it&apos;s totally fine and counts as partial progress towards answering the question</li><li>If a question seems really unusually hard, it might be your unlucky day, but it might also be the case that it&apos;s hard for everyone. Many places rate interview performance on problems &quot;on a curve&quot;</li><li>For some questions, you can get more &quot;points&quot; by sharing high quality thinking out loud without arriving at the right answer than by just jumping to the right answer. What this looks like is: asking the right questions, thinking about the right things, coming up with tractable, reasonable sounding ideas (that may not ultimately work)</li><li>If you know someone who is in tech, call in some favors and do practice interview sessions and get feedback</li><li>The interview process can be pretty long; I think my longest interview processes had 5 rounds (where the last one is an &quot;onsite,&quot; though these are often over Zoom now instead of physically onsite), and I&apos;ve heard of more. Have a constant pipeline of interviews, and don&apos;t wait for one interview process to end before looking for another. </li><li>Don&apos;t get too attached to any one job. I made this mistake a few times, and still ultimately ended up with a job I was very happy to get each time (and which was likely better than the counterfactual offer anyways)</li></ul><h2 id="a-note-on-levels">A note on &quot;levels&quot;</h2><p>One of the great (IMO) features of tech careers is the notion of a &quot;level.&quot; Although they aren&apos;t fully standardized between different companies (and it takes some experience to know what to expect from a &quot;senior&quot; engineer coming from different types of orgs), I still find them a pretty useful framework for understanding jobs in tech. A typical sequence of levels at tech companies (or quant sometimes, for larger firms) looks something like the following, where if you&apos;re on say, a data scientist track, you might replace &quot;software engineer&quot; with &quot;data scientist&quot; everywhere.</p><ul><li>Junior<ul><li>Sometimes can be a title like &quot;Software Engineer 1&quot; or &quot;SDE 1&quot; or &quot;SWE 1&quot;</li><li>Entry level for bachelors</li><li>Very low expectations, your job at this level is mostly to learn. Companies typically view juniors as an investment more than a clear source of value.</li></ul></li><li>Mid<ul><li>Often comes without a title prefix, e.g. just &quot;Software Engineer&quot; and not &quot;Mid-level Software Engineer&quot;</li><li>Otherwise might look like &quot;Software Engineer 2&quot; or &quot;SDE 2&quot; or &quot;SWE 2&quot;</li></ul></li><li>Senior<ul><li>Typically considered a &quot;terminal&quot; role in tech &#x2013; there are many levels beyond this, but this is where you&apos;re considered good enough for companies to be happy to have you sit here indefinitely and provide this level of value without growing further. Many people do just that.</li><li>Typical expected time is ~5 yrs to reach this level (for the median engineer that makes it to senior level, not for the median engineer <em>with a PhD</em> that makes it to senior &#x2013; you can make it to senior much faster if you want but you do not have to)</li><li>When interviewing for this level and above in software, you will start to need to know <a href="https://github.com/Jeevan-kumar-Raj/Grokking-System-Design?ref=georgeyw.com" rel="noreferrer">system design</a>. Expect at least one technical interview to be on system design problems.</li></ul></li><li>Staff<ul><li>This is where there&apos;s an option to move into the management track, which has its own levels that parallel the &quot;individual contributor&quot; (IC) track<ul><li>Most people actually move into management instead of moving up the IC ladder, see notes on that below</li></ul></li><li>It is harder to define what your role is at this level and above, staff+ ICs tend to get increasingly specialized roles based on their individual capabilities and the needs of the company</li><li>Compensation was already stupid, but this is where it gets real dumb. 500k+ total compensation is common at FAANG</li></ul></li><li>Principal<ul><li>Here levels start to break down further at different companies. Sometimes there is a &quot;senior staff&quot; before this level and sometimes there is a &quot;senior principal&quot; after this level. Sometimes staff is skipped, sometimes principal doesn&apos;t exist.</li><li>Compensation gets downright offensive: FAANG principal+ can earn 7 figures</li></ul></li></ul><p>Management track has similar levels (and compensation) starting with &quot;Engineering Manager&quot; or something similar being the equivalent of something like &quot;Staff Software Engineer.&quot; It then has its own titles like &quot;Senior Engineering Manager,&quot; and may go towards titles like &quot;VP of Engineering.&quot; All of these levels (for both IC and manager track) have their own expectations, which I won&apos;t cover here. There are good/comprehensive resources online, like <a href="https://dresscode.renttherunway.com/blog/ladder?ref=georgeyw.com" rel="noreferrer">this one</a>.</p><h3 id="benchmarking-levels-against-academia">Benchmarking levels against academia</h3><p>Ok, here&apos;s where the real hot takes begin. I&apos;m not going to bother justifying these, so take them with a grain of salt, but I think this is what I believe from my limited perspective.</p><ul><li>If you don&apos;t have a PhD in computer science or machine learning, expect to enter tech or quant at the junior or mid-level. You might be able to get a data science role at a senior level if you did e.g. probability and do well on your interviews.<ul><li>The startup I first worked for didn&apos;t have levels in a formal sense until a year after I joined, but I think I was retroactively initially placed at the junior level and then promoted to mid-level after a year.</li></ul></li><li>If you can get a PhD in math, you can reach senior level in tech, full stop. Tech jobs are high paying because 1) tech companies make a shitload of money, and marginal engineers really do actually bring a lot of value to the company at scale and 2) because the jobs are competitive and hard and have high expectations. But crucially, they are competitive and hard and have high expectations <em>relative to the typical job in America</em>. They are <em>not</em> more competitive or harder or have higher expectations than earning a PhD (modulo personal fit for tech, some people may just find software easier than others).</li><li>Postdocs up through tenured professors are probably somewhere around the &quot;staff&quot; level equivalent for academia.</li><li>I&apos;m less sure about this, but I&apos;d guess that full professor in academia is comparable to principal level in tech or to the equivalent management track role (hey, remember that 7 figure compensation number above?)</li><li>These don&apos;t mean they&apos;re where your <em>cap</em> is, but I would confidently place money at 1:1 odds (and probably much worse odds, though I don&apos;t want to commit to figure those out here) that someone with the given level of academic accomplishment can first reach the given comparable tech level in a typical amount of time</li><li>In general, the number of people who reach each progressive level in the tech ladder tends to decrease exponentially, and the difficulty and expectations of the levels also increase substantially past senior. It is <em>totally fine</em> to hit ~senior level and just chill there forever and do other stuff with your time, like fun side projects or your math research or just having more time to enjoy life in general.</li></ul><h2 id="a-personal-request">A personal request</h2><p>If you&apos;ve gotten this far, hopefully this resource has been useful to you. I am happy if this improves your life in any way. My only request to you is that you please try to avoid working <em>directly</em> on AI capabilities as a result of this advice.</p><p>I&apos;ve <a href="https://www.georgeyw.com/slowing-down-pandoras-box/" rel="noreferrer">written about</a> AI safety previously and have <a href="https://www.georgeyw.com/slowing-down-pandoras-box/" rel="noreferrer">pivoted my career</a> to technical safety research because I think this is such an important problem. This is not asking you to say, turn down the only offer you get and not be able to support yourself or your family, but to please think about this if you find yourself having a choice in the matter. Most job opportunities don&apos;t fall under this umbrella, and it&apos;ll probably be clear if a given job is aimed at significantly advancing AI, so you mostly have to be specifically looking for such jobs. </p>]]></content:encoded></item><item><title><![CDATA[Sharing is caring: in defense of ordinary things]]></title><description><![CDATA[On foggy mornings and small talk]]></description><link>https://www.georgeyw.com/sharing-is-caring/</link><guid isPermaLink="false">66a51b54e68b640a7230caf3</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sun, 19 Nov 2023 17:53:11 GMT</pubDate><content:encoded><![CDATA[<p>I visited Oxford last week for a conference-workshop-retreat-thing (working title), which means I&apos;m super jet lagged this week and have been temporarily consigned to the life of a morning person. I&apos;m normally an extreme night owl, and the rare times I&apos;m awake at daybreak are at the wrong end of the day. So when I begrudgingly got out of bed at 6am this morning (a notable improvement from 3am), I figured I&apos;d go to the park and have my coffee with the sunrise for once*.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.georgeyw.com/content/images/2024/07/20231116_071113.jpeg" class="kg-image" alt loading="lazy" width="2000" height="1500"><figcaption><span style="white-space: pre-wrap;">nice, quiet Silent Hill vibes</span></figcaption></figure><p>Dear reader, it was foggy and overcast. There was no sunrise. I&apos;m not upset, but thanks for asking.</p><p>While I was walking, I thought back to the conversations I had in Oxford and to previous work retreats before that. I&apos;ve only experienced remote work, and a lesson that I keep relearning is how impactful meeting in-person is. It&apos;s clearly helpful for productivity &#x2013; for about a week, everyone&apos;s primary focus is on planning, on sharing knowledge, on aligning the big picture. But I don&apos;t think this is actually the most important part. There&apos;s a key social cohesion created, and not just from intentionally designed icebreaker activities. </p><p>When you first interact with someone, you get the broad strokes that everyone gets. He&apos;s a teacher, she&apos;s an engineer, they&apos;re a parent, a spouse, a sibling. Work intros often start with everyone rattling off education history, past jobs, and maybe a well-rehearsed fun fact. Basic knowledge is of course a necessary starting point, but how much of a picture do you really get of someone from that?</p><p>These are the things that often come to mind when we think of our identity, but billions of other people in the world are, say, cat or dog people, and I&apos;d take short odds that there&apos;s more variance within one of those groups than between them. You can learn a dozen facts about a person and still be no closer to understanding them or building a close relationship.</p><p>Most people never see the the small, mundane, in-between moments in our lives, but I think this is where intimacy lies. I don&apos;t identify with a foggy morning; it doesn&apos;t feel like a part of who I am in any way. And yet, I&apos;d only share this kind of story with a few people (and you of course, dear reader). I don&apos;t broadcast this to the world (kindly ignore that I&apos;m publicly writing about it) because it isn&apos;t that important, and most people just won&apos;t care. For the people that do, they care because it&apos;s a story about <em>me</em>, because someone in their tribe experienced it, and because it means something to be one of the first people that comes to mind when anything happens.</p><p>I&apos;m not saying that spending a week in person gets you anywhere near that level of closeness. But having such dense overlap in organic moments sure seems like a bigger step forward than as many video calls to hash out business logistics. There&apos;s no choice but to experience the ordinary together and much more space to fill with mundane questions (how was your morning?) and mundane stories (the funniest thing happened...). Maybe there&apos;s something like the <a href="https://en.wikipedia.org/wiki/Ben_Franklin_effect?ref=georgeyw.com" rel="noreferrer">Ben Franklin effect</a> happening too &#x2013; if sharing is caring, then perhaps caring follows sharing as much as the reverse.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.georgeyw.com/content/images/size/w1000/2024/07/Untitled_Artwork.png" class="kg-image" alt loading="lazy" width="1000" height="944"><figcaption><span style="white-space: pre-wrap;">the frontier of sharing advances or recedes with closeness (</span><a href="https://twitter.com/RichDecibels/status/1716140468751204826?ref=georgeyw.com"><span style="white-space: pre-wrap;">h/t</span></a><span style="white-space: pre-wrap;"> for format)</span></figcaption></figure><p>I also notice that a drop-off in mundane sharing predicts an eventual drop-off in closeness**. I think this makes sense; most of what happens in our lives isn&apos;t news-worthy, so not talking about the boring stuff means much less contact overall. But beyond this, there&apos;s probably something powerful in gradually losing the urge to share and in the worry that you might be bothering someone with something uninteresting rather than feeling secure in the idea that anything is interesting to them if you&apos;re the main character. Sometimes the last things to develop are the first to go.</p><hr><p>*Technically this happened a few days ago by time of publishing, but I wrote that line on the day it happened. Don&apos;t sue me.</p><p>**I&apos;m also noticing while writing that I&apos;m generally not sharing much of the boring stuff with <em>anyone</em> right now (so for my close friends reading this, don&apos;t go off thinking I secretly hate you in particular because I didn&apos;t tell you about a non-sunrise). This seems like a sign that I&apos;m getting too isolated and need to give more attention to my social wellbeing.</p>]]></content:encoded></item><item><title><![CDATA[Moving to AI safety]]></title><description><![CDATA[Kicked up some dust this year, some was settled]]></description><link>https://www.georgeyw.com/moving-to-ai-safety/</link><guid isPermaLink="false">66a51b54e68b640a7230caf2</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Tue, 14 Nov 2023 16:44:00 GMT</pubDate><content:encoded><![CDATA[<figure class="kg-card kg-gallery-card kg-width-wide kg-card-hascaption"><div class="kg-gallery-container"><div class="kg-gallery-row"><div class="kg-gallery-image"><img src="https://www.georgeyw.com/content/images/2024/07/boat-sunset.jpg" width="2000" height="2666" loading="lazy" alt></div><div class="kg-gallery-image"><img src="https://www.georgeyw.com/content/images/2024/07/boat-whiteboard-blurred.png" width="1700" height="1536" loading="lazy" alt></div><div class="kg-gallery-image"><img src="https://www.georgeyw.com/content/images/2024/07/fall-leaves.jpg" width="2000" height="2667" loading="lazy" alt></div></div></div><figcaption><p><span style="white-space: pre-wrap;">Regrettably no photos of laser tag in a castle - have a picture of a lecture on a lake instead</span></p></figcaption></figure><p>It&apos;s been an eventful year! I quit my job, tried to start a company, had a summer fling with freelancing (it was awesome), and did a ton of soul searching. It&apos;s also been a bit of an emotional rollercoaster: lots of stress, lots of confusion, lots of fun (from New England sunset boat rides to laser tag in a <a href="https://www.wythamabbey.org/?ref=georgeyw.com" rel="noreferrer">castle</a>). The year&apos;s not over yet, so there&apos;s still time for curveballs, but it looks like some dust is finally settling.</p><p>The dominating consequence of my soul search was the conviction that AI safety is the <a href="https://www.lesswrong.com/posts/P5k3PGzebd5yYrYqd/the-hamming-question?ref=georgeyw.com" rel="noreferrer">most important thing I can work on</a> right now. I spent the better part of the last 6 months exploring how rapid AI development will transform our world, and the world in turn kept smacking me in the face with <a href="https://www.georgeyw.com/slowing-down-pandoras-box/" rel="noreferrer">how wide the cone of possibilities</a> was. Unfortunately, not all outcomes are positive &#x2013; I now believe that this is the <a href="https://www.lesswrong.com/s/yYxggfHYRrqnJXuRx?ref=georgeyw.com" rel="noreferrer">most perilous century</a> for humanity, with AI being the primary driver. </p><p>The hype around AI &#x2013; that it could solve cancer, poverty, conflict, disease, even aging and death &#x2013; massively accelerates its development. To be clear, AI safety shouldn&apos;t be <a href="https://a16z.com/the-techno-optimist-manifesto/?ref=georgeyw.com" rel="noreferrer">conflated</a> with the belief that approaching <a href="https://www.lesswrong.com/posts/Py3uGnncqXuEfPtQp/interpersonal-entanglement?ref=georgeyw.com" rel="noreferrer">Reedspacer&apos;s Lower Bound</a> is inherently bad. I hope we cure all disease! I hope we get functionally infinite prosperity! I am <em>so</em> on board with solving death! It would be tragic if we <em>never</em> achieved safe artificial superintelligence.</p><p>But hype breeds recklessness. Almost all of our forecasted extinction risk this century <a href="https://www.metaculus.com/questions/2568/ragnar%25C3%25B6k-seriesresults-so-far/?ref=georgeyw.com#comment-133441" rel="noreferrer">comes from AI</a> &#x2013;  many times more than from nuclear war, climate change, and bio-risk combined. And development goes on accelerating anyways. I want all the great things that AI promises, I just think we can get them for a better price. </p><figure class="kg-card kg-image-card"><img src="https://www.georgeyw.com/content/images/2024/07/image.png" class="kg-image" alt loading="lazy" width="481" height="147"></figure><p>So what&apos;s next for me? I&apos;m excited to work towards a safer future with <a href="https://www.lesswrong.com/posts/nN7bHuHZYaWv9RDJL/announcing-timaeus?ref=georgeyw.com" rel="noreferrer">Timaeus</a>, a brand spankin&apos; new technical AI safety research org. I&apos;m hopping on board to help scope out a new field of <a href="https://www.lesswrong.com/tag/interpretability-ml-and-ai?ref=georgeyw.com" rel="noreferrer">interpretability</a> research as a research assistant/engineer for at least the next ~6 months. I&apos;m also frantically cramming Spanish for the <a href="https://aifuturesfellowship.org/?ref=georgeyw.com" rel="noreferrer">AI Futures Fellowship</a> in Mexico City early next year, where I&apos;ll hopefully develop a broader base of knowledge, meet brilliant people, and eat far too much food.</p><h3 id="the-future-is-unclear">The future is unclear</h3><p>I think more people should spend time thinking hard about AI. Whatever consequences you think there will be, it seems hard to get away from the conclusion that they&apos;ll be transformative in <em>some</em> direction. But <a href="https://twitter.com/littIeramblings/status/1708945586496446796?ref=georgeyw.com" rel="noreferrer">sensemaking</a> about AI is <a href="https://michaelnotebook.com/xrisk/index.html?ref=georgeyw.com" rel="noreferrer">really hard</a>, especially if you don&apos;t have a technical background. Even as someone with the relevant background, I spent months highly confused about the right path to take and still feel fairly uncertain. We can and should at least still strive to be directionally correct though, and voting for safe policies is just as critical as technical research (if not more). </p><p>Finally, I think this is the ~correct direction (or <em>a</em> correct direction) for me to be moving in right now, but it&apos;s all pretty nuanced, and I&apos;m always open to more perspectives. Please reach out if you&apos;re curious or have thoughts you&apos;d like to share. Don&apos;t be a stranger, odds are I&apos;d love to chat &#x2013; everyone who&apos;s let me talk their ear off so far can attest to that.</p>]]></content:encoded></item><item><title><![CDATA[Slowing down Pandora's Box]]></title><description><![CDATA[Editorial for the Canadian Science Policy Centre]]></description><link>https://www.georgeyw.com/slowing-down-pandoras-box/</link><guid isPermaLink="false">66a51b54e68b640a7230caf1</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Tue, 22 Aug 2023 04:53:17 GMT</pubDate><content:encoded><![CDATA[<p><em>Originally written for an </em><a href="https://sciencepolicy.ca/posts/slowing-down-pandoras-box/?ref=georgeyw.com"><em>editorial series</em></a><em> for the Canadian Science Policy Centre.</em></p><hr><p>This thought keeps coming back to me lately: <em>everything is about to get really weird</em>. For the past two years, I&#x2019;ve had a front-row seat to the primordial soup of AI research and startups. I&#x2019;ve built a couple of products based on language models, consulted on generative music tools, and dabbled across the board of generative AI, including image, audio transcription, voice, and more. A massive wave of commercial products is coming, for which ChatGPT is just the tip of the iceberg.</p><p>AI has been fairly consistent in progressing exponentially. The thing about exponential growth is that for a long time, it looks like nothing is happening, then it looks like something is finally happening, and then while you&#x2019;re still processing the something that&#x2019;s happening, the growth explodes and&#x2026; uh oh, suddenly everything is weird now. We&#x2019;re still in the &#x201C;huh, it looks like something is finally happening&#x201D; stage with AI. </p><p>A certain recent pandemic demonstrated that we&#x2019;re not so good at dealing with the next phase of exponential growth. But covid didn&#x2019;t care what was or wasn&#x2019;t intuitive to us; the world just changed overnight anyway. We scrambled to react and flatten the curve, but not before it was already a foregone conclusion that there <em>would</em> be a huge curve. When it comes to AI, it&#x2019;s fortunately not all downside like it is with a global pandemic. AI has incredible potential to bring prosperity &#x2013; some have even compared it to the discovery of fire or electricity. But there&#x2019;s a darker side, and that side grows exponentially too.</p><p>While writing this, I hopped on X (formerly known as Twitter, RIP) and saw a post on <a href="https://twitter.com/nearcyan/status/1687945164378058752?ref=georgeyw.com">practical acoustic side-channel attacks on keyboards</a>. The model in that paper can correctly identify 93% of your keystrokes from the audio that a Zoom call picks up. I see news like this more often than is comfortable, and then I predictably think to myself that everything is about to get really weird. You have to really stand out these days to be noteworthy &#x2013; a few months ago, a research team managed to <a href="https://medarc-ai.github.io/mindeye/?ref=georgeyw.com">read the human mind</a>(!) with &gt;90% accuracy by training a model on fMRI scans. [<em>Note: since writing this, we&apos;ve managed to </em><a href="https://news.berkeley.edu/2023/08/15/releases-20230811?ref=georgeyw.com"><em>reconstruct music</em></a><em> by reading brainwaves too.</em>]</p><p>The optimistic among us would hope in vain that AI tech isn&#x2019;t abused. FraudGPT and WormGPT are two cousins of ChatGPT, repurposed for enabling cyberattacks and scams. ChaosGPT was a (fortunately ineffective) bot whose task was literally to destroy humanity. What happens as we progress along the exponential growth curve and the people building these harmful AIs get access to shiny new toys? Dario Amodei, CEO of Anthropic, <a href="https://www.judiciary.senate.gov/imo/media/doc/2023-07-26_-_testimony_-_amodei.pdf?ref=georgeyw.com">testified to the US Congress</a> that a straightforward extrapolation of today&#x2019;s systems suggests that large-scale biological attacks could be possible in only two or three years. He <a href="https://twitter.com/profoundlyyyy/status/1684333960753565696?ref=georgeyw.com">went on to say</a> that if we don&#x2019;t have mechanisms in place to restrain AI systems by 2025-26, &#x201C;we&#x2019;re gonna have a really bad time.&#x201D;</p><p>Two to three years isn&#x2019;t a lot of time, and we don&#x2019;t even know what the right restraining mechanisms are yet! It would be great to have a bit more time to figure things out before the exponential progress curve blows past us. Fortunately, unlike with the pandemic, we can actually make that happen for AI. The Future of Life Institute wrote an <a href="https://futureoflife.org/open-letter/pause-giant-ai-experiments/?ref=georgeyw.com">open letter</a> with over 33,000 signatures calling for a six-month moratorium on AI development, signed by many of the biggest names in AI research and tech. It&#x2019;s not clear that the specific plan of a six-month moratorium would be most effective, but this seems directionally correct. Policymakers might consider prophylactic legislation that would allow such a moratorium to come into effect if certain conditions are met. Compute governance is a similar slowing mechanism, where large training runs of AI models are limited in the amount of compute power they&#x2019;re allowed to use. The threshold can be adjusted over time to have some control over the pace of development without it being all or nothing. These aren&#x2019;t long-term solutions, but they would buy time to figure out more permanent measures.</p><p>Besides short timelines to serious risks, another core challenge with making AI safe is the asymmetry between offense and defense. There are simply too many angles of attack to be able to cover all vulnerabilities &#x2013; but policy can reshape the AI landscape into something more defensible. When advanced AI is too openly accessible, any individual can choose to take one of those many angles of attack, so an obvious solution might be to prevent advanced AI from being too openly accessible. </p><p>We can accomplish that via regulatory action, but this is a pretty controversial topic in the tech world; open-sourcing (making publicly available) your research and your code is usually synonymous with nice things such as transparency, collaboration, and even security since you have more eyes on fixing vulnerabilities. Furthermore, a common counterargument to closing off access to AI technology is that it centralizes too much power in the hands of the few major AI labs and that we can only have a fair balance of power if the technology is available to everyone. </p><p>But the world is not always a safer place when people have more access to technology. The US has a lot of guns and &#x2013; surprise &#x2013; <a href="https://www.nytimes.com/2017/11/07/world/americas/mass-shootings-us-international.html?ref=georgeyw.com">a lot of mass shootings</a>. If everyone had a nuke in their pocket, it would only be a matter of time before most urban cities disappeared in a mushroom cloud. All it takes is one person doing the wrong thing. And if everyone has a large-scale-biological-attack-capable AI on their home computer, Dario Amodei put it well: we&#x2019;re gonna have a really bad time.</p><p>It doesn&#x2019;t seem likely that we can entirely close Pandora&#x2019;s box on AI, but we can make it manageable. Everything is about to get really weird, but we still have the power to decide if that&#x2019;ll be a good weird or a bad one.<br></p>]]></content:encoded></item><item><title><![CDATA[My PhD cost me $1.17M (but I'd do it again)]]></title><description><![CDATA[PhD advice + reflections on opportunity cost ]]></description><link>https://www.georgeyw.com/my-phd-costed-me-1-17m-so-far-but-id-do-it-again/</link><guid isPermaLink="false">66a51b54e68b640a7230caef</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sun, 02 Jul 2023 20:07:47 GMT</pubDate><content:encoded><![CDATA[<p>Not from my bank account, just from opportunity cost. It&#x2019;s the first time I&#x2019;ve sat down and calculated that. I don&#x2019;t think I was ready for the sticker shock, I just thought it could make for a catchy title. Honestly, this might say as much about big tech salaries as it does about PhDs, but the counterfactual history was a reasonable one where I took a job at Google and took a standard path to senior engineer.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh6.googleusercontent.com/shqDEBVd58zG1q1AFGk43NuC5iw2UTK2lMRh6aPxzmCUesKSEXpjbdem5letzILQG_P1NBMHQhleW7zKaKLSUQWDWYGaV_UIPm493z1FZHTfWxA3CzspY-fDQEmUzoL9VXx1YUO9Ev8Y3P0BXlEkCxA" class="kg-image" alt loading="lazy" width="561" height="312"><figcaption><i><em class="italic" style="white-space: pre-wrap;">Current Google Senior SWE compensation, courtesy </em></i><a href="http://levels.fyi/?ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">levels.fyi</em></i></a></figcaption></figure><p>To get that number, I made some salary approximations, calculated taxes and cost of living, and was naively surprised at <a href="https://www.investopedia.com/ask/answers/042415/what-average-annual-return-sp-500.asp?ref=georgeyw.com">S&amp;P 500 returns</a> in a zero interest rate environment, which managed to account for a substantial amount of the difference. The total gap will increase over time too, through the interminable power of compounding returns. Your ultimate career path might not have quite such inflated salaries, and zero interest rates probably aren&#x2019;t coming back any time soon, but the opportunity cost for a PhD may still be much higher than you expect.</p><p>Some people hold the position that they should do their best to deter anyone and everyone they can from getting a PhD. Their belief is that you&#x2019;re fit for one if and only if you ignore that advice and boldly forge ahead anyways.</p><p>I think this is unnecessary. There are plenty of real reasons not to do a PhD (I have at least 1.17M of them), and one should have a better reason to do one than an inability to absorb advice. These considerations, and my own story, are the best I can offer to you if you&apos;re a prospective student; the final decision is up to you.</p><h2 id="the-happy-path-of-grad-school">The happy path of grad school</h2><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh5.googleusercontent.com/MrP9MkK0xrLlWYvtLsw_GbNMrWQKtvTlXkgZZ5E6Hyl6DRzkSjsHuQ8Z9E4sdNOCUuzoZSBd9GsW9_oRVD_AOSSPcU8xQvGz3dtPxJ29S23ODbztv1JrWBe1YMswN_k7L6AVgK0CZEIneChsPSz3UZs" class="kg-image" alt loading="lazy" width="223" height="311"><figcaption><a href="https://gatherer.wizards.com/Pages/Card/Details.aspx?multiverseid=394746&amp;ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">This</em></i></a><i><em class="italic" style="white-space: pre-wrap;"> could be you! (Fun fact: </em></i><a href="https://en.wikipedia.org/wiki/Richard_Garfield?ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">Richard Garfield</em></i></a><i><em class="italic" style="white-space: pre-wrap;">, the creator of Magic: The Gathering, went to the same grad program as me. No wonder he&#x2019;s so successful!)</em></i></figcaption></figure><p>Whatever path you choose in life, you&#x2019;ll take some form of opportunity cost along the way; the important thing is just to notice the tradeoff. For example, last I checked, it was quite expensive to raise children in the US. Most people go on to have children anyways, and many go on to be happier with their kids than the money would make them. An econ PhD friend of mine talks about a concept of &#x201C;Life <a href="https://en.wikipedia.org/wiki/Expected_value?ref=georgeyw.com">EV</a>&#x201D;: sometimes we do things that are financially suboptimal because they sufficiently improve our life along other dimensions (he often yells this as he&apos;s halfway through leaping into a clearly bad decision because it seems fun).  </p><p>But just as you shouldn&#x2019;t have kids for the sole reason of hoping you have someone to visit you in your nursing home, you shouldn&#x2019;t only think of a PhD in terms of what comes after. The best reason to do a PhD, and one which is both necessary and sufficient, is that you simply want to experience the journey.</p><p>That journey can be a beautiful one, if you&#x2019;re <a href="https://www.georgeyw.com/you-dont-take-enough-risks-to-be-lucky/">lucky</a>. You&#x2019;ll meet other like-minded, bright-eyed grad students. You&#x2019;ll meet generous mentors who pay things forward like it is their moral duty. You&#x2019;ll travel to conferences (on someone else&#x2019;s dime!) <a href="https://fpsac.org/confs/?ref=georgeyw.com">across the world</a>, and you&#x2019;ll meet more people that care about your <a href="https://matt.might.net/articles/phd-school-in-pictures/?ref=georgeyw.com">tiny corner of reality</a> than you could have imagined existed (there were <em>dozens</em> of us). </p><p>You can work on the kind of problem that follows you to the shower. The kind that itches your brain at night. The kind that jolts you awake as you lay in bed, because a piece finally clicked into place and you can&#x2019;t hope to sleep until you write it down. The kind that gives you an adrenaline high for days when you figure it out, because you&#x2019;ve been working on it for months or years.</p><p>There are other reasons that help, too. It&#x2019;s hard to believe you can tackle a challenge that no one else has, especially if the pressure of completing your thesis hangs over you. A PhD is training and proof of your ability to grapple with the unknown. </p><p>You might want to do something you can only do with a PhD, or which a PhD can open doors for. In the world of math, this could mean working at a university, but it could also mean working at a hedge fund, at a national lab, or at an AI lab of a big tech company. Most things don&#x2019;t need a PhD in the strictest sense, but people do pay attention when you have one.</p><p>These things are icing on the cake, but at the end of the day, the surest way to make it a positive, worthwhile experience is to <em>want the experience</em>. A PhD should not feel like paying your dues, and it is not noble to suffer through one. An important corollary is that if you are decidedly <em>not</em> enjoying the experience, it&#x2019;s ok to leave!</p><h3 id="a-note-on-slack">A note on &quot;slack&quot;</h3><p>It can be a little paradoxical to describe grad school as a time with a lot of <a href="https://slatestarcodex.com/2020/05/12/studies-on-slack/?ref=georgeyw.com">slack</a>. After all, grad students have a reputation for being over-worked, under-paid, over-caffeinated, and under-rested. But it isn&apos;t always like that, and there is a unique thread of intellectual freedom through all of it.</p><p>For me, that freedom manifested as space to grow. Once I finished most of my coursework (first two years), the only expectation seemed to be that I spend a lot of time thinking. In theory, that means thinking about your research, but in practice, my mind still managed to wander for many hours a day. </p><p>That space expanded considerably in my last two years; the pandemic happened, I&apos;d already decided to leave academia, and I had enough research to graduate. I had few expectations to fulfill, and any remaining research work was mainly tying up loose ends. This was an unusual abundance of slack, but it led to substantially more hours being alone in my head. </p><p>I used to think this time was a bit wasted. The amount of concrete things that came out of those last couple of years felt pretty dismal in number. I was probably burnt out for a substantial portion of it, and the social isolation was hard.</p><p>Looking back, those years were probably some of the richest years of personal growth that I&apos;ve ever had. Among other things, I developed a lot more self-awareness and confronted large chunks of emotional baggage that I never previously noticed.</p><p>This constitutes a large part of why I would make the same choice to do a PhD, if I could go back and choose again. It also highlights another kind of opportunity cost, which is of time. Whether I spent 5 years in grad school or at Google, those 5 years would pass either way. I only got to spend my early 20s once, and I&apos;m glad that I spent them the way I did. Just as money in the bank would compound year over year, I can benefit from that space to grow for the rest of my life.</p><h2 id="the-darker-side-of-grad-school">The darker side of grad school</h2><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh3.googleusercontent.com/hmwn6EPkFGl1pRNWskNMMkhA-K3EMp-qvJxoi_8n-gnU_I-GC6Yg3KAN-iE0yuaadMGQ-3nq13dPVjSLm4mjYgpw1AQWdhtoZj4X46sfRBLGy4x3D32iLpffSCjTMuO6-c4d1lra6DoxMIr4bnewaX0" class="kg-image" alt loading="lazy" width="265" height="370"><figcaption><a href="https://gatherer.wizards.com/Pages/Card/Details.aspx?multiverseid=547907&amp;ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">This</em></i></a><i><em class="italic" style="white-space: pre-wrap;"> could be you too!</em></i></figcaption></figure><p>The PhD journey can be a beautiful one, if you&#x2019;re lucky. If you&#x2019;re unlucky, academia is a <a href="https://www.benkuhn.net/grad/?ref=georgeyw.com">meat grinder</a>. There are various opportunity costs to consider, but there are real costs too, mostly in the form of stress to the mind and spirit. From my limited view, the following things were the main culprits.</p><h3 id="impostor-syndrome">Impostor syndrome</h3><p>Academia has a high density of brilliant people. Many academics, including ones you will think are far more brilliant than yourself, <a href="https://en.wikipedia.org/wiki/Impostor_syndrome?ref=georgeyw.com">feel they don&#x2019;t belong</a>. They do, and you will too if you choose to join them. There is an entire post that can be written about this topic, but the short version is that it is important to know that you may feel this way, to know that it is a normal and well-worn path, and to know that there are resources for this.</p><h3 id="social-isolation">Social isolation</h3><p>It can be intoxicating to be consumed by your work, but the darker side is that you can end up disconnected from and left behind by your social network. PhD programs are selective, so you can&#x2019;t be picky about which city you end up in. I think this is especially hard for overseas students, where even phone calls are difficult to coordinate thanks to time zones. </p><p>You&#x2019;ll see other people moving on to the next stage of their life, starting their careers, starting families, buying houses, cars, vacations, and other shiny things. Your family and friends may valiantly try to understand your work, but they probably won&#x2019;t. Your cohort might end up a bit south of social and outgoing (n.b. you are <a href="https://notebook.drmaciver.com/posts/2020-03-17-11:40.html?ref=georgeyw.com">allowed</a> to be the one that organizes social events!).</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh6.googleusercontent.com/bgoPjg9zskSnJEivHUIHDispRDEsTSBd-OxfjOoHhTv_ssXZ-b6WKfytrRB3KFnN61Pe-UoW8FOyi_GkG54uvCAujvp1IxPNTbkZllrMMg5O4OIRQNyd4evoG5z69zjt0tsISxIVk0iIxMwlGTytDjE" class="kg-image" alt loading="lazy" width="443" height="425"><figcaption><i><em class="italic" style="white-space: pre-wrap;">Pictured: one of the reasons I made it through in one piece</em></i></figcaption></figure><p>My advice for this is to find a hobby with a strong community around it (preferably a physical one; you get exercise and get to meet people in person). I took up rock climbing in my first year, and it changed my life.</p><h3 id="a-bad-advisor-relationship">A bad advisor relationship</h3><p>Professors are people too, and in any group of people there will be unsavory individuals. Some rare professors have a reputation for dealing in politics, gossip, and lies. It is not always easy for a fresh grad student to figure out who these are, but one should do everything in their power to avoid such professors. </p><p>Some other professors have very particular (and antagonistic) approaches to mentorship. One who shall not be named would intentionally fail students on their <a href="https://en.wikipedia.org/wiki/Comprehensive_examination?ref=georgeyw.com">quals</a>. Their alleged motivation for doing so was to use the pressure to pass quals as a motivational whip or <a href="https://en.wikipedia.org/wiki/Forcing_function?ref=georgeyw.com">forcing function</a> on their students. By doing so, they could compel their students to study additional foundational material for the second attempt. Think twice before working with someone like this.</p><p>Setting aside extreme examples, you might just not get along well with your advisor through no fault of yours or theirs. Working chemistry is fickle, and you&#x2019;re allowed to decide it isn&#x2019;t a good match. Shop around for advisors aggressively until you find a fit (or at least do the best you can with the size of your department), since from the point of picking an advisor onwards, they&#x2019;ll be the biggest factor in your journey. It can be worth picking your second favorite topic if it means working with your favorite professor.</p><h3 id="academia-or-bust">Academia or bust</h3><p>One of my undergrad math professors once told me how they got their job. While preparing to defend their thesis, their advisor got a phone call from a colleague asking if he had any good students finishing soon. That was it. No interview, no <a href="https://en.wikipedia.org/wiki/Postdoctoral_researcher?ref=georgeyw.com">postdoc</a> needed (typically a 3 year contract), they just discovered the next day that they had a job waiting for them. That was some 60 odd years ago. </p><p>Today, you&#x2019;ll fight to earn 1-2 postdocs, then fight even harder to earn a tenure track position, then work on earning tenure (usually a 6 year process). Although it&#x2019;s uncommon to be denied tenure once you have a tenure-track position, that&#x2019;s still a total of ~16 years from the start of grad school before you can be completely sure of stable employment. A <a href="https://forum.effectivealtruism.org/posts/3TQTec6FKcMSRBT2T/estimation-of-probabilities-to-get-tenure-track-in-academia?ref=georgeyw.com">pretty low</a> (but not insignificant) portion of PhD students that want tenure eventually make it.</p><p>I believe in chasing dreams; if tenure is your dream, then aim for tenure. Just don&#x2019;t hinge all of your expectations and happiness on that outcome. Luck is enough of a factor that brilliancy won&#x2019;t inoculate you. <a href="https://en.wikipedia.org/wiki/Yitang_Zhang?ref=georgeyw.com">Yitang Zhang</a> proved a <a href="https://terrytao.wordpress.com/career-advice/on-the-importance-of-partial-progress/?ref=georgeyw.com">partial result</a> of one of the most <a href="https://en.wikipedia.org/wiki/Twin_prime?ref=georgeyw.com#Twin_prime_conjecture">famous open problems</a> in math in 2013. He was incredibly persistent, but prior to this major result, still found himself living out of his car and working at Subway (I think this is exceptionally <em>unlucky </em>though).</p><p>At the end of the day, there are plenty of other <a href="https://en.wikipedia.org/wiki/Deep_tech?ref=georgeyw.com">places</a> <a href="https://www.nasa.gov/?ref=georgeyw.com">you</a> <a href="https://nationallabs.org/?ref=georgeyw.com">can</a> <a href="https://en.wikipedia.org/wiki/Big_Tech?ref=georgeyw.com">do</a> <a href="https://en.wikipedia.org/wiki/Hedge_fund?ref=georgeyw.com">research</a> <a href="https://en.wikipedia.org/wiki/Secondary_school?ref=georgeyw.com">or</a> <a href="https://en.wikipedia.org/wiki/Community_college?ref=georgeyw.com">teach</a>.</p><h3 id="unexpected-regrets">Unexpected regrets</h3><p>Few journeys end with no regrets. 5 years is a long time for regrets to find you, and they may not be what you expect. I am happy to have done a PhD, and I wouldn&#x2019;t change that decision. But my biggest regret was that it kept me away from home for what ended up being the last years of our family cat&#x2019;s life. She unexpectedly passed away shortly after defending my thesis, and before I could make it home to say goodbye.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh3.googleusercontent.com/9x_BFoRRyIhnEnnom7l080--Jp3MWvltKjgKVbNaz3vTRayrxqmm4fPC8oJBUQBHeVIf0GZ-mF8HsMyOW0ZfPwo0LbrxNVW4N2c2jcM9M8ASj9VzyYV7g9LlYYshCOpm7ueeFB1Ole839VPXBpGYGgw" class="kg-image" alt loading="lazy" width="598" height="83"><figcaption><i><em class="italic" style="white-space: pre-wrap;">From </em></i><a href="https://repository.upenn.edu/edissertations/4098/?ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">my thesis</em></i></a></figcaption></figure><hr><h2 id="appendix-miscellaneous-advice-other-articles">Appendix: miscellaneous advice + other articles</h2><h3 id="external-links">External links</h3><ul>
<li>Read Terry Tao&#x2019;s <a href="https://terrytao.wordpress.com/career-advice/?ref=georgeyw.com">career advice</a>, it generalizes well. Some standouts for me:
<ul>
<li><a href="https://terrytao.wordpress.com/career-advice/does-one-have-to-be-a-genius-to-do-maths/?ref=georgeyw.com">Does one have to be a genius to do maths?</a></li>
<li><a href="https://terrytao.wordpress.com/career-advice/continually-aim-just-beyond-your-current-range/?ref=georgeyw.com">Continually aim just beyond your current range</a></li>
</ul>
</li>
<li>Sara Billey&#x2019;s advice on <a href="https://sites.math.washington.edu//~billey/advice/timely.fashion.pdf?ref=georgeyw.com">getting a PhD in a timely fashion</a>.
<ul>
<li>Choose your advisor systematically. Talk to current and former grad students. Are they happy? Do they have the kind of job that you want? Do they have jobs <em>at all</em>?</li>
<li>20 hours per week of focus time is enough. Hours in front of the TV don&#x2019;t count. Hours interrupted by texting and checking your phone don&#x2019;t count.</li>
</ul>
</li>
<li>John Baez&#x2019;s <a href="https://math.ucr.edu/home/baez/advice.html?ref=georgeyw.com">advice for young scientists</a>.</li>
<li>A great <a href="https://twitter.com/overthink_pod/status/1670952255217086464?ref=georgeyw.com">list of questions</a> before starting a philosophy degree.</li>
<li>A <a href="https://matt.might.net/articles/books-papers-materials-for-graduate-students/?ref=georgeyw.com">list of resources</a> by Matt Might.</li>
</ul>
<h3 id="on-teaching">On teaching</h3><ul><li>Put your heart into it when you do it, because you have a responsibility to your students, but set reasonable boundaries, because you have a responsibility to yourself, too.</li><li>It&#x2019;s easy to put infinite time into teaching. It feels productive compared to slamming your head against a research problem with no visible progress. Resist this urge. In the language of the <a href="https://asana.com/resources/eisenhower-matrix?ref=georgeyw.com">Eisenhower Matrix</a>, teaching is frequently important and urgent, while research is frequently important but not urgent. The latter category contributes most to long-term success.</li><li>Grade efficiently. If you can choose the grading rubric, simpler point rubrics are easier and faster to grade. The difference between an answer that is 2 out of 3 points and 3 out of 3 points is much easier to spot than 17 or 18 out of 20.</li><li>Grade fairly and predictably. Students <em>will</em> compare grades and answers and bring papers back to you with complaints (even if you grade perfectly, students will do that anyway, it&apos;ll just cost you less time).</li></ul><h3 id="other-notes">Other notes</h3><ul><li>Develop connections with people outside of your department. If you&apos;re gunning for academic positions, aim for at least one recommendation letter from someone outside. Your advisor can and should help you via warm intros.</li><li>People who brag about how much they work don&#x2019;t work as much as they say they do. Don&#x2019;t let it get in your head, and don&#x2019;t try to catch up. Grad school is a marathon and not a sprint.</li><li>People who brag about how <em>little</em> work they do often work more than they say they do. Some people really are built different, but for the most part, it isn&apos;t the norm to work very little and make visible progress.</li><li>There&#x2019;s some folk wisdom that you should go to a different institution at each stage of your career. The idea behind this is that each department (or research group/lab) has its own school of thought. This might hold true beyond grad school, but it&#x2019;s overblown at the undergrad -&gt; grad transition. Don&#x2019;t be afraid to stay at your undergrad institution for your PhD.</li><li>Prestige isn&#x2019;t everything, but <a href="https://www.insidehighered.com/news/2022/09/23/new-study-finds-80-faculty-trained-20-institutions?ref=georgeyw.com">it still matters more than it should</a>. If you want to stay in academia, then the prestige of your advisor can open doors for you. If you want to leave academia, people still pay attention to the university brand next to your degree.</li></ul><hr><p>Thank you to those who volunteered to read drafts: Alex, Felix, Ross</p><p>Feedback appreciated!</p>]]></content:encoded></item><item><title><![CDATA[You don't take enough risks to be lucky]]></title><description><![CDATA[How I think about long-tailed outcomes]]></description><link>https://www.georgeyw.com/you-dont-take-enough-risks-to-be-lucky/</link><guid isPermaLink="false">66a51b54e68b640a7230caec</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Tue, 20 Jun 2023 06:10:08 GMT</pubDate><content:encoded><![CDATA[<p>I quit my job a month ago, which will probably end up among the best turning points of my career. To be clear, there was nothing <em>wrong </em>with that job, the alternative is just that much better. My social and professional networks have grown dramatically. I have the most intellectual freedom I&#x2019;ve ever had, letting me discover and work on [redacted], [redacted], and [redacted]. Above all, the degree of agency and ownership of my life and success is intoxicating. I might sober up eventually, but for now it&#x2019;s hard to imagine wanting to work a regular 9-5 any time soon.</p><p>There might be some collective wisdom on this. I honestly expected a lot more concern and muted skepticism whenever I revealed that I&#x2019;d quit my job. What I received instead was congratulations, excitement, and curiosity. My guess is that good things seem to happen when someone voluntarily makes a big life decision. Maybe they have an amazing opportunity waiting for them, maybe leaving their comfort zone creates amazing opportunities. Maybe I&#x2019;m just surrounded by relatively fortunate young people without enough financial commitments to make unemployment truly scary.</p><p>Anyways. The point of this isn&#x2019;t to convince you to quit your job. This was a fairly lucky outcome, but I didn&#x2019;t plan for that, and the point isn&#x2019;t to brag about it either. I&#x2019;ve had a lot of lucky outcomes in my life, and the point of this is to share how I made that happen and how you can be luckier, too.</p><h2 id="what-do-you-need-to-be-lucky">What do you need to be lucky?</h2><p>Luck is random, but it&#x2019;s not purely random. If we think of being lucky as having net positive impact from random chance in your life, then we can be skillful by selecting more opportunities where the odds are stacked in our favor. This is hard because the odds are usually too fuzzy to compute.</p><p>This implies a few prerequisites for practicing luck as a skill. First and foremost, you must have an <em>appropriate </em>amount of belief that you can have an impact on the seemingly random outcomes in your life. Our belief in this is heavily conditioned on how lucky we&#x2019;ve been, and we tend to over attribute skill when good things happen and over attribute random chance when bad things happen. As a result, there are people for whom this feels obvious and people for whom this feels impossible, but both parties are off the mark. Underconfidence means you never take enough shots on goal, but overconfidence means you&#x2019;ll bite off more than you can chew, and you&#x2019;ll really be gambling more than practicing luck as a skill.</p><p>You must also have an appetite and tolerance for risk. The goal here is to select more opportunities with odds stacked in your favor, but that still means taking a lot of opportunities where odds are involved. That means you will still lose a lot of bets. Humans are naturally risk-averse, and it will feel bad to lose bets. Dealing with this over time requires a high tolerance for failure, and you must have tolerance both in your mental game and in having <a href="https://en.wikipedia.org/wiki/Risk_of_ruin?ref=georgeyw.com">sufficient resources to absorb the negative outcomes</a>. Resources here can mean things like time, energy, money, social status, etc, and having more resources makes the mental game easier. Quitting my job was made much easier by knowing how long my savings could last and that I was fortunate to have a valuable skill set.</p><h2 id="it%E2%80%99s-a-rigged-game">It&#x2019;s a rigged game</h2><p>Fortune is not evenly distributed. Some people are born richer, some with more social status, some are more neurotypical, some more attractive, and on and on. People with more of these resources have access to more opportunities, and as a result, access to more opportunities with favorable odds. It is easier to afford the necessary risk of these opportunities if you have more resources. Although this post is not about the distribution of these resources, make no mistake: this is the truth and this is not fair.</p><p>That being said, you can make a meaningful impact on your future luck. Whatever your starting point is, <a href="https://danluu.com/p95-skill/?ref=georgeyw.com">you can beat a majority of people who don&#x2019;t try</a>, and most people don&#x2019;t try. It also gets easier over time; luck begets luck. Quitting my job was easier because of my savings and skill set, but the size of my nest egg mostly came from a couple of unique bets that turned out well. I chose my career from the overlap of things I like and things that pay well, which increased my access to good opportunities. Those in turn came from earlier, smaller things. Wherever you start, your luck can gradually snowball and change your life.</p><p>A closely related concept is <a href="https://articles.starcitygames.com/articles/learning-to-truly-play-to-your-outs/?ref=georgeyw.com">playing to your outs</a>, and nearly every life situation has outs, however improbable. These might only serve to make a terrible situation into a slightly less bad one, but they exist for those with the desire to find them. I sincerely hope you are not in such a terrible situation, but if you are, you probably aren&#x2019;t an exception. As harsh as that sounds, it&#x2019;s actually a good thing. The upshot is that <em>you can make a meaningful impact on your future luck</em>. The game is rigged, and the game is hard. We&#x2019;re stuck playing it anyway, so let&#x2019;s play with our eyes open.</p><h2 id="luck-is-making-chances-to-dig-for-outliers">Luck is making chances to dig for outliers</h2><p>The first hard thing about being lucky is realizing that <a href="https://milan.cvitkovic.net/writing/things_youre_allowed_to_do/?ref=georgeyw.com">you can just do things</a>. This is like the extreme version of noticing you can just choose to have cake for dinner. You have ownership over your life, and you are allowed to do more things than you realize to make it better.</p><p>The second hard thing is noticing all the places in your life where you might <em>want </em>to do something. When I say all, I really mean <em>all</em>. Noticing that you can quit your job isn&#x2019;t hard. Neither is noticing that you can go on a backpacking trip through Europe, that you can have kids, that you can exercise more, or that you can move to that city you&#x2019;ve always dreamed of living in. People talk about these things all the time. They&#x2019;re worth thinking about, but far from the only things worth thinking about. Sometimes exploring the full space of Bets You Can Make looks like <a href="https://www.benkuhn.net/abyss/?ref=georgeyw.com">staring into the abyss</a>. Sometimes it looks like ordering a dish you&#x2019;ve never tried before that sounds kind of weird.</p><p>The next step is evaluation; not all of these bets are good bets to take. I&#x2019;m going to focus on a particular type of bet, which I think is fortunately both the easiest to talk about and the most useful. That class is one where there&#x2019;s potential for huge outliers in outcomes. This is what that looks like as a graph:</p><figure class="kg-card kg-image-card"><img src="https://www.georgeyw.com/content/images/size/w1000/2024/07/risk-2.png" class="kg-image" alt loading="lazy" width="1000" height="750"></figure><p>This type of distribution is called long-tailed, heavy-tailed, or right skewed. Line 1 shows roughly where the median outcome is. If you only take a few of these long-tailed bets, then you&#x2019;re likely to only end up with approximately median outcomes, and your luck will feel average. Line 2 shows roughly where the mean outcome is. The key here is that the downside outcomes are minimal and bounded, and the upside outcomes are massive and practically unbounded. If you take enough bets with this distribution of outcomes, then you&#x2019;ll eventually pick up enough outliers (line 3) to bring your net outcome well above the median. This is also <a href="https://www.nber.org/papers/w25113?ref=georgeyw.com">a theory</a> for how diversified portfolios manage to beat treasury returns. Most individual stocks end up with lower return than treasuries, but they&#x2019;re incredibly long-tailed, and it&#x2019;s the few outliers that cause a diversified portfolio to outperform treasuries.</p><p>Ben Kuhn has a great post on systematically <a href="https://www.benkuhn.net/outliers/?ref=georgeyw.com">searching for outliers</a> within specific domains, such as job selection, blog post writing, or startup investing. These tasks have common features:</p><ul><li>You can keep resampling from the distribution by taking new jobs, writing new posts, or investing in more startups</li><li>It&#x2019;s hard to tell how good the outlier outcomes can actually be, especially with low sample size</li><li>Scoring a few outliers covers the cost of searching and then some (and usually a lot more than some)</li></ul><p>One of the key takeaways is that finding success in those domains is largely a function of how well you can repeatedly sample from their outcome distribution.</p><p>My addition to this process is that you don&#x2019;t have to compartmentalize it to each specific domain. Suppose you have 1 domain with a specific outcome distribution and 100 other domains that all happen to have a similar outcome distribution. Assuming you&#x2019;re equally happy to get outliers in any of these domains, you can get as many expected outliers by sampling from the first, single domain 100 times, or by sampling from each of the 100 unique domains once each. In either case, you get to the same number of attempts at scoring outliers.</p><p>Long-tailed results also typically take a long time to be realized. You don&#x2019;t get immediate feedback on whether the job you picked or the partner(s) you&#x2019;re with is an outlier. There&#x2019;s a serial nature to these selection tasks that limits the number of shots you have in your lifespan. If it takes a couple of years to figure out if the job you&#x2019;re at is really exceptional, then focusing solely on resampling new jobs might leave you with only a couple dozen shots on goal in your entire life (and the longer it takes to find an outlier, the less you can reap the benefits). You can roughly double your shots on goal to find (any) outliers in life if you parallelize the process of resampling your job with resampling your romantic partner.</p><p>Why stop at two parallel processes? There are more unique domains than you realize that have long-tailed outcomes and that you can parallelize your sampling from <em>all </em>these domains. </p><p>Unfortunately, this is where the whole thing gets much harder. The biggest source of parallelization is going to be from decisions that are very specific to your life path and that you might only get to make a few times in your life. You might miss them, and if you see them, you might misevaluate them. There are a lot of these types of decisions, and not all of them have this long-tailed outcome distribution. Here is a quick, non-exhaustive list of things you might not get to do very often and which have varying degrees of difficulty, impact, and endorsement.</p><ul><li>Picking an advisor in grad school</li><li>Getting married</li><li>Having an expensive wedding</li><li>Getting divorced</li><li>Having kids or not</li><li>The number of kids to have</li><li>Adopting a kid</li><li>Adopting a pet</li><li>Choosing a career (especially a high-commitment one, like being a doctor)</li><li>Trade on a potential <a href="https://en.wikipedia.org/wiki/Black_swan_theory?ref=georgeyw.com">Black Swan</a> event</li><li>Committing major financial crimes</li><li>Buying a house</li><li>Emigrating to a new country</li><li>Donating a kidney</li><li>Undergoing gender-affirming care</li><li>Following a particular religion</li><li>Taking a particularly unique job abroad</li><li>Whistleblowing</li></ul><p>These are still things that are common or high profile enough that anyone can think of them, but access to opportunities is highly <a href="https://en.wikipedia.org/wiki/Path_dependence?ref=georgeyw.com">path dependent</a>.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh5.googleusercontent.com/TnOIOqxVHGsY42ZfvLMkA6wAaJ-i_3ckVfANrsaFmBg7t6kGUT5K4lF6bCT9pA8ioXBlfdPb03buGLr9W3pv7dv1wZ74TLP-rUxwkLFkgFR8PrVhxYJiXZd09YxiC_qbYKcFF4zeC-fVKdEc8WT4uFY" class="kg-image" alt loading="lazy" width="1478" height="926"><figcaption><span style="white-space: pre-wrap;">Credit: </span><a href="https://twitter.com/waitbutwhy/status/1367871165319049221?lang=en&amp;ref=georgeyw.com"><span style="white-space: pre-wrap;">Wait but Why?</span></a></figcaption></figure><p>Your full list of opportunities will look very different from mine, and it will be much longer than the list above. Your list will be unique and constantly changing, because it depends on your skills, resources, connections, and the state of the world around you. To be luckier, you must be able to frequently and accurately update your list and pick which ones to take the leap on (please don&apos;t pick <a href="https://en.wikipedia.org/wiki/Sam_Bankman-Fried?ref=georgeyw.com">major financial crimes</a>).</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh5.googleusercontent.com/TLUI4YVOSfg41Bqw1jK6XDQw9RYngnvqCFxPNY9Dg2ZnfntiK2_Mp5ACKLxGa2YHHsW_JZ4OlOvcem-R0LgqsISAdpmt8ujsSIaEXO_d9bHCTniz0kfwMI0TLR9fLu2QCckt6RXC-tNitfI7qNWIB9M" class="kg-image" alt loading="lazy" width="1600" height="800"><figcaption><span style="white-space: pre-wrap;">QR codes generated by ControlNet</span></figcaption></figure><p>A recent example that comes to mind are these beautiful, functional(!) <a href="https://mp-weixin-qq-com.translate.goog/s/i4WR5ULH1ZZYl8Watf3EPw?_x_tr_sl=auto&amp;_x_tr_tl=en&amp;_x_tr_hl=en-US&amp;_x_tr_pto=wapp&amp;ref=georgeyw.com">QR codes generated by ControlNet</a>. I&#x2019;m familiar with Stable Diffusion and other generative AI models, but my brain contains basically zero knowledge about QR codes (it turns out QR codes are <a href="https://news.ycombinator.com/item?id=36128082&amp;ref=georgeyw.com">quite malleable</a>). I wouldn&#x2019;t have considered this to be on my personal list of opportunities, but for this person, it was and they took it. This wasn&#x2019;t just a once-in-your-lifetime opportunity, this was once-in-anyone&#x2019;s-lifetime. Now they get to be &#x201C;that person that made super cool QR codes&#x201D; forever, and no one else does. Depending on what you&#x2019;re optimizing for (visibility, career opportunities, internet fame), this could be a fantastic outlier outcome for a speculative project.</p><p>Going for those one-off chances doesn&#x2019;t lend itself to a clean framework of resampling on the same decision, but it does expose you to many more potential outliers over a lifetime. You can get lucky in a particular domain by understanding how to search for outliers in that domain and by repeatedly sampling from that domain. This is good, but this still limits your shots on goal. You also might not find a life-changing outlier until late in life. </p><p>You can get even luckier, and sooner in life, by being able to accurately identify <em>all</em> the domains of long-tailed decisions that are accessible to you and learning to search for these outliers <em>everywhere</em>. Even if some of them are once-in-a-lifetime opportunities, once-in-a-lifetime opportunities abound.</p><h2 id="%E2%80%9Cbut-x-is-already-pretty-good%E2%80%9D">&#x201C;But X is already pretty good&#x201D;</h2><p>At the job that I left, I previously held the belief that it was an exceptionally good place to work. My basis for believing this was that many of my coworkers frequently commented on it being the best place they&#x2019;ve ever worked at. Surely if so many people think so, there must be some truth to it being exceptional?</p><p>I still believe it was a great place to work, but (as Ben Kuhn also discusses in his post) there&#x2019;s a difference between being at the 90th percentile (great) and being at the 99.9th percentile (exceptional). If you&#x2019;re at a 90th percentile company and most of your colleagues have worked at around 5 places in total, there&#x2019;s a 65.6% chance that your 90th percentile company is better than all of the previous 4 places they&#x2019;ve been. So being at a 90th percentile company might look like two thirds of your coworkers truthfully being able to say it&#x2019;s the best job they&#x2019;ve ever had, even though 1 in 10 places are better.</p><p>Truly strong evidence of an exceptional outcome looks stronger than two thirds of people with something comparable saying it&#x2019;s the best outcome they&#x2019;ve ever had. Evidence of an exceptional outcome looks more like zero people around you with something comparable. Without this kind of evidence, chances are that a given situation in your life is closer to a median outcome than you think it is. If so, the risk for rolling the dice again is also lower than you think it is. In general, we bias away from change, but we should try to embrace it to become lucky. In 2020, a <a href="https://academic.oup.com/restud/article-abstract/88/1/378/5834495?redirectedFrom=fulltext&amp;login=false&amp;ref=georgeyw.com">study</a> asked 20,000 people to make some difficult decision on their mind with a coin flip:</p><blockquote>For important decisions (e.g. quitting a job or ending a relationship), individuals who are told by the coin toss to make a change are more likely to make a change, more satisfied with their decisions, and happier six months later than those whose coin toss instructed maintaining the status quo.</blockquote><p>(n.b. If you&apos;re a former coworker, don&apos;t take this as advice or judgement. I haven&apos;t even worked at 5 places yet!)</p><h2 id="what-you-can-do-to-get-luckier">What you can do to get luckier</h2><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.georgeyw.com/content/images/size/w1000/2024/07/risk-1-1.png" class="kg-image" alt loading="lazy" width="1000" height="750"><figcaption><span style="white-space: pre-wrap;">The basic feedback loop</span></figcaption></figure><p>Network more. Every new person that you develop a connection with, either socially or professionally, can be a long-tailed outcome. That might sound a little sociopathic, but the quality of our relationships predict a lot of our life happiness, so we should be more intentional in creating chances for outliers. Most people will be acquaintances, but some can be lifelong friends, romantic partners, and/or collaborators. Furthermore, a rich and healthy network might also spot long-tailed opportunities that are uniquely suited for you and bring them to you. Noticing good bets is easier when you have more bets to choose from, and having a strong network is much faster than searching alone (pay this forward!).</p><p>Be curious, too. It costs less energy to spend time learning about things that excite you. Doing <a href="http://www.paulgraham.com/greatwork.html?ref=georgeyw.com">great work</a> often looks like following your curiosity to places where people haven&apos;t tread before. This can mean going deeper than anyone else has, but it can also mean going broad enough that you find a unique combination of knowledge, as with the earlier QR code art.</p><p>In the long term, it&#x2019;s crucial to develop a strong probabilistic mindset and intuition, which makes it easier to reason about whether a decision is potentially long-tailed or not. This also helps avoid certain types of opportunities, such as the lottery, which is arguably long-tailed due to massive outliers, but not actually worth it. Finally, this helps identify downside risk, so that you can avoid making a decision that leads to a complete blow up. As you&#x2019;re exposed to more chances, you&#x2019;ll develop a better sense of what makes an opportunity good or bad in this framework, including making more accurate assessments for more unique long-tailed decisions that you only get to make once or twice.</p><p>There isn&#x2019;t a lot of general advice that I can offer beyond this. The ideal situation for developing a skill is one where you can do deliberate practice, but this framework <a href="https://commoncog.com/the-problems-with-deliberate-practice/?ref=georgeyw.com">can&#x2019;t be applied to the domain of luck</a>. There are no well-established training methods, and as far as I can tell, no one officially coaches for being lucky. A lot of my personal skill in this domain falls under gut feeling, or <a href="https://commoncog.com/the-tacit-knowledge-series/?ref=georgeyw.com">tacit knowledge</a>, and I haven&#x2019;t thought explicitly about this enough to make that gut feeling explicit. If you&#x2019;ve thought about this and have ideas, please share them. In lieu of a perfectly written guide, here are some things you might start with:</p><ul><li>Interview occasionally at jobs that sound interesting, even when you aren&#x2019;t looking for a new job</li><li>If you&#x2019;re introverted, get adopted by an extroverted friend</li><li>Say yes to invitations as a default</li><li>Do more things where the worst that can happen is nothing</li><li>Do fewer things that look like sitting at home and watching Netflix (I am guilty of sitting at home and playing video games)</li><li>Go to random events or activities on <a href="https://www.meetup.com/home/?ref=georgeyw.com">Meetup</a> or <a href="https://www.eventbrite.com/?ref=georgeyw.com">Eventbrite</a></li><li>Start saving up some resources to be ready for opportunities when they come; remember that your resources are resources, and you should spend them (wisely)</li><li>Don&#x2019;t do pyramid schemes</li></ul><p>Mileage is the best teacher available, so find some risk you can tolerate, and start taking some shots.</p><hr><p>This is my first post. It was inspired by this <a href="https://notebook.drmaciver.com/posts/2022-05-25-09:47.html?ref=georgeyw.com">unfinished draft</a> by David R. MacIver and by this post on <a href="https://www.benkuhn.net/writing/?ref=georgeyw.com">starting a blog</a> by Ben Kuhn. I recently re-read it for the third or fourth time and finally got off my ass to try my luck with it.</p><p>Thank you to those who volunteered to read drafts: AJ, Cyrus, Felix, Ross.</p><p>Any feedback is appreciated!</p>]]></content:encoded></item></channel></rss>