<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:media="http://search.yahoo.com/mrss/"><channel><title><![CDATA[This Is My True Name]]></title><description><![CDATA[This Is My True Name]]></description><link>https://www.georgeyw.com/</link><image><url>https://www.georgeyw.com/favicon.png</url><title>This Is My True Name</title><link>https://www.georgeyw.com/</link></image><generator>Ghost 5.79</generator><lastBuildDate>Thu, 22 Feb 2024 15:02:24 GMT</lastBuildDate><atom:link href="https://www.georgeyw.com/rss/" rel="self" type="application/rss+xml"/><ttl>60</ttl><item><title><![CDATA[Calibrating your gut]]></title><description><![CDATA[<p>By <em>gut</em> or <em>gut feeling</em> I mean our non-verbal physical and emotional responses to a decision that are pointed to by words like <em>intuition</em> or <em>instinct</em> and phrases like <em>follow your heart</em>. When it cooperates, our gut can be a source of deeply felt conviction, even in murky waters, or</p>]]></description><link>https://www.georgeyw.com/calibrating-your-gut/</link><guid isPermaLink="false">65c15ff3ba7cd00001c6d2cc</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sat, 17 Feb 2024 03:20:18 GMT</pubDate><content:encoded><![CDATA[<p>By <em>gut</em> or <em>gut feeling</em> I mean our non-verbal physical and emotional responses to a decision that are pointed to by words like <em>intuition</em> or <em>instinct</em> and phrases like <em>follow your heart</em>. When it cooperates, our gut can be a source of deeply felt conviction, even in murky waters, or &#x2013; when it doesn&apos;t &#x2013; a source of uncertainty, even when we know all the information. Gut feelings act on both a short timescale (the milliseconds before conscious thought even starts) and on a long timescale (the hours, days, weeks of deliberation on a difficult life choice). Whether we notice or not, our gut feelings influence the way we move through the world.</p><p>Okay, and what about <em>calibrating</em>? By calibrating, I mean something like &quot;bringing into alignment with reality,&quot; i.e. making it so that your gut feelings seem to be good predictors of real-world outcomes. But words like intuition and instinct seem to imply a somewhat immutable relationship to your gut &#x2013; what is there to <em>do</em> about it? </p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://i.kym-cdn.com/photos/images/original/000/572/078/d6d.jpg" class="kg-image" alt loading="lazy" width="500" height="407"><figcaption><span style="white-space: pre-wrap;">&quot;just trust your gut&quot; -- unfortunately unhelpful if you already know how to </span><i><em class="italic" style="white-space: pre-wrap;">and</em></i><span style="white-space: pre-wrap;"> if you don&apos;t</span></figcaption></figure><p>My answer to this is good, old-fashioned, high-quality, educated speculation grounded in a sample size of one. As with many things, it starts with the <a href="https://agentyduck.blogspot.com/p/noticing.html?ref=georgeyw.com" rel="noreferrer">art of noticing</a>. That is, to &quot;listen to your gut,&quot; but not in the sense of do-what-your-gut-tells-you, more in the sense of try-to-notice-that-your-gut-is-saying-anything-at-all-even-if-you-have-no-idea-what-it&apos;s-going-on-about. In fact, you should be careful of the first sense, because at this stage your gut might still be wrong.</p><h2 id="noticing">Noticing</h2><p>There&apos;s a video game series I&apos;m fond of called Fire Emblem. It&apos;s a tactical RPG where you try to strategically hit things and not get hit back too badly. If you do this well, eventually you get to hit a dragon until it dies and you win. There are various game mechanics involved, but the most basic one is your hit chance. For this, Fire Emblem uses a system called True Hit. Naturally, it lies to you.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.escapistmagazine.com/wp-content/uploads/2023/06/fire-emblem-gba-nintendo-switch-online-june-23.jpg?resize=800%2C400" class="kg-image" alt="Fire Emblem Joins GBA Nintendo Switch Online on June 23" loading="lazy" width="800" height="400"><figcaption><span style="white-space: pre-wrap;">right: player character a 93% hit chance is displayed, but the actual hit chance is ??%</span></figcaption></figure><p>The details of how exactly it lies to you aren&apos;t that important. The important effect is that displayed hit chances that are above 50% are actually higher in practice, and those below 50% are actually lower in practice. Fire Emblem does this because most players, on a <em>gut level</em>, feel that a 95% chance event ought to happen much more often than 95% of the time, and that a 99% chance event is an <em>almost certainty</em> &#x2013; it feels like an injustice when it doesn&apos;t happen! In that sense, Fire Emblem calibrates the game mechanics to better fit a player&apos;s gut feelings to make the game feel more intuitive. But what this really does is just hide what your gut is trying to say, even if it&apos;s wrong.</p><p>Reality typically won&apos;t recalibrate itself to match your gut feelings, but that means we can <em>notice</em> what happens when our gut is saying something. The exact &quot;motions&quot; (mostly involuntary and mental rather than physical) happening aren&apos;t super important, only the noticing of them.  There might be a certain way that we plan <em>around</em> an event, taking it for granted when we shouldn&apos;t. We might mentally dwell too much on an irrelevant fact or gloss over something important. We might flinch away from an uncomfortable thought or fidget or feel various physical sensations. There might be a specific flavor of mental unease that pushes us to defer entirely to spreadsheets. Whatever it is, there&apos;s something there to attend to.</p><h2 id="calibrating-yourself-to-your-gut">Calibrating yourself to your gut</h2><p>Last year, I attended an ad hoc workshop on improving your calibration for prediction markets. There was surprisingly little quantitative analysis &#x2013; it was mostly about learning to pay conscious attention to the specific gears moving in our brains that led to them spitting out a probability. The point was to be able to correctly associate these gut feelings / gears / mental motions to reality. That is, suppose you think something is the right call 90% of the time and that comes with some kind of internal subjective experience composed of various feelings and motions. Actually though, suppose it turns out that&apos;s only true 70% of the time. If we&apos;re paying attention, we can learn to re-associate our mental connection with that cluster of emotions from being &quot;what we feel when something is right 90% of the time&quot; to &quot;what we feel when something is right 70% of the time.&quot;</p><p>That exercise was limited to prediction markets and assignment of probabilities, but I suspect this is much more broadly applicable. If we&apos;re listening to a new argument for something, we can pay attention to where we feel resistance, if our nervous system gets activated, or where our mind tries to pattern match, and later cross-check that against a more objective judgment. When presented with an unfamiliar situation, we can notice the things that happen internally as we move through the situation and make decisions and later review if our internal feelings seemed to correspond well with what turned out to be the right decision or not. </p><p>As a personal example, I&apos;ve historically struggled with writing the bottom line first and working backwards to justify it. I recently started noticing &#x2013; <em>feeling in my gut</em> &#x2013; what it feels like when I&apos;m rationalizing something, before I consciously realize I&apos;m doing so (and I then try to follow that up by explicitly acknowledging it and positively reinforcing myself for successfully noticing). This has lots of smaller, day-to-day consequences, but it was also a larger scale version of this noticing (that I was rationalizing my trajectory at the time) that became a catalyst for my recent career shift to <a href="https://www.georgeyw.com/moving-to-ai-safety/" rel="noreferrer">AI safety</a>. </p><h2 id="actually-trusting-your-gut">Actually trusting your gut</h2><p>So far we&apos;ve been talking about how to listen to and interpret what your gut is saying. That seems pretty valuable to me already, but sometimes you might have to do the <em>opposite</em> of what your gut is saying if you&apos;ve noticed it&apos;s consistently getting something wrong (such as being uneasy with hard conversations despite them typically turning out well). I claim that at this point, most of the difficult work is done, and you just have to give it time.</p><p>If you maintain this explicit mental attention / feedback loop between your internal motions and outcomes, then your gut feelings will gradually evolve. They&apos;ll tend to ease their discomfort in situations where you notice that such discomfort is unnecessary and to strengthen their conviction where you notice it is warranted. If you stay in tune with these changes, then better calibration &#x2013; and being able to <em>trust</em> your gut &#x2013; is something that just naturally follows.</p>]]></content:encoded></item><item><title><![CDATA[Alternatives to (math) academia and how to get there]]></title><description><![CDATA[This is a reference post for math PhDs]]></description><link>https://www.georgeyw.com/math-alternatives/</link><guid isPermaLink="false">656bf27815752c0001c2ad83</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sun, 03 Dec 2023 07:08:55 GMT</pubDate><content:encoded><![CDATA[<p><em>[This is a reference post aimed at mathematicians (PhDs, postdocs, professors) who might be interested in options outside of academia. It outlines common career paths and how to find a foothold in those directions. This might be of some limited use to those who are interested in the careers but come from a different background. This is unlikely to be of use to people with neither the background nor the career interest.]</em></p><p>I&apos;ve had this conversation a bunch of times now, and I&apos;m happy to talk about it each time, but it seems like it would be more convenient for all parties if there were just a compiled list of my advice somewhere, so here&apos;s an attempt at such a compilation. Apologies in advance for any gaps, please feel free to email me with questions, and answers may get added to this reference, though I think I&apos;m approximately aiming for an 80/20 rule here. </p><h2 id="why-leave">Why leave?</h2><p>Whatever reasons there are will be highly personal. I can list some of my own reasons below, as well as some that I&apos;ve heard, but if you&apos;re reading this, you might have some of your own as well. Note that this is unashamedly biased towards leaving academia, but I don&apos;t think someone is wrong if they conclude that staying in academia is right for them, so long as the decision is well-informed. </p><p>But my first lukewarm take is that the discoverability of other options is generally terrible, and the friction of figuring out what options there are and how to actually achieve those options is enough to stop many people from seriously considering an exit.</p><p>I&apos;ll also note a personal mental obstacle that I had as a young grad student. For whatever reason, I had a sort of cached mental impression that not &quot;making it&quot; to an academic job was a failure. It&apos;s hard to pinpoint exactly where that impression came from. I&apos;m sure some of it was internal, but there&apos;s also a degree to which this view is baked in around us &#x2013; for example, it&apos;s common to refer to leaving academia as selling out, which has some negative connotation. I don&apos;t think holding the view that it&apos;s a failure is healthy, nor is it true. </p><p>In reality, I think <strong>most of the reasons why someone might remain in academia are superseded by the appropriate arrangement outside of academia</strong> (of which there are <em>many</em>), up to and including being able to do <em>more</em> research outside of academia. Read on for more details.</p><h3 id="reasons-to-leave">Reasons to leave</h3><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.georgeyw.com/content/images/2023/12/image.png" class="kg-image" alt loading="lazy" width="1180" height="750" srcset="https://www.georgeyw.com/content/images/size/w600/2023/12/image.png 600w, https://www.georgeyw.com/content/images/size/w1000/2023/12/image.png 1000w, https://www.georgeyw.com/content/images/2023/12/image.png 1180w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">Do you really need more convincing than this? Scroll down, I guess... (</span><a href="https://twitter.com/nearcyan/status/1681166975303598080?ref=georgeyw.com" rel="noreferrer"><span style="white-space: pre-wrap;">source</span></a><span style="white-space: pre-wrap;">)</span></figcaption></figure><ul><li>You make (a lot) more money outside of academia<ul><li>By the time you would get tenure in academia, you could counterfactually be a senior or staff level researcher/engineer (see later on for a note on &quot;levels&quot;) in tech and earn <a href="https://www.levels.fyi/?compare=Apple%2CFacebook%2CAmazon%2CMicrosoft%2CGoogle&amp;track=Software+Engineer&amp;ref=georgeyw.com" rel="noreferrer">300-500k+ total compensation</a> (combined salary + equity grant, which is as good as cash)</li><li>The <a href="https://www.georgeyw.com/my-phd-costed-me-1-17m-so-far-but-id-do-it-again/" rel="noreferrer">opportunity cost</a> of staying in academia is well into the six figures per year, just from the raw salary difference, not even accounting for compounding growth of investments. If you weren&apos;t already in academia would you <em>pay</em> $200k+ per year in order to <em>be</em> in academia? (if you look further down, my estimation of the opportunity cost at the tenured professor level is much higher)</li></ul></li><li>The job demands, stress, and time commitment are typically lower<ul><li>The demands are often low enough that you can hold a full time job and <em>still have more time for research than in academia</em></li><li>The Google employees in the tweet above are probably on the low end, but not <em>that </em>far from the true median.</li></ul></li><li>You have more flexibility in where you live (especially with remote work)</li><li>Remote work in tech specifically also comes with extremely flexible hours</li><li>Less administrative burden, zero course preps, no outreach/service expectations</li><li>If you learn certain practical skills, you are incredibly employable and don&apos;t <em>need</em> the job security guarantees of tenure (and make enough money anyways that you can absorb unlikely periods of unemployment, or just straight up retire after ~10-20 relatively chill years and just do whatever the hell you want after)</li><li>There may be a personal situation tying you to your job. Depending on that situation (e.g. with visa status), you may find a solution anyways outside of industry (many tech companies sponsor visas)</li></ul><h2 id="where-do-i-go">Where do I go?</h2><p>Ok, so let&apos;s say you&apos;re considering leaving. What are your career options? When I first Googled &quot;math major career options&quot; in undergrad, the top result was being an actuary. Nothing against actuaries, they&apos;re probably cool people, but I don&apos;t want to personally find out. Your career options are much more varied than this.</p><ul><li>Tech / Silicon Valley<ul><li>This can mean researcher (e.g. machine learning, data science, etc) or software engineer</li><li>This is broadly split into Big Tech (FAANG companies) and startups</li><li>Big Tech has the highest compensation between the two and the lowest job demand (which is paradoxical, but FAANG makes a <em>lot</em> of money, and it&apos;s hard to be ultra efficient at their scale)<ul><li>As an intuition pump: Elon caught a lot of flak for firing like, 80% of X (f.k.a. Twitter), but it&apos;s still running (albeit with some performance degradation here and there). I was among the doubters, but so far it seems like he was right on this point. Many big tech companies could probably survive &gt;50% cuts if they had the guts to do it.</li></ul></li><li>Startups have the most exciting work, but lower compensation (higher than academia) and higher job demand (comparable to academia). If you&apos;re an early employee at a startup that eventually IPOs, you&apos;re probably going to walk away very rich, but don&apos;t count on it. The attraction here is working on problems that feel like they&apos;re either building the future or really fixing existing problems in a meaningful way. Be careful not to choose a startup whose mission you don&apos;t care about. The <em>entire point</em> of working at a small startup is mission alignment. Working on a mission you <em>do </em>care about feels amazing. Some real examples of startups that you can really work on:<ul><li>Flying houses (mission: tiny houses that can be airlifted into disaster relief areas)</li><li>Lab grown meat (mission: end animal suffering due to factory farming)</li><li>Alternative energy (mission: solve world energy supply)</li><li>Wearable tech (biometrics: live healthier lives / AI assistants: extend our cognition with new tools)</li><li>Self-driving cars (lots of potential missions, e.g. solve car accident mortality)</li><li>Literal space travel (mission: go to SPACE and maybe COLONIZE MARS)</li></ul></li><li>Those are just the eye-catching ones. If you can think of a real problem in the world, even if it&apos;s incredibly unsexy, there&apos;s probably a startup somewhere working on it (lots of startups also suck though and are working on fake problems, but evaluating startups is an entire 5k word guide on its own)</li></ul></li><li>Quant finance<ul><li>Pay is similar to big tech</li><li>Great if you are specifically extremely competitive, resilient to stress, and can be happy with high variance positive EV decision making even if the realized consequence was negative (think activities like poker)</li><li>There is a particular type of person for whom this is a good fit, but most people are not, and that is fine</li></ul></li><li>NSA<ul><li>The NSA employs a lot of mathematicians, but I&apos;m not familiar with the precise kind of work they do. Presumably a lot of cryptography related work.</li><li>I hear the work-life balance is great at the NSA</li></ul></li><li>CCR campuses<ul><li>Apparently the NSA also has more &quot;relaxed&quot; research centers called Center(s) for Communications Research</li><li>Examples: <a href="https://idaccr.org/?ref=georgeyw.com" rel="noreferrer">Princeton</a>, <a href="https://www.ida.org/en/ida-ffrdcs/center-for-communications-and-computing/center-for-communications-research-la-jolla?ref=georgeyw.com" rel="noreferrer">La Jolla</a> (there are more if you Google, for some reason they don&apos;t seem to be aggregated anywhere)</li></ul></li><li>NASA<ul><li>You can do cool space stuff!</li></ul></li><li>Government labs<ul><li>There are a variety of national labs that fund fundamental research</li><li>If such a lab happens to fund the type of research you do, you can enjoy what is essentially an academic job but without the teaching</li></ul></li><li>Healthcare / biostats<ul><li>A close friend of mine does work in epidemiology using statistics and gets to do covid modeling</li></ul></li><li>Honestly, almost anything<ul><li>Your most powerful tool is being agentic. You can, in some sense, really do whatever you put your mind to &#x2013; most people are <a href="https://drmaciver.substack.com/p/learning-to-walk-through-walls?ref=georgeyw.com" rel="noreferrer">limited by artificial walls</a>. A PhD is <em>proof that you can do hard stuff on your own</em>. The degree of career freedom you have outside of academia is incredible. Not all artificial walls are worth tearing down, but before you think &quot;I could never do that,&quot; look with fresh eyes and ask if that&apos;s really true.</li><li>There are plenty of careers that there aren&apos;t degrees for. America employs ~160M people, and my undergrad institution had like ~120 majors. Lots of people have jobs that they didn&apos;t formally study for, and you can do that too. </li></ul></li><ul><li>Things that I did or almost did at various points (in the span of only 2.5 years!) that no one had to give me permission or a formal list of prereqs for: education tools, music tech, science communication, game design, sports betting (for American football &#x2013; and I <em>still don&apos;t even know what a tight end is</em>), algorithmic trading, software engineering, blockchain/web3, tech consulting/advising for early startups, data science, supply chain optimization, AI safety research, and probably a lot more that I&apos;ve forgotten already</li></ul><li>Other thoughts<ul><li>Perspectives that people have on PhDs are mixed, but there&apos;s more than enough people who look at it and say &quot;oh shit, they have a PhD, they&apos;re probably really smart&quot; rather than &quot;their thesis isn&apos;t relevant, let&apos;s not hire them&quot; that I think it&apos;s net positive (you only need one good hiring manager to think positively towards a PhD)</li><li>There is a ton of interesting work that happens outside of academia.</li></ul></li></ul><h2 id="how-do-i-actually-concretely-get-a-job">How do I actually, concretely get a job?</h2><p><em>[This is from the perspective of a student, but postdocs and professors can adapt the same general strategies.]</em></p><p>For most of the common career paths, the primary new skill you need to pick up is some degree of programming (coding combinatorics experiments is a good starting point, but not representative of the skills you need) and/or some degree of statistics/data science/machine learning (these three all blend together to different degrees at different jobs). Everything else you can learn on the job (and don&apos;t worry &#x2013; everyone does). If you&apos;re picking the &quot;choose-your-own-adventure&quot; option in the previous section, then I&apos;ll leave it up to you to determine how much of what skill you need to learn and how you learn it.</p><ul><li>You probably don&apos;t need to learn much programming, if any, for jobs like at the NSA or at government labs (or if you do, they might just teach you on the job)</li><li>You need to learn a little programming (probably leaning towards statistical languages, e.g. Python, R) but mostly data science concepts for things like biostats and data science</li><li>You need to learn a decent amount of programming if you want to do software engineering</li><li>You need both programming (less than for software engineering) and data science/ML for quant finance</li><li>You need both if you want to specialize to certain parts of software engineering, such as machine learning engineering</li><li>Lukewarm take, but I think basically any job in the realm of knowledge work will benefit from some proper programming experience, even if the job description doesn&apos;t explicitly call for it (and even if none of your coworkers know any)</li></ul><p><strong>From here on, the main focus will be on getting tech/quant/data science jobs. The reason for this is that these are 1) what I&apos;m most familiar with and 2) what I expect will be the most involved processes / require the most domain specific prep knowledge.</strong></p><h3 id="learning-data-science-and-machine-learning">Learning data science and machine learning</h3><p>Your core resource is <a href="https://www.kaggle.com/?ref=georgeyw.com" rel="noreferrer">Kaggle</a>. Start with the <a href="https://www.kaggle.com/competitions/titanic?ref=georgeyw.com" rel="noreferrer">Titanic</a> and <a href="https://www.kaggle.com/competitions/house-prices-advanced-regression-techniques?ref=georgeyw.com" rel="noreferrer">Housing Prices</a> intro competitions (in that order). Find some example notebooks, or if you&apos;re braver, try to analyze the dataset and generate predictions on your own without hints. These notebooks exist to familiarize you with the common tools, languages, and techniques that are used in data science. Pay attention to the questions asked, but also the heuristics and rigor. Notice the types of questions which are <em>not</em> asked. Data science is a different field from mathematics with different standards for what is or isn&apos;t a good answer to a question.</p><p>Work your way up to trying some <a href="https://www.kaggle.com/competitions?ref=georgeyw.com" rel="noreferrer">live competitions</a>, including competitions with real prize money. Pick one that seems interesting, maybe find a few friends who are also trying to learn data science and form a team. Give yourself somewhere from a few weeks to a couple months and dig really deep into that competition. There is no substitute to learning by doing.</p><p>As a reference point for the level of difficulty of questions you might get in interviews:</p><ul><li>One of my onsite questions at a quant hedge fund was essentially a series of questions about the housing prices dataset</li><li>Another ML/data science role asked me to code the k-nearest neighbors algorithm from scratch (I later learned that they didn&apos;t mean <em>literally</em> from scratch, it was meant to be a test of how well I knew the numpy library in Python)</li><li>Various questions on linear regression. These types of questions did not inherently require knowledge beyond the undergraduate level, but they <em>did</em> require a depth of understanding of those concepts that wouldn&apos;t normally be expected of an undergraduate. <ul><li>E.g. suppose you have a dataset on the 2D plane, where data is uniformly distributed in a rectangle. What happens to the linear regression line as you rotate that rectangle 90 degrees clockwise or counter-clockwise?</li><li>Similar questions about normal distributions are common</li></ul></li></ul><h3 id="learning-to-program">Learning to program</h3><p>Programming experience in the software industry is typically split into &quot;production&quot; and &quot;not production&quot; level experience. &quot;Production&quot; means work that was done in a commercial setting or otherwise at a commercial level. This typically implies things such as: real money is on the line, code is run live without supervision most of the time, and coding is done as a team in a shared repository with peer reviews, among other things. Unfortunately, within the software industry, things like personal projects or coding experiments for math research are not typically considered production-level coding experience. </p><p>At first glance, there is an apparent catch-22 that it&apos;s hard to find a job if you don&apos;t have production-level coding experience, but you can&apos;t get production-level coding experience if you never get a job. Internships can be one solution, but maybe you&apos;re applying in your last year and you don&apos;t have any summers left, or maybe you&apos;re a postdoc/professor already. In that case, here is a &quot;minimum viable solution&quot; to getting out of the catch-22:</p><ul><li>Watch the MIT OCW <a href="https://ocw.mit.edu/courses/6-006-introduction-to-algorithms-spring-2020/?ref=georgeyw.com" rel="noreferrer">Introduction to Algorithms</a> videos. Doing problem sets is optional. <ul><li>Become comfortable with the data structures and algorithms presented as well as in the analysis of time and space complexity of solutions.</li></ul></li><li>Doing problem sets in the previous step is optional because doing problems in this step is <em>not</em> optional. Make a <a href="https://leetcode.com/?ref=georgeyw.com" rel="noreferrer">LeetCode</a> account, consider if a premium subscription is useful to you (I bought one month of premium at one point and found it useful enough for one month but not more), and then do a ton of problems.<ul><li>Start with easy problems, then move up to medium once you feel good about easy problems.</li></ul></li><ul><li>There is a (now somewhat infamous maybe) list of <a href="https://gist.github.com/krishnadey30/88c4e2f601e96597974c00185e479532?ref=georgeyw.com" rel="noreferrer">75 LeetCode problems</a> that aim to be representative of the types of problems you&apos;ll run into in interviews. If you&apos;re pressed for time (e.g. graduating soon), this is the minimal set of LC problems you should aim to solve</li><li>LeetCode is seriously the core of the current <a href="https://en.wikipedia.org/wiki/Metagame?ref=georgeyw.com" rel="noreferrer">meta</a> of (entry level) tech interviews. It is a fake skill in the sense that the connection with ability to do real work is dubious (though I don&apos;t think it&apos;s totally indefensible), but it is a real skill in the sense that you can really study to the exam and get good at LC style problems and this is highly instrumental for getting a job in the first place.</li><li>A good benchmark is being able to consistently do LeetCode medium problems on your own, with no hints, in 15-30 minutes. The vast majority of interview problems are not harder than LeetCode medium. Past this point, you start to get diminishing returns on the time spent on LeetCode, but it is (IMO) quite hard to accidentally get to a point where marginal practice is pointless.</li><li>Quick plug for a fun alternative to LeetCode: <a href="https://adventofcode.com/?ref=georgeyw.com" rel="noreferrer">Advent of Code</a>. This is an annual advent calendar of coding problems that increases in difficulty as it approaches Christmas and has a silly story to go along with it.</li></ul><li>Finally, work on some actual personal projects. Find something interesting and just do it. <ul><li>You can sometimes get a 2 for 1 with data science/ML experience by doing some kind of coding-intensive competition on Kaggle.</li><li>The first project I ever worked on that made it onto my resume was an automated arbitrage bot for <a href="https://www.mtgo.com/en/mtgo?ref=georgeyw.com" rel="noreferrer">Magic The Gathering: Online</a>. The code was trash, it was a buggy mess (in our defense, this was partly because MTGO itself was a buggy mess), and we got banned from the platform shortly after. We made some actual money off of running the bot, but the real value was returns on career capital &#x2013; it has one of the highest return rates of any single week I&apos;ve ever spent coding. It is the most common talking point in my interviews and is still proudly displayed on my resume.</li><li>Check out this <a href="https://news.ycombinator.com/item?id=38511280&amp;ref=georgeyw.com" rel="noreferrer">Hacker News thread</a> for examples of long-tail outcomes of interesting projects</li></ul></li><li>Another avenue to show off some of your coding (and to actually get some real code review/feedback) is to make open source contributions on Github. I won&apos;t get into how to do this, but some employers view this very favorably and like to see actual code examples as a way to de-risk hiring you. </li><li>A longer path (which you can take if you decide well in advance that this is the path you&apos;d like to take) is to take some computer science classes, and maybe even get a masters in CS. This seems effective if you&apos;re willing to put in the time commitment and can plan ahead for it.</li></ul><h3 id="misc-quant-questions">Misc quant questions</h3><p>There is a small industry built around certain types of quantitative problems and brain teasers that get asked almost exclusively in quant interviews. Getting into these problems is probably out of scope for this document, but you can find plenty of examples online.</p><h3 id="general-interview-advice">General interview advice</h3><p>Interviewing is a skill, too. Your first few interviews will probably feel like garbage. If I haven&apos;t interviewed for a while, my interview skills get rusty, and it takes a few warm-up interviews to get comfortable with them again. Bias towards interviewing earlier in your interview prep process and just get the crappy first few interviews out of the way. I promise it gets better, and there are enough companies out there that flubbing a few interviews for middle-range targets isn&apos;t an issue.</p><p>For most of the quantitative jobs above, the vast, vast majority of your interviews will be technical interviews. Other interviews (such as with HR or recruiters) will be more like formalities in the process, mostly just checking that you&apos;re reasonably easy to talk to. I didn&apos;t have a single real behavioral interview until one point where I was speculatively interviewing for a totally different kind of job.</p><p>For technical interviews, a full guide is also out of scope for this reference, but here are some general pointers:</p><ul><li>Think out loud (seriously, it is sometimes a red flag for interviewers if you are too quiet)</li><li>Ask clarifying questions (green flag if you ask them, red flag if you don&apos;t + were wrong about interpreting the question)</li><li>Don&apos;t be afraid to start with naive solutions to programming problems where the time complexity sucks and is obviously not optimal. As long as you <em>say that out loud</em> and then iterate from there, it&apos;s totally fine and counts as partial progress towards answering the question</li><li>If a question seems really unusually hard, it might be your unlucky day, but it might also be the case that it&apos;s hard for everyone. Many places rate interview performance on problems &quot;on a curve&quot;</li><li>For some questions, you can get more &quot;points&quot; by sharing high quality thinking out loud without arriving at the right answer than by just jumping to the right answer. What this looks like is: asking the right questions, thinking about the right things, coming up with tractable, reasonable sounding ideas (that may not ultimately work)</li><li>If you know someone who is in tech, call in some favors and do practice interview sessions and get feedback</li><li>The interview process can be pretty long; I think my longest interview processes had 5 rounds (where the last one is an &quot;onsite,&quot; though these are often over Zoom now instead of physically onsite), and I&apos;ve heard of more. Have a constant pipeline of interviews, and don&apos;t wait for one interview process to end before looking for another. </li><li>Don&apos;t get too attached to any one job. I made this mistake a few times, and still ultimately ended up with a job I was very happy to get each time (and which was likely better than the counterfactual offer anyways)</li></ul><h2 id="a-note-on-levels">A note on &quot;levels&quot;</h2><p>One of the great (IMO) features of tech careers is the notion of a &quot;level.&quot; Although they aren&apos;t fully standardized between different companies (and it takes some experience to know what to expect from a &quot;senior&quot; engineer coming from different types of orgs), I still find them a pretty useful framework for understanding jobs in tech. A typical sequence of levels at tech companies (or quant sometimes, for larger firms) looks something like the following, where if you&apos;re on say, a data scientist track, you might replace &quot;software engineer&quot; with &quot;data scientist&quot; everywhere.</p><ul><li>Junior<ul><li>Sometimes can be a title like &quot;Software Engineer 1&quot; or &quot;SDE 1&quot; or &quot;SWE 1&quot;</li><li>Entry level for bachelors</li><li>Very low expectations, your job at this level is mostly to learn. Companies typically view juniors as an investment more than a clear source of value.</li></ul></li><li>Mid<ul><li>Often comes without a title prefix, e.g. just &quot;Software Engineer&quot; and not &quot;Mid-level Software Engineer&quot;</li><li>Otherwise might look like &quot;Software Engineer 2&quot; or &quot;SDE 2&quot; or &quot;SWE 2&quot;</li></ul></li><li>Senior<ul><li>Typically considered a &quot;terminal&quot; role in tech &#x2013; there are many levels beyond this, but this is where you&apos;re considered good enough for companies to be happy to have you sit here indefinitely and provide this level of value without growing further. Many people do just that.</li><li>Typical expected time is ~5 yrs to reach this level (for the median engineer that makes it to senior level, not for the median engineer <em>with a PhD</em> that makes it to senior &#x2013; you can make it to senior much faster if you want but you do not have to)</li><li>When interviewing for this level and above in software, you will start to need to know <a href="https://github.com/Jeevan-kumar-Raj/Grokking-System-Design?ref=georgeyw.com" rel="noreferrer">system design</a>. Expect at least one technical interview to be on system design problems.</li></ul></li><li>Staff<ul><li>This is where there&apos;s an option to move into the management track, which has its own levels that parallel the &quot;individual contributor&quot; (IC) track<ul><li>Most people actually move into management instead of moving up the IC ladder, see notes on that below</li></ul></li><li>It is harder to define what your role is at this level and above, staff+ ICs tend to get increasingly specialized roles based on their individual capabilities and the needs of the company</li><li>Compensation was already stupid, but this is where it gets real dumb. 500k+ total compensation is common at FAANG</li></ul></li><li>Principal<ul><li>Here levels start to break down further at different companies. Sometimes there is a &quot;senior staff&quot; before this level and sometimes there is a &quot;senior principal&quot; after this level. Sometimes staff is skipped, sometimes principal doesn&apos;t exist.</li><li>Compensation gets downright offensive: FAANG principal+ can earn 7 figures</li></ul></li></ul><p>Management track has similar levels (and compensation) starting with &quot;Engineering Manager&quot; or something similar being the equivalent of something like &quot;Staff Software Engineer.&quot; It then has its own titles like &quot;Senior Engineering Manager,&quot; and may go towards titles like &quot;VP of Engineering.&quot; All of these levels (for both IC and manager track) have their own expectations, which I won&apos;t cover here. There are good/comprehensive resources online, like <a href="https://dresscode.renttherunway.com/blog/ladder?ref=georgeyw.com" rel="noreferrer">this one</a>.</p><h3 id="benchmarking-levels-against-academia">Benchmarking levels against academia</h3><p>Ok, here&apos;s where the real hot takes begin. I&apos;m not going to bother justifying these, so take them with a grain of salt, but I think this is what I believe from my limited perspective.</p><ul><li>If you don&apos;t have a PhD in computer science or machine learning, expect to enter tech or quant at the junior or mid-level. You might be able to get a data science role at a senior level if you did e.g. probability and do well on your interviews.<ul><li>The startup I first worked for didn&apos;t have levels in a formal sense until a year after I joined, but I think I was retroactively initially placed at the junior level and then promoted to mid-level after a year.</li></ul></li><li>If you can get a PhD in math, you can reach senior level in tech, full stop. Tech jobs are high paying because 1) tech companies make a shitload of money, and marginal engineers really do actually bring a lot of value to the company at scale and 2) because the jobs are competitive and hard and have high expectations. But crucially, they are competitive and hard and have high expectations <em>relative to the typical job in America</em>. They are <em>not</em> more competitive or harder or have higher expectations than earning a PhD (modulo personal fit for tech, some people may just find software easier than others).</li><li>Postdocs up through tenured professors are probably somewhere around the &quot;staff&quot; level equivalent for academia.</li><li>I&apos;m less sure about this, but I&apos;d guess that full professor in academia is comparable to principal level in tech or to the equivalent management track role (hey, remember that 7 figure compensation number above?)</li><li>These don&apos;t mean they&apos;re where your <em>cap</em> is, but I would confidently place money at 1:1 odds (and probably much worse odds, though I don&apos;t want to commit to figure those out here) that someone with the given level of academic accomplishment can first reach the given comparable tech level in a typical amount of time</li><li>In general, the number of people who reach each progressive level in the tech ladder tends to decrease exponentially, and the difficulty and expectations of the levels also increase substantially past senior. It is <em>totally fine</em> to hit ~senior level and just chill there forever and do other stuff with your time, like fun side projects or your math research or just having more time to enjoy life in general.</li></ul><h2 id="a-personal-request">A personal request</h2><p>If you&apos;ve gotten this far, hopefully this resource has been useful to you. I am happy if this improves your life in any way. My only request to you is that you please try to avoid working <em>directly</em> on AI capabilities as a result of this advice.</p><p>I&apos;ve <a href="https://www.georgeyw.com/slowing-down-pandoras-box/" rel="noreferrer">written about</a> AI safety previously and have <a href="https://www.georgeyw.com/slowing-down-pandoras-box/" rel="noreferrer">pivoted my career</a> to technical safety research because I think this is such an important problem. This is not asking you to say, turn down the only offer you get and not be able to support yourself or your family, but to please think about this if you find yourself having a choice in the matter. Most job opportunities don&apos;t fall under this umbrella, and it&apos;ll probably be clear if a given job is aimed at significantly advancing AI, so you mostly have to be specifically looking for such jobs. </p>]]></content:encoded></item><item><title><![CDATA[Sharing is caring: in defense of ordinary things]]></title><description><![CDATA[On foggy mornings and small talk]]></description><link>https://www.georgeyw.com/sharing-is-caring/</link><guid isPermaLink="false">65561af72bc2fa000111b8c2</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sun, 19 Nov 2023 17:53:11 GMT</pubDate><content:encoded><![CDATA[<p>I visited Oxford last week for a conference-workshop-retreat-thing (working title), which means I&apos;m super jet lagged this week and have been temporarily consigned to the life of a morning person. I&apos;m normally an extreme night owl, and the rare times I&apos;m awake at daybreak are at the wrong end of the day. So when I begrudgingly got out of bed at 6am this morning (a notable improvement from 3am), I figured I&apos;d go to the park and have my coffee with the sunrise for once*.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.georgeyw.com/content/images/2023/11/fog_atx.jpeg" class="kg-image" alt loading="lazy" width="2000" height="1190" srcset="https://www.georgeyw.com/content/images/size/w600/2023/11/fog_atx.jpeg 600w, https://www.georgeyw.com/content/images/size/w1000/2023/11/fog_atx.jpeg 1000w, https://www.georgeyw.com/content/images/size/w1600/2023/11/fog_atx.jpeg 1600w, https://www.georgeyw.com/content/images/size/w2400/2023/11/fog_atx.jpeg 2400w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">nice, quiet Silent Hill vibes</span></figcaption></figure><p>Dear reader, it was foggy and overcast. There was no sunrise. I&apos;m not upset, but thanks for asking.</p><p>While I was walking, I thought back to the conversations I had in Oxford and to previous work retreats before that. I&apos;ve only experienced remote work, and a lesson that I keep relearning is how impactful meeting in-person is. It&apos;s clearly helpful for productivity &#x2013; for about a week, everyone&apos;s primary focus is on planning, on sharing knowledge, on aligning the big picture. But I don&apos;t think this is actually the most important part. There&apos;s a key social cohesion created, and not just from intentionally designed icebreaker activities. </p><p>When you first interact with someone, you get the broad strokes that everyone gets. He&apos;s a teacher, she&apos;s an engineer, they&apos;re a parent, a spouse, a sibling. Work intros often start with everyone rattling off education history, past jobs, and maybe a well-rehearsed fun fact. Basic knowledge is of course a necessary starting point, but how much of a picture do you really get of someone from that?</p><p>These are the things that often come to mind when we think of our identity, but billions of other people in the world are, say, cat or dog people, and I&apos;d take short odds that there&apos;s more variance within one of those groups than between them. You can learn a dozen facts about a person and still be no closer to understanding them or building a close relationship.</p><p>Most people never see the the small, mundane, in-between moments in our lives, but I think this is where intimacy lies. I don&apos;t identify with a foggy morning; it doesn&apos;t feel like a part of who I am in any way. And yet, I&apos;d only share this kind of story with a few people (and you of course, dear reader). I don&apos;t broadcast this to the world (kindly ignore that I&apos;m publicly writing about it) because it isn&apos;t that important, and most people just won&apos;t care. For the people that do, they care because it&apos;s a story about <em>me</em>, because someone in their tribe experienced it, and because it means something to be one of the first people that comes to mind when anything happens.</p><p>I&apos;m not saying that spending a week in person gets you anywhere near that level of closeness. But having such dense overlap in organic moments sure seems like a bigger step forward than as many video calls to hash out business logistics. There&apos;s no choice but to experience the ordinary together and much more space to fill with mundane questions (how was your morning?) and mundane stories (the funniest thing happened...). Maybe there&apos;s something like the <a href="https://en.wikipedia.org/wiki/Ben_Franklin_effect?ref=georgeyw.com" rel="noreferrer">Ben Franklin effect</a> happening too &#x2013; if sharing is caring, then perhaps caring follows sharing as much as the reverse.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.georgeyw.com/content/images/2023/11/Untitled_Artwork-1.png" class="kg-image" alt loading="lazy" width="1800" height="1700" srcset="https://www.georgeyw.com/content/images/size/w600/2023/11/Untitled_Artwork-1.png 600w, https://www.georgeyw.com/content/images/size/w1000/2023/11/Untitled_Artwork-1.png 1000w, https://www.georgeyw.com/content/images/size/w1600/2023/11/Untitled_Artwork-1.png 1600w, https://www.georgeyw.com/content/images/2023/11/Untitled_Artwork-1.png 1800w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">the frontier of sharing advances or recedes with closeness (</span><a href="https://twitter.com/RichDecibels/status/1716140468751204826?ref=georgeyw.com"><span style="white-space: pre-wrap;">h/t</span></a><span style="white-space: pre-wrap;"> for format)</span></figcaption></figure><p>I also notice that a drop-off in mundane sharing predicts an eventual drop-off in closeness**. I think this makes sense; most of what happens in our lives isn&apos;t news-worthy, so not talking about the boring stuff means much less contact overall. But beyond this, there&apos;s probably something powerful in gradually losing the urge to share and in the worry that you might be bothering someone with something uninteresting rather than feeling secure in the idea that anything is interesting to them if you&apos;re the main character. Sometimes the last things to develop are the first to go.</p><hr><p>*Technically this happened a few days ago by time of publishing, but I wrote that line on the day it happened. Don&apos;t sue me.</p><p>**I&apos;m also noticing while writing that I&apos;m generally not sharing much of the boring stuff with <em>anyone</em> right now (so for my close friends reading this, don&apos;t go off thinking I secretly hate you in particular because I didn&apos;t tell you about a non-sunrise). This seems like a sign that I&apos;m getting too isolated and need to give more attention to my social wellbeing.</p>]]></content:encoded></item><item><title><![CDATA[Moving to AI safety]]></title><description><![CDATA[Kicked up some dust this year, some was settled]]></description><link>https://www.georgeyw.com/moving-to-ai-safety/</link><guid isPermaLink="false">653f7b2a473c9b0001aae606</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Tue, 14 Nov 2023 16:44:00 GMT</pubDate><content:encoded><![CDATA[<figure class="kg-card kg-gallery-card kg-width-wide kg-card-hascaption"><div class="kg-gallery-container"><div class="kg-gallery-row"><div class="kg-gallery-image"><img src="https://www.georgeyw.com/content/images/2023/11/boat-sunset.jpg" width="2000" height="2666" loading="lazy" alt srcset="https://www.georgeyw.com/content/images/size/w600/2023/11/boat-sunset.jpg 600w, https://www.georgeyw.com/content/images/size/w1000/2023/11/boat-sunset.jpg 1000w, https://www.georgeyw.com/content/images/size/w1600/2023/11/boat-sunset.jpg 1600w, https://www.georgeyw.com/content/images/size/w2400/2023/11/boat-sunset.jpg 2400w" sizes="(min-width: 720px) 720px"></div><div class="kg-gallery-image"><img src="https://www.georgeyw.com/content/images/2023/11/boat-whiteboard-blurred-2.png" width="1700" height="1536" loading="lazy" alt srcset="https://www.georgeyw.com/content/images/size/w600/2023/11/boat-whiteboard-blurred-2.png 600w, https://www.georgeyw.com/content/images/size/w1000/2023/11/boat-whiteboard-blurred-2.png 1000w, https://www.georgeyw.com/content/images/size/w1600/2023/11/boat-whiteboard-blurred-2.png 1600w, https://www.georgeyw.com/content/images/2023/11/boat-whiteboard-blurred-2.png 1700w" sizes="(min-width: 720px) 720px"></div><div class="kg-gallery-image"><img src="https://www.georgeyw.com/content/images/2023/11/fall-leaves.jpg" width="2000" height="2667" loading="lazy" alt srcset="https://www.georgeyw.com/content/images/size/w600/2023/11/fall-leaves.jpg 600w, https://www.georgeyw.com/content/images/size/w1000/2023/11/fall-leaves.jpg 1000w, https://www.georgeyw.com/content/images/size/w1600/2023/11/fall-leaves.jpg 1600w, https://www.georgeyw.com/content/images/size/w2400/2023/11/fall-leaves.jpg 2400w" sizes="(min-width: 720px) 720px"></div></div></div><figcaption><p><span style="white-space: pre-wrap;">Regrettably no photos of laser tag in a castle - have a picture of a lecture on a lake instead</span></p></figcaption></figure><p>It&apos;s been an eventful year! I quit my job, tried to start a company, had a summer fling with freelancing (it was awesome), and did a ton of soul searching. It&apos;s also been a bit of an emotional rollercoaster: lots of stress, lots of confusion, lots of fun (from New England sunset boat rides to laser tag in a <a href="https://www.wythamabbey.org/?ref=georgeyw.com" rel="noreferrer">castle</a>). The year&apos;s not over yet, so there&apos;s still time for curveballs, but it looks like some dust is finally settling.</p><p>The dominating consequence of my soul search was the conviction that AI safety is the <a href="https://www.lesswrong.com/posts/P5k3PGzebd5yYrYqd/the-hamming-question?ref=georgeyw.com" rel="noreferrer">most important thing I can work on</a> right now. I spent the better part of the last 6 months exploring how rapid AI development will transform our world, and the world in turn kept smacking me in the face with <a href="https://www.georgeyw.com/slowing-down-pandoras-box/" rel="noreferrer">how wide the cone of possibilities</a> was. Unfortunately, not all outcomes are positive &#x2013; I now believe that this is the <a href="https://www.lesswrong.com/s/yYxggfHYRrqnJXuRx?ref=georgeyw.com" rel="noreferrer">most perilous century</a> for humanity, with AI being the primary driver. </p><p>The hype around AI &#x2013; that it could solve cancer, poverty, conflict, disease, even aging and death &#x2013; massively accelerates its development. To be clear, AI safety shouldn&apos;t be <a href="https://a16z.com/the-techno-optimist-manifesto/?ref=georgeyw.com" rel="noreferrer">conflated</a> with the belief that approaching <a href="https://www.lesswrong.com/posts/Py3uGnncqXuEfPtQp/interpersonal-entanglement?ref=georgeyw.com" rel="noreferrer">Reedspacer&apos;s Lower Bound</a> is inherently bad. I hope we cure all disease! I hope we get functionally infinite prosperity! I am <em>so</em> on board with solving death! It would be tragic if we <em>never</em> achieved safe artificial superintelligence.</p><p>But hype breeds recklessness. Almost all of our forecasted extinction risk this century <a href="https://www.metaculus.com/questions/2568/ragnar%25C3%25B6k-seriesresults-so-far/?ref=georgeyw.com#comment-133441" rel="noreferrer">comes from AI</a> &#x2013;  many times more than from nuclear war, climate change, and bio-risk combined. And development goes on accelerating anyways. I want all the great things that AI promises, I just think we can get them for a better price. </p><figure class="kg-card kg-image-card"><img src="https://www.georgeyw.com/content/images/2023/11/image.png" class="kg-image" alt loading="lazy" width="478" height="150"></figure><p>So what&apos;s next for me? I&apos;m excited to work towards a safer future with <a href="https://www.lesswrong.com/posts/nN7bHuHZYaWv9RDJL/announcing-timaeus?ref=georgeyw.com" rel="noreferrer">Timaeus</a>, a brand spankin&apos; new technical AI safety research org. I&apos;m hopping on board to help scope out a new field of <a href="https://www.lesswrong.com/tag/interpretability-ml-and-ai?ref=georgeyw.com" rel="noreferrer">interpretability</a> research as a research assistant/engineer for at least the next ~6 months. I&apos;m also frantically cramming Spanish for the <a href="https://aifuturesfellowship.org/?ref=georgeyw.com" rel="noreferrer">AI Futures Fellowship</a> in Mexico City early next year, where I&apos;ll hopefully develop a broader base of knowledge, meet brilliant people, and eat far too much food.</p><h3 id="the-future-is-unclear">The future is unclear</h3><p>I think more people should spend time thinking hard about AI. Whatever consequences you think there will be, it seems hard to get away from the conclusion that they&apos;ll be transformative in <em>some</em> direction. But <a href="https://twitter.com/littIeramblings/status/1708945586496446796?ref=georgeyw.com" rel="noreferrer">sensemaking</a> about AI is <a href="https://michaelnotebook.com/xrisk/index.html?ref=georgeyw.com" rel="noreferrer">really hard</a>, especially if you don&apos;t have a technical background. Even as someone with the relevant background, I spent months highly confused about the right path to take and still feel fairly uncertain. We can and should at least still strive to be directionally correct though, and voting for safe policies is just as critical as technical research (if not more). </p><p>Finally, I think this is the ~correct direction (or <em>a</em> correct direction) for me to be moving in right now, but it&apos;s all pretty nuanced, and I&apos;m always open to more perspectives. Please reach out if you&apos;re curious or have thoughts you&apos;d like to share. Don&apos;t be a stranger, odds are I&apos;d love to chat &#x2013; everyone who&apos;s let me talk their ear off so far can attest to that.</p>]]></content:encoded></item><item><title><![CDATA[Slowing down Pandora's Box]]></title><description><![CDATA[Editorial for the Canadian Science Policy Centre]]></description><link>https://www.georgeyw.com/slowing-down-pandoras-box/</link><guid isPermaLink="false">64e43d29460d3c0001af6b7f</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Tue, 22 Aug 2023 04:53:17 GMT</pubDate><content:encoded><![CDATA[<p><em>Originally written for an </em><a href="https://sciencepolicy.ca/posts/slowing-down-pandoras-box/?ref=georgeyw.com"><em>editorial series</em></a><em> for the Canadian Science Policy Centre.</em></p><hr><p>This thought keeps coming back to me lately: <em>everything is about to get really weird</em>. For the past two years, I&#x2019;ve had a front-row seat to the primordial soup of AI research and startups. I&#x2019;ve built a couple of products based on language models, consulted on generative music tools, and dabbled across the board of generative AI, including image, audio transcription, voice, and more. A massive wave of commercial products is coming, for which ChatGPT is just the tip of the iceberg.</p><p>AI has been fairly consistent in progressing exponentially. The thing about exponential growth is that for a long time, it looks like nothing is happening, then it looks like something is finally happening, and then while you&#x2019;re still processing the something that&#x2019;s happening, the growth explodes and&#x2026; uh oh, suddenly everything is weird now. We&#x2019;re still in the &#x201C;huh, it looks like something is finally happening&#x201D; stage with AI. </p><p>A certain recent pandemic demonstrated that we&#x2019;re not so good at dealing with the next phase of exponential growth. But covid didn&#x2019;t care what was or wasn&#x2019;t intuitive to us; the world just changed overnight anyway. We scrambled to react and flatten the curve, but not before it was already a foregone conclusion that there <em>would</em> be a huge curve. When it comes to AI, it&#x2019;s fortunately not all downside like it is with a global pandemic. AI has incredible potential to bring prosperity &#x2013; some have even compared it to the discovery of fire or electricity. But there&#x2019;s a darker side, and that side grows exponentially too.</p><p>While writing this, I hopped on X (formerly known as Twitter, RIP) and saw a post on <a href="https://twitter.com/nearcyan/status/1687945164378058752?ref=georgeyw.com">practical acoustic side-channel attacks on keyboards</a>. The model in that paper can correctly identify 93% of your keystrokes from the audio that a Zoom call picks up. I see news like this more often than is comfortable, and then I predictably think to myself that everything is about to get really weird. You have to really stand out these days to be noteworthy &#x2013; a few months ago, a research team managed to <a href="https://medarc-ai.github.io/mindeye/?ref=georgeyw.com">read the human mind</a>(!) with &gt;90% accuracy by training a model on fMRI scans. [<em>Note: since writing this, we&apos;ve managed to </em><a href="https://news.berkeley.edu/2023/08/15/releases-20230811?ref=georgeyw.com"><em>reconstruct music</em></a><em> by reading brainwaves too.</em>]</p><p>The optimistic among us would hope in vain that AI tech isn&#x2019;t abused. FraudGPT and WormGPT are two cousins of ChatGPT, repurposed for enabling cyberattacks and scams. ChaosGPT was a (fortunately ineffective) bot whose task was literally to destroy humanity. What happens as we progress along the exponential growth curve and the people building these harmful AIs get access to shiny new toys? Dario Amodei, CEO of Anthropic, <a href="https://www.judiciary.senate.gov/imo/media/doc/2023-07-26_-_testimony_-_amodei.pdf?ref=georgeyw.com">testified to the US Congress</a> that a straightforward extrapolation of today&#x2019;s systems suggests that large-scale biological attacks could be possible in only two or three years. He <a href="https://twitter.com/profoundlyyyy/status/1684333960753565696?ref=georgeyw.com">went on to say</a> that if we don&#x2019;t have mechanisms in place to restrain AI systems by 2025-26, &#x201C;we&#x2019;re gonna have a really bad time.&#x201D;</p><p>Two to three years isn&#x2019;t a lot of time, and we don&#x2019;t even know what the right restraining mechanisms are yet! It would be great to have a bit more time to figure things out before the exponential progress curve blows past us. Fortunately, unlike with the pandemic, we can actually make that happen for AI. The Future of Life Institute wrote an <a href="https://futureoflife.org/open-letter/pause-giant-ai-experiments/?ref=georgeyw.com">open letter</a> with over 33,000 signatures calling for a six-month moratorium on AI development, signed by many of the biggest names in AI research and tech. It&#x2019;s not clear that the specific plan of a six-month moratorium would be most effective, but this seems directionally correct. Policymakers might consider prophylactic legislation that would allow such a moratorium to come into effect if certain conditions are met. Compute governance is a similar slowing mechanism, where large training runs of AI models are limited in the amount of compute power they&#x2019;re allowed to use. The threshold can be adjusted over time to have some control over the pace of development without it being all or nothing. These aren&#x2019;t long-term solutions, but they would buy time to figure out more permanent measures.</p><p>Besides short timelines to serious risks, another core challenge with making AI safe is the asymmetry between offense and defense. There are simply too many angles of attack to be able to cover all vulnerabilities &#x2013; but policy can reshape the AI landscape into something more defensible. When advanced AI is too openly accessible, any individual can choose to take one of those many angles of attack, so an obvious solution might be to prevent advanced AI from being too openly accessible. </p><p>We can accomplish that via regulatory action, but this is a pretty controversial topic in the tech world; open-sourcing (making publicly available) your research and your code is usually synonymous with nice things such as transparency, collaboration, and even security since you have more eyes on fixing vulnerabilities. Furthermore, a common counterargument to closing off access to AI technology is that it centralizes too much power in the hands of the few major AI labs and that we can only have a fair balance of power if the technology is available to everyone. </p><p>But the world is not always a safer place when people have more access to technology. The US has a lot of guns and &#x2013; surprise &#x2013; <a href="https://www.nytimes.com/2017/11/07/world/americas/mass-shootings-us-international.html?ref=georgeyw.com">a lot of mass shootings</a>. If everyone had a nuke in their pocket, it would only be a matter of time before most urban cities disappeared in a mushroom cloud. All it takes is one person doing the wrong thing. And if everyone has a large-scale-biological-attack-capable AI on their home computer, Dario Amodei put it well: we&#x2019;re gonna have a really bad time.</p><p>It doesn&#x2019;t seem likely that we can entirely close Pandora&#x2019;s box on AI, but we can make it manageable. Everything is about to get really weird, but we still have the power to decide if that&#x2019;ll be a good weird or a bad one.<br></p>]]></content:encoded></item><item><title><![CDATA[My PhD cost me $1.17M (but I'd do it again)]]></title><description><![CDATA[PhD advice + reflections on opportunity cost ]]></description><link>https://www.georgeyw.com/my-phd-costed-me-1-17m-so-far-but-id-do-it-again/</link><guid isPermaLink="false">64987776bea67900017b92af</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Sun, 02 Jul 2023 20:07:47 GMT</pubDate><content:encoded><![CDATA[<p>Not from my bank account, just from opportunity cost. It&#x2019;s the first time I&#x2019;ve sat down and calculated that. I don&#x2019;t think I was ready for the sticker shock, I just thought it could make for a catchy title. Honestly, this might say as much about big tech salaries as it does about PhDs, but the counterfactual history was a reasonable one where I took a job at Google and took a standard path to senior engineer.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh6.googleusercontent.com/shqDEBVd58zG1q1AFGk43NuC5iw2UTK2lMRh6aPxzmCUesKSEXpjbdem5letzILQG_P1NBMHQhleW7zKaKLSUQWDWYGaV_UIPm493z1FZHTfWxA3CzspY-fDQEmUzoL9VXx1YUO9Ev8Y3P0BXlEkCxA" class="kg-image" alt loading="lazy" width="561" height="312"><figcaption><i><em class="italic" style="white-space: pre-wrap;">Current Google Senior SWE compensation, courtesy </em></i><a href="http://levels.fyi/?ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">levels.fyi</em></i></a></figcaption></figure><p>To get that number, I made some salary approximations, calculated taxes and cost of living, and was naively surprised at <a href="https://www.investopedia.com/ask/answers/042415/what-average-annual-return-sp-500.asp?ref=georgeyw.com">S&amp;P 500 returns</a> in a zero interest rate environment, which managed to account for a substantial amount of the difference. The total gap will increase over time too, through the interminable power of compounding returns. Your ultimate career path might not have quite such inflated salaries, and zero interest rates probably aren&#x2019;t coming back any time soon, but the opportunity cost for a PhD may still be much higher than you expect.</p><p>Some people hold the position that they should do their best to deter anyone and everyone they can from getting a PhD. Their belief is that you&#x2019;re fit for one if and only if you ignore that advice and boldly forge ahead anyways.</p><p>I think this is unnecessary. There are plenty of real reasons not to do a PhD (I have at least 1.17M of them), and one should have a better reason to do one than an inability to absorb advice. These considerations, and my own story, are the best I can offer to you if you&apos;re a prospective student; the final decision is up to you.</p><h2 id="the-happy-path-of-grad-school">The happy path of grad school</h2><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh5.googleusercontent.com/MrP9MkK0xrLlWYvtLsw_GbNMrWQKtvTlXkgZZ5E6Hyl6DRzkSjsHuQ8Z9E4sdNOCUuzoZSBd9GsW9_oRVD_AOSSPcU8xQvGz3dtPxJ29S23ODbztv1JrWBe1YMswN_k7L6AVgK0CZEIneChsPSz3UZs" class="kg-image" alt loading="lazy" width="223" height="311"><figcaption><a href="https://gatherer.wizards.com/Pages/Card/Details.aspx?multiverseid=394746&amp;ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">This</em></i></a><i><em class="italic" style="white-space: pre-wrap;"> could be you! (Fun fact: </em></i><a href="https://en.wikipedia.org/wiki/Richard_Garfield?ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">Richard Garfield</em></i></a><i><em class="italic" style="white-space: pre-wrap;">, the creator of Magic: The Gathering, went to the same grad program as me. No wonder he&#x2019;s so successful!)</em></i></figcaption></figure><p>Whatever path you choose in life, you&#x2019;ll take some form of opportunity cost along the way; the important thing is just to notice the tradeoff. For example, last I checked, it was quite expensive to raise children in the US. Most people go on to have children anyways, and many go on to be happier with their kids than the money would make them. An econ PhD friend of mine talks about a concept of &#x201C;Life <a href="https://en.wikipedia.org/wiki/Expected_value?ref=georgeyw.com">EV</a>&#x201D;: sometimes we do things that are financially suboptimal because they sufficiently improve our life along other dimensions (he often yells this as he&apos;s halfway through leaping into a clearly bad decision because it seems fun).  </p><p>But just as you shouldn&#x2019;t have kids for the sole reason of hoping you have someone to visit you in your nursing home, you shouldn&#x2019;t only think of a PhD in terms of what comes after. The best reason to do a PhD, and one which is both necessary and sufficient, is that you simply want to experience the journey.</p><p>That journey can be a beautiful one, if you&#x2019;re <a href="https://www.georgeyw.com/you-dont-take-enough-risks-to-be-lucky/">lucky</a>. You&#x2019;ll meet other like-minded, bright-eyed grad students. You&#x2019;ll meet generous mentors who pay things forward like it is their moral duty. You&#x2019;ll travel to conferences (on someone else&#x2019;s dime!) <a href="https://fpsac.org/confs/?ref=georgeyw.com">across the world</a>, and you&#x2019;ll meet more people that care about your <a href="https://matt.might.net/articles/phd-school-in-pictures/?ref=georgeyw.com">tiny corner of reality</a> than you could have imagined existed (there were <em>dozens</em> of us). </p><p>You can work on the kind of problem that follows you to the shower. The kind that itches your brain at night. The kind that jolts you awake as you lay in bed, because a piece finally clicked into place and you can&#x2019;t hope to sleep until you write it down. The kind that gives you an adrenaline high for days when you figure it out, because you&#x2019;ve been working on it for months or years.</p><p>There are other reasons that help, too. It&#x2019;s hard to believe you can tackle a challenge that no one else has, especially if the pressure of completing your thesis hangs over you. A PhD is training and proof of your ability to grapple with the unknown. </p><p>You might want to do something you can only do with a PhD, or which a PhD can open doors for. In the world of math, this could mean working at a university, but it could also mean working at a hedge fund, at a national lab, or at an AI lab of a big tech company. Most things don&#x2019;t need a PhD in the strictest sense, but people do pay attention when you have one.</p><p>These things are icing on the cake, but at the end of the day, the surest way to make it a positive, worthwhile experience is to <em>want the experience</em>. A PhD should not feel like paying your dues, and it is not noble to suffer through one. An important corollary is that if you are decidedly <em>not</em> enjoying the experience, it&#x2019;s ok to leave!</p><h3 id="a-note-on-slack">A note on &quot;slack&quot;</h3><p>It can be a little paradoxical to describe grad school as a time with a lot of <a href="https://slatestarcodex.com/2020/05/12/studies-on-slack/?ref=georgeyw.com">slack</a>. After all, grad students have a reputation for being over-worked, under-paid, over-caffeinated, and under-rested. But it isn&apos;t always like that, and there is a unique thread of intellectual freedom through all of it.</p><p>For me, that freedom manifested as space to grow. Once I finished most of my coursework (first two years), the only expectation seemed to be that I spend a lot of time thinking. In theory, that means thinking about your research, but in practice, my mind still managed to wander for many hours a day. </p><p>That space expanded considerably in my last two years; the pandemic happened, I&apos;d already decided to leave academia, and I had enough research to graduate. I had few expectations to fulfill, and any remaining research work was mainly tying up loose ends. This was an unusual abundance of slack, but it led to substantially more hours being alone in my head. </p><p>I used to think this time was a bit wasted. The amount of concrete things that came out of those last couple of years felt pretty dismal in number. I was probably burnt out for a substantial portion of it, and the social isolation was hard.</p><p>Looking back, those years were probably some of the richest years of personal growth that I&apos;ve ever had. Among other things, I developed a lot more self-awareness and confronted large chunks of emotional baggage that I never previously noticed.</p><p>This constitutes a large part of why I would make the same choice to do a PhD, if I could go back and choose again. It also highlights another kind of opportunity cost, which is of time. Whether I spent 5 years in grad school or at Google, those 5 years would pass either way. I only got to spend my early 20s once, and I&apos;m glad that I spent them the way I did. Just as money in the bank would compound year over year, I can benefit from that space to grow for the rest of my life.</p><h2 id="the-darker-side-of-grad-school">The darker side of grad school</h2><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh3.googleusercontent.com/hmwn6EPkFGl1pRNWskNMMkhA-K3EMp-qvJxoi_8n-gnU_I-GC6Yg3KAN-iE0yuaadMGQ-3nq13dPVjSLm4mjYgpw1AQWdhtoZj4X46sfRBLGy4x3D32iLpffSCjTMuO6-c4d1lra6DoxMIr4bnewaX0" class="kg-image" alt loading="lazy" width="265" height="370"><figcaption><a href="https://gatherer.wizards.com/Pages/Card/Details.aspx?multiverseid=547907&amp;ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">This</em></i></a><i><em class="italic" style="white-space: pre-wrap;"> could be you too!</em></i></figcaption></figure><p>The PhD journey can be a beautiful one, if you&#x2019;re lucky. If you&#x2019;re unlucky, academia is a <a href="https://www.benkuhn.net/grad/?ref=georgeyw.com">meat grinder</a>. There are various opportunity costs to consider, but there are real costs too, mostly in the form of stress to the mind and spirit. From my limited view, the following things were the main culprits.</p><h3 id="impostor-syndrome">Impostor syndrome</h3><p>Academia has a high density of brilliant people. Many academics, including ones you will think are far more brilliant than yourself, <a href="https://en.wikipedia.org/wiki/Impostor_syndrome?ref=georgeyw.com">feel they don&#x2019;t belong</a>. They do, and you will too if you choose to join them. There is an entire post that can be written about this topic, but the short version is that it is important to know that you may feel this way, to know that it is a normal and well-worn path, and to know that there are resources for this.</p><h3 id="social-isolation">Social isolation</h3><p>It can be intoxicating to be consumed by your work, but the darker side is that you can end up disconnected from and left behind by your social network. PhD programs are selective, so you can&#x2019;t be picky about which city you end up in. I think this is especially hard for overseas students, where even phone calls are difficult to coordinate thanks to time zones. </p><p>You&#x2019;ll see other people moving on to the next stage of their life, starting their careers, starting families, buying houses, cars, vacations, and other shiny things. Your family and friends may valiantly try to understand your work, but they probably won&#x2019;t. Your cohort might end up a bit south of social and outgoing (n.b. you are <a href="https://notebook.drmaciver.com/posts/2020-03-17-11:40.html?ref=georgeyw.com">allowed</a> to be the one that organizes social events!).</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh6.googleusercontent.com/bgoPjg9zskSnJEivHUIHDispRDEsTSBd-OxfjOoHhTv_ssXZ-b6WKfytrRB3KFnN61Pe-UoW8FOyi_GkG54uvCAujvp1IxPNTbkZllrMMg5O4OIRQNyd4evoG5z69zjt0tsISxIVk0iIxMwlGTytDjE" class="kg-image" alt loading="lazy" width="443" height="425"><figcaption><i><em class="italic" style="white-space: pre-wrap;">Pictured: one of the reasons I made it through in one piece</em></i></figcaption></figure><p>My advice for this is to find a hobby with a strong community around it (preferably a physical one; you get exercise and get to meet people in person). I took up rock climbing in my first year, and it changed my life.</p><h3 id="a-bad-advisor-relationship">A bad advisor relationship</h3><p>Professors are people too, and in any group of people there will be unsavory individuals. Some rare professors have a reputation for dealing in politics, gossip, and lies. It is not always easy for a fresh grad student to figure out who these are, but one should do everything in their power to avoid such professors. </p><p>Some other professors have very particular (and antagonistic) approaches to mentorship. One who shall not be named would intentionally fail students on their <a href="https://en.wikipedia.org/wiki/Comprehensive_examination?ref=georgeyw.com">quals</a>. Their alleged motivation for doing so was to use the pressure to pass quals as a motivational whip or <a href="https://en.wikipedia.org/wiki/Forcing_function?ref=georgeyw.com">forcing function</a> on their students. By doing so, they could compel their students to study additional foundational material for the second attempt. Think twice before working with someone like this.</p><p>Setting aside extreme examples, you might just not get along well with your advisor through no fault of yours or theirs. Working chemistry is fickle, and you&#x2019;re allowed to decide it isn&#x2019;t a good match. Shop around for advisors aggressively until you find a fit (or at least do the best you can with the size of your department), since from the point of picking an advisor onwards, they&#x2019;ll be the biggest factor in your journey. It can be worth picking your second favorite topic if it means working with your favorite professor.</p><h3 id="academia-or-bust">Academia or bust</h3><p>One of my undergrad math professors once told me how they got their job. While preparing to defend their thesis, their advisor got a phone call from a colleague asking if he had any good students finishing soon. That was it. No interview, no <a href="https://en.wikipedia.org/wiki/Postdoctoral_researcher?ref=georgeyw.com">postdoc</a> needed (typically a 3 year contract), they just discovered the next day that they had a job waiting for them. That was some 60 odd years ago. </p><p>Today, you&#x2019;ll fight to earn 1-2 postdocs, then fight even harder to earn a tenure track position, then work on earning tenure (usually a 6 year process). Although it&#x2019;s uncommon to be denied tenure once you have a tenure-track position, that&#x2019;s still a total of ~16 years from the start of grad school before you can be completely sure of stable employment. A <a href="https://forum.effectivealtruism.org/posts/3TQTec6FKcMSRBT2T/estimation-of-probabilities-to-get-tenure-track-in-academia?ref=georgeyw.com">pretty low</a> (but not insignificant) portion of PhD students that want tenure eventually make it.</p><p>I believe in chasing dreams; if tenure is your dream, then aim for tenure. Just don&#x2019;t hinge all of your expectations and happiness on that outcome. Luck is enough of a factor that brilliancy won&#x2019;t inoculate you. <a href="https://en.wikipedia.org/wiki/Yitang_Zhang?ref=georgeyw.com">Yitang Zhang</a> proved a <a href="https://terrytao.wordpress.com/career-advice/on-the-importance-of-partial-progress/?ref=georgeyw.com">partial result</a> of one of the most <a href="https://en.wikipedia.org/wiki/Twin_prime?ref=georgeyw.com#Twin_prime_conjecture">famous open problems</a> in math in 2013. He was incredibly persistent, but prior to this major result, still found himself living out of his car and working at Subway (I think this is exceptionally <em>unlucky </em>though).</p><p>At the end of the day, there are plenty of other <a href="https://en.wikipedia.org/wiki/Deep_tech?ref=georgeyw.com">places</a> <a href="https://www.nasa.gov/?ref=georgeyw.com">you</a> <a href="https://nationallabs.org/?ref=georgeyw.com">can</a> <a href="https://en.wikipedia.org/wiki/Big_Tech?ref=georgeyw.com">do</a> <a href="https://en.wikipedia.org/wiki/Hedge_fund?ref=georgeyw.com">research</a> <a href="https://en.wikipedia.org/wiki/Secondary_school?ref=georgeyw.com">or</a> <a href="https://en.wikipedia.org/wiki/Community_college?ref=georgeyw.com">teach</a>.</p><h3 id="unexpected-regrets">Unexpected regrets</h3><p>Few journeys end with no regrets. 5 years is a long time for regrets to find you, and they may not be what you expect. I am happy to have done a PhD, and I wouldn&#x2019;t change that decision. But my biggest regret was that it kept me away from home for what ended up being the last years of our family cat&#x2019;s life. She unexpectedly passed away shortly after defending my thesis, and before I could make it home to say goodbye.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh3.googleusercontent.com/9x_BFoRRyIhnEnnom7l080--Jp3MWvltKjgKVbNaz3vTRayrxqmm4fPC8oJBUQBHeVIf0GZ-mF8HsMyOW0ZfPwo0LbrxNVW4N2c2jcM9M8ASj9VzyYV7g9LlYYshCOpm7ueeFB1Ole839VPXBpGYGgw" class="kg-image" alt loading="lazy" width="598" height="83"><figcaption><i><em class="italic" style="white-space: pre-wrap;">From </em></i><a href="https://repository.upenn.edu/edissertations/4098/?ref=georgeyw.com"><i><em class="italic" style="white-space: pre-wrap;">my thesis</em></i></a></figcaption></figure><hr><h2 id="appendix-miscellaneous-advice-other-articles">Appendix: miscellaneous advice + other articles</h2><h3 id="external-links">External links</h3><ul>
<li>Read Terry Tao&#x2019;s <a href="https://terrytao.wordpress.com/career-advice/?ref=georgeyw.com">career advice</a>, it generalizes well. Some standouts for me:
<ul>
<li><a href="https://terrytao.wordpress.com/career-advice/does-one-have-to-be-a-genius-to-do-maths/?ref=georgeyw.com">Does one have to be a genius to do maths?</a></li>
<li><a href="https://terrytao.wordpress.com/career-advice/continually-aim-just-beyond-your-current-range/?ref=georgeyw.com">Continually aim just beyond your current range</a></li>
</ul>
</li>
<li>Sara Billey&#x2019;s advice on <a href="https://sites.math.washington.edu//~billey/advice/timely.fashion.pdf?ref=georgeyw.com">getting a PhD in a timely fashion</a>.
<ul>
<li>Choose your advisor systematically. Talk to current and former grad students. Are they happy? Do they have the kind of job that you want? Do they have jobs <em>at all</em>?</li>
<li>20 hours per week of focus time is enough. Hours in front of the TV don&#x2019;t count. Hours interrupted by texting and checking your phone don&#x2019;t count.</li>
</ul>
</li>
<li>John Baez&#x2019;s <a href="https://math.ucr.edu/home/baez/advice.html?ref=georgeyw.com">advice for young scientists</a>.</li>
<li>A great <a href="https://twitter.com/overthink_pod/status/1670952255217086464?ref=georgeyw.com">list of questions</a> before starting a philosophy degree.</li>
<li>A <a href="https://matt.might.net/articles/books-papers-materials-for-graduate-students/?ref=georgeyw.com">list of resources</a> by Matt Might.</li>
</ul>
<h3 id="on-teaching">On teaching</h3><ul><li>Put your heart into it when you do it, because you have a responsibility to your students, but set reasonable boundaries, because you have a responsibility to yourself, too.</li><li>It&#x2019;s easy to put infinite time into teaching. It feels productive compared to slamming your head against a research problem with no visible progress. Resist this urge. In the language of the <a href="https://asana.com/resources/eisenhower-matrix?ref=georgeyw.com">Eisenhower Matrix</a>, teaching is frequently important and urgent, while research is frequently important but not urgent. The latter category contributes most to long-term success.</li><li>Grade efficiently. If you can choose the grading rubric, simpler point rubrics are easier and faster to grade. The difference between an answer that is 2 out of 3 points and 3 out of 3 points is much easier to spot than 17 or 18 out of 20.</li><li>Grade fairly and predictably. Students <em>will</em> compare grades and answers and bring papers back to you with complaints (even if you grade perfectly, students will do that anyway, it&apos;ll just cost you less time).</li></ul><h3 id="other-notes">Other notes</h3><ul><li>Develop connections with people outside of your department. If you&apos;re gunning for academic positions, aim for at least one recommendation letter from someone outside. Your advisor can and should help you via warm intros.</li><li>People who brag about how much they work don&#x2019;t work as much as they say they do. Don&#x2019;t let it get in your head, and don&#x2019;t try to catch up. Grad school is a marathon and not a sprint.</li><li>People who brag about how <em>little</em> work they do often work more than they say they do. Some people really are built different, but for the most part, it isn&apos;t the norm to work very little and make visible progress.</li><li>There&#x2019;s some folk wisdom that you should go to a different institution at each stage of your career. The idea behind this is that each department (or research group/lab) has its own school of thought. This might hold true beyond grad school, but it&#x2019;s overblown at the undergrad -&gt; grad transition. Don&#x2019;t be afraid to stay at your undergrad institution for your PhD.</li><li>Prestige isn&#x2019;t everything, but <a href="https://www.insidehighered.com/news/2022/09/23/new-study-finds-80-faculty-trained-20-institutions?ref=georgeyw.com">it still matters more than it should</a>. If you want to stay in academia, then the prestige of your advisor can open doors for you. If you want to leave academia, people still pay attention to the university brand next to your degree.</li></ul><hr><p>Thank you to those who volunteered to read drafts: Alex, Felix, Ross</p><p>Feedback appreciated!</p>]]></content:encoded></item><item><title><![CDATA[You don't take enough risks to be lucky]]></title><description><![CDATA[How I think about long-tailed outcomes]]></description><link>https://www.georgeyw.com/you-dont-take-enough-risks-to-be-lucky/</link><guid isPermaLink="false">648e4ed91566530001d97a98</guid><dc:creator><![CDATA[George Wang]]></dc:creator><pubDate>Tue, 20 Jun 2023 06:10:08 GMT</pubDate><content:encoded><![CDATA[<p>I quit my job a month ago, which will probably end up among the best turning points of my career. To be clear, there was nothing <em>wrong </em>with that job, the alternative is just that much better. My social and professional networks have grown dramatically. I have the most intellectual freedom I&#x2019;ve ever had, letting me discover and work on [redacted], [redacted], and [redacted]. Above all, the degree of agency and ownership of my life and success is intoxicating. I might sober up eventually, but for now it&#x2019;s hard to imagine wanting to work a regular 9-5 any time soon.</p><p>There might be some collective wisdom on this. I honestly expected a lot more concern and muted skepticism whenever I revealed that I&#x2019;d quit my job. What I received instead was congratulations, excitement, and curiosity. My guess is that good things seem to happen when someone voluntarily makes a big life decision. Maybe they have an amazing opportunity waiting for them, maybe leaving their comfort zone creates amazing opportunities. Maybe I&#x2019;m just surrounded by relatively fortunate young people without enough financial commitments to make unemployment truly scary.</p><p>Anyways. The point of this isn&#x2019;t to convince you to quit your job. This was a fairly lucky outcome, but I didn&#x2019;t plan for that, and the point isn&#x2019;t to brag about it either. I&#x2019;ve had a lot of lucky outcomes in my life, and the point of this is to share how I made that happen and how you can be luckier, too.</p><h2 id="what-do-you-need-to-be-lucky">What do you need to be lucky?</h2><p>Luck is random, but it&#x2019;s not purely random. If we think of being lucky as having net positive impact from random chance in your life, then we can be skillful by selecting more opportunities where the odds are stacked in our favor. This is hard because the odds are usually too fuzzy to compute.</p><p>This implies a few prerequisites for practicing luck as a skill. First and foremost, you must have an <em>appropriate </em>amount of belief that you can have an impact on the seemingly random outcomes in your life. Our belief in this is heavily conditioned on how lucky we&#x2019;ve been, and we tend to over attribute skill when good things happen and over attribute random chance when bad things happen. As a result, there are people for whom this feels obvious and people for whom this feels impossible, but both parties are off the mark. Underconfidence means you never take enough shots on goal, but overconfidence means you&#x2019;ll bite off more than you can chew, and you&#x2019;ll really be gambling more than practicing luck as a skill.</p><p>You must also have an appetite and tolerance for risk. The goal here is to select more opportunities with odds stacked in your favor, but that still means taking a lot of opportunities where odds are involved. That means you will still lose a lot of bets. Humans are naturally risk-averse, and it will feel bad to lose bets. Dealing with this over time requires a high tolerance for failure, and you must have tolerance both in your mental game and in having <a href="https://en.wikipedia.org/wiki/Risk_of_ruin?ref=georgeyw.com">sufficient resources to absorb the negative outcomes</a>. Resources here can mean things like time, energy, money, social status, etc, and having more resources makes the mental game easier. Quitting my job was made much easier by knowing how long my savings could last and that I was fortunate to have a valuable skill set.</p><h2 id="it%E2%80%99s-a-rigged-game">It&#x2019;s a rigged game</h2><p>Fortune is not evenly distributed. Some people are born richer, some with more social status, some are more neurotypical, some more attractive, and on and on. People with more of these resources have access to more opportunities, and as a result, access to more opportunities with favorable odds. It is easier to afford the necessary risk of these opportunities if you have more resources. Although this post is not about the distribution of these resources, make no mistake: this is the truth and this is not fair.</p><p>That being said, you can make a meaningful impact on your future luck. Whatever your starting point is, <a href="https://danluu.com/p95-skill/?ref=georgeyw.com">you can beat a majority of people who don&#x2019;t try</a>, and most people don&#x2019;t try. It also gets easier over time; luck begets luck. Quitting my job was easier because of my savings and skill set, but the size of my nest egg mostly came from a couple of unique bets that turned out well. I chose my career from the overlap of things I like and things that pay well, which increased my access to good opportunities. Those in turn came from earlier, smaller things. Wherever you start, your luck can gradually snowball and change your life.</p><p>A closely related concept is <a href="https://articles.starcitygames.com/articles/learning-to-truly-play-to-your-outs/?ref=georgeyw.com">playing to your outs</a>, and nearly every life situation has outs, however improbable. These might only serve to make a terrible situation into a slightly less bad one, but they exist for those with the desire to find them. I sincerely hope you are not in such a terrible situation, but if you are, you probably aren&#x2019;t an exception. As harsh as that sounds, it&#x2019;s actually a good thing. The upshot is that <em>you can make a meaningful impact on your future luck</em>. The game is rigged, and the game is hard. We&#x2019;re stuck playing it anyway, so let&#x2019;s play with our eyes open.</p><h2 id="luck-is-making-chances-to-dig-for-outliers">Luck is making chances to dig for outliers</h2><p>The first hard thing about being lucky is realizing that <a href="https://milan.cvitkovic.net/writing/things_youre_allowed_to_do/?ref=georgeyw.com">you can just do things</a>. This is like the extreme version of noticing you can just choose to have cake for dinner. You have ownership over your life, and you are allowed to do more things than you realize to make it better.</p><p>The second hard thing is noticing all the places in your life where you might <em>want </em>to do something. When I say all, I really mean <em>all</em>. Noticing that you can quit your job isn&#x2019;t hard. Neither is noticing that you can go on a backpacking trip through Europe, that you can have kids, that you can exercise more, or that you can move to that city you&#x2019;ve always dreamed of living in. People talk about these things all the time. They&#x2019;re worth thinking about, but far from the only things worth thinking about. Sometimes exploring the full space of Bets You Can Make looks like <a href="https://www.benkuhn.net/abyss/?ref=georgeyw.com">staring into the abyss</a>. Sometimes it looks like ordering a dish you&#x2019;ve never tried before that sounds kind of weird.</p><p>The next step is evaluation; not all of these bets are good bets to take. I&#x2019;m going to focus on a particular type of bet, which I think is fortunately both the easiest to talk about and the most useful. That class is one where there&#x2019;s potential for huge outliers in outcomes. This is what that looks like as a graph:</p><figure class="kg-card kg-image-card"><img src="https://www.georgeyw.com/content/images/2023/06/image-3.png" class="kg-image" alt loading="lazy" width="2000" height="1266" srcset="https://www.georgeyw.com/content/images/size/w600/2023/06/image-3.png 600w, https://www.georgeyw.com/content/images/size/w1000/2023/06/image-3.png 1000w, https://www.georgeyw.com/content/images/size/w1600/2023/06/image-3.png 1600w, https://www.georgeyw.com/content/images/size/w2400/2023/06/image-3.png 2400w" sizes="(min-width: 720px) 720px"></figure><p>This type of distribution is called long-tailed, heavy-tailed, or right skewed. Line 1 shows roughly where the median outcome is. If you only take a few of these long-tailed bets, then you&#x2019;re likely to only end up with approximately median outcomes, and your luck will feel average. Line 2 shows roughly where the mean outcome is. The key here is that the downside outcomes are minimal and bounded, and the upside outcomes are massive and practically unbounded. If you take enough bets with this distribution of outcomes, then you&#x2019;ll eventually pick up enough outliers (line 3) to bring your net outcome well above the median. This is also <a href="https://www.nber.org/papers/w25113?ref=georgeyw.com">a theory</a> for how diversified portfolios manage to beat treasury returns. Most individual stocks end up with lower return than treasuries, but they&#x2019;re incredibly long-tailed, and it&#x2019;s the few outliers that cause a diversified portfolio to outperform treasuries.</p><p>Ben Kuhn has a great post on systematically <a href="https://www.benkuhn.net/outliers/?ref=georgeyw.com">searching for outliers</a> within specific domains, such as job selection, blog post writing, or startup investing. These tasks have common features:</p><ul><li>You can keep resampling from the distribution by taking new jobs, writing new posts, or investing in more startups</li><li>It&#x2019;s hard to tell how good the outlier outcomes can actually be, especially with low sample size</li><li>Scoring a few outliers covers the cost of searching and then some (and usually a lot more than some)</li></ul><p>One of the key takeaways is that finding success in those domains is largely a function of how well you can repeatedly sample from their outcome distribution.</p><p>My addition to this process is that you don&#x2019;t have to compartmentalize it to each specific domain. Suppose you have 1 domain with a specific outcome distribution and 100 other domains that all happen to have a similar outcome distribution. Assuming you&#x2019;re equally happy to get outliers in any of these domains, you can get as many expected outliers by sampling from the first, single domain 100 times, or by sampling from each of the 100 unique domains once each. In either case, you get to the same number of attempts at scoring outliers.</p><p>Long-tailed results also typically take a long time to be realized. You don&#x2019;t get immediate feedback on whether the job you picked or the partner(s) you&#x2019;re with is an outlier. There&#x2019;s a serial nature to these selection tasks that limits the number of shots you have in your lifespan. If it takes a couple of years to figure out if the job you&#x2019;re at is really exceptional, then focusing solely on resampling new jobs might leave you with only a couple dozen shots on goal in your entire life (and the longer it takes to find an outlier, the less you can reap the benefits). You can roughly double your shots on goal to find (any) outliers in life if you parallelize the process of resampling your job with resampling your romantic partner.</p><p>Why stop at two parallel processes? There are more unique domains than you realize that have long-tailed outcomes and that you can parallelize your sampling from <em>all </em>these domains. </p><p>Unfortunately, this is where the whole thing gets much harder. The biggest source of parallelization is going to be from decisions that are very specific to your life path and that you might only get to make a few times in your life. You might miss them, and if you see them, you might misevaluate them. There are a lot of these types of decisions, and not all of them have this long-tailed outcome distribution. Here is a quick, non-exhaustive list of things you might not get to do very often and which have varying degrees of difficulty, impact, and endorsement.</p><ul><li>Picking an advisor in grad school</li><li>Getting married</li><li>Having an expensive wedding</li><li>Getting divorced</li><li>Having kids or not</li><li>The number of kids to have</li><li>Adopting a kid</li><li>Adopting a pet</li><li>Choosing a career (especially a high-commitment one, like being a doctor)</li><li>Trade on a potential <a href="https://en.wikipedia.org/wiki/Black_swan_theory?ref=georgeyw.com">Black Swan</a> event</li><li>Committing major financial crimes</li><li>Buying a house</li><li>Emigrating to a new country</li><li>Donating a kidney</li><li>Undergoing gender-affirming care</li><li>Following a particular religion</li><li>Taking a particularly unique job abroad</li><li>Whistleblowing</li></ul><p>These are still things that are common or high profile enough that anyone can think of them, but access to opportunities is highly <a href="https://en.wikipedia.org/wiki/Path_dependence?ref=georgeyw.com">path dependent</a>.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh5.googleusercontent.com/TnOIOqxVHGsY42ZfvLMkA6wAaJ-i_3ckVfANrsaFmBg7t6kGUT5K4lF6bCT9pA8ioXBlfdPb03buGLr9W3pv7dv1wZ74TLP-rUxwkLFkgFR8PrVhxYJiXZd09YxiC_qbYKcFF4zeC-fVKdEc8WT4uFY" class="kg-image" alt loading="lazy" width="624" height="391"><figcaption><span style="white-space: pre-wrap;">Credit: </span><a href="https://twitter.com/waitbutwhy/status/1367871165319049221?lang=en&amp;ref=georgeyw.com"><span style="white-space: pre-wrap;">Wait but Why?</span></a></figcaption></figure><p>Your full list of opportunities will look very different from mine, and it will be much longer than the list above. Your list will be unique and constantly changing, because it depends on your skills, resources, connections, and the state of the world around you. To be luckier, you must be able to frequently and accurately update your list and pick which ones to take the leap on (please don&apos;t pick <a href="https://en.wikipedia.org/wiki/Sam_Bankman-Fried?ref=georgeyw.com">major financial crimes</a>).</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://lh5.googleusercontent.com/TLUI4YVOSfg41Bqw1jK6XDQw9RYngnvqCFxPNY9Dg2ZnfntiK2_Mp5ACKLxGa2YHHsW_JZ4OlOvcem-R0LgqsISAdpmt8ujsSIaEXO_d9bHCTniz0kfwMI0TLR9fLu2QCckt6RXC-tNitfI7qNWIB9M" class="kg-image" alt loading="lazy" width="624" height="312"><figcaption><span style="white-space: pre-wrap;">QR codes generated by ControlNet</span></figcaption></figure><p>A recent example that comes to mind are these beautiful, functional(!) <a href="https://mp-weixin-qq-com.translate.goog/s/i4WR5ULH1ZZYl8Watf3EPw?_x_tr_sl=auto&amp;_x_tr_tl=en&amp;_x_tr_hl=en-US&amp;_x_tr_pto=wapp&amp;ref=georgeyw.com">QR codes generated by ControlNet</a>. I&#x2019;m familiar with Stable Diffusion and other generative AI models, but my brain contains basically zero knowledge about QR codes (it turns out QR codes are <a href="https://news.ycombinator.com/item?id=36128082&amp;ref=georgeyw.com">quite malleable</a>). I wouldn&#x2019;t have considered this to be on my personal list of opportunities, but for this person, it was and they took it. This wasn&#x2019;t just a once-in-your-lifetime opportunity, this was once-in-anyone&#x2019;s-lifetime. Now they get to be &#x201C;that person that made super cool QR codes&#x201D; forever, and no one else does. Depending on what you&#x2019;re optimizing for (visibility, career opportunities, internet fame), this could be a fantastic outlier outcome for a speculative project.</p><p>Going for those one-off chances doesn&#x2019;t lend itself to a clean framework of resampling on the same decision, but it does expose you to many more potential outliers over a lifetime. You can get lucky in a particular domain by understanding how to search for outliers in that domain and by repeatedly sampling from that domain. This is good, but this still limits your shots on goal. You also might not find a life-changing outlier until late in life. </p><p>You can get even luckier, and sooner in life, by being able to accurately identify <em>all</em> the domains of long-tailed decisions that are accessible to you and learning to search for these outliers <em>everywhere</em>. Even if some of them are once-in-a-lifetime opportunities, once-in-a-lifetime opportunities abound.</p><h2 id="%E2%80%9Cbut-x-is-already-pretty-good%E2%80%9D">&#x201C;But X is already pretty good&#x201D;</h2><p>At the job that I left, I previously held the belief that it was an exceptionally good place to work. My basis for believing this was that many of my coworkers frequently commented on it being the best place they&#x2019;ve ever worked at. Surely if so many people think so, there must be some truth to it being exceptional?</p><p>I still believe it was a great place to work, but (as Ben Kuhn also discusses in his post) there&#x2019;s a difference between being at the 90th percentile (great) and being at the 99.9th percentile (exceptional). If you&#x2019;re at a 90th percentile company and most of your colleagues have worked at around 5 places in total, there&#x2019;s a 65.6% chance that your 90th percentile company is better than all of the previous 4 places they&#x2019;ve been. So being at a 90th percentile company might look like two thirds of your coworkers truthfully being able to say it&#x2019;s the best job they&#x2019;ve ever had, even though 1 in 10 places are better.</p><p>Truly strong evidence of an exceptional outcome looks stronger than two thirds of people with something comparable saying it&#x2019;s the best outcome they&#x2019;ve ever had. Evidence of an exceptional outcome looks more like zero people around you with something comparable. Without this kind of evidence, chances are that a given situation in your life is closer to a median outcome than you think it is. If so, the risk for rolling the dice again is also lower than you think it is. In general, we bias away from change, but we should try to embrace it to become lucky. In 2020, a <a href="https://academic.oup.com/restud/article-abstract/88/1/378/5834495?redirectedFrom=fulltext&amp;login=false&amp;ref=georgeyw.com">study</a> asked 20,000 people to make some difficult decision on their mind with a coin flip:</p><blockquote>For important decisions (e.g. quitting a job or ending a relationship), individuals who are told by the coin toss to make a change are more likely to make a change, more satisfied with their decisions, and happier six months later than those whose coin toss instructed maintaining the status quo.</blockquote><p>(n.b. If you&apos;re a former coworker, don&apos;t take this as advice or judgement. I haven&apos;t even worked at 5 places yet!)</p><h2 id="what-you-can-do-to-get-luckier">What you can do to get luckier</h2><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://www.georgeyw.com/content/images/2023/06/image-2.png" class="kg-image" alt loading="lazy" width="2000" height="988" srcset="https://www.georgeyw.com/content/images/size/w600/2023/06/image-2.png 600w, https://www.georgeyw.com/content/images/size/w1000/2023/06/image-2.png 1000w, https://www.georgeyw.com/content/images/size/w1600/2023/06/image-2.png 1600w, https://www.georgeyw.com/content/images/size/w2400/2023/06/image-2.png 2400w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">The basic feedback loop</span></figcaption></figure><p>Network more. Every new person that you develop a connection with, either socially or professionally, can be a long-tailed outcome. That might sound a little sociopathic, but the quality of our relationships predict a lot of our life happiness, so we should be more intentional in creating chances for outliers. Most people will be acquaintances, but some can be lifelong friends, romantic partners, and/or collaborators. Furthermore, a rich and healthy network might also spot long-tailed opportunities that are uniquely suited for you and bring them to you. Noticing good bets is easier when you have more bets to choose from, and having a strong network is much faster than searching alone (pay this forward!).</p><p>Be curious, too. It costs less energy to spend time learning about things that excite you. Doing <a href="http://www.paulgraham.com/greatwork.html?ref=georgeyw.com">great work</a> often looks like following your curiosity to places where people haven&apos;t tread before. This can mean going deeper than anyone else has, but it can also mean going broad enough that you find a unique combination of knowledge, as with the earlier QR code art.</p><p>In the long term, it&#x2019;s crucial to develop a strong probabilistic mindset and intuition, which makes it easier to reason about whether a decision is potentially long-tailed or not. This also helps avoid certain types of opportunities, such as the lottery, which is arguably long-tailed due to massive outliers, but not actually worth it. Finally, this helps identify downside risk, so that you can avoid making a decision that leads to a complete blow up. As you&#x2019;re exposed to more chances, you&#x2019;ll develop a better sense of what makes an opportunity good or bad in this framework, including making more accurate assessments for more unique long-tailed decisions that you only get to make once or twice.</p><p>There isn&#x2019;t a lot of general advice that I can offer beyond this. The ideal situation for developing a skill is one where you can do deliberate practice, but this framework <a href="https://commoncog.com/the-problems-with-deliberate-practice/?ref=georgeyw.com">can&#x2019;t be applied to the domain of luck</a>. There are no well-established training methods, and as far as I can tell, no one officially coaches for being lucky. A lot of my personal skill in this domain falls under gut feeling, or <a href="https://commoncog.com/the-tacit-knowledge-series/?ref=georgeyw.com">tacit knowledge</a>, and I haven&#x2019;t thought explicitly about this enough to make that gut feeling explicit. If you&#x2019;ve thought about this and have ideas, please share them. In lieu of a perfectly written guide, here are some things you might start with:</p><ul><li>Interview occasionally at jobs that sound interesting, even when you aren&#x2019;t looking for a new job</li><li>If you&#x2019;re introverted, get adopted by an extroverted friend</li><li>Say yes to invitations as a default</li><li>Do more things where the worst that can happen is nothing</li><li>Do fewer things that look like sitting at home and watching Netflix (I am guilty of sitting at home and playing video games)</li><li>Go to random events or activities on <a href="https://www.meetup.com/home/?ref=georgeyw.com">Meetup</a> or <a href="https://www.eventbrite.com/?ref=georgeyw.com">Eventbrite</a></li><li>Start saving up some resources to be ready for opportunities when they come; remember that your resources are resources, and you should spend them (wisely)</li><li>Don&#x2019;t do pyramid schemes</li></ul><p>Mileage is the best teacher available, so find some risk you can tolerate, and start taking some shots.</p><hr><p>This is my first post. It was inspired by this <a href="https://notebook.drmaciver.com/posts/2022-05-25-09:47.html?ref=georgeyw.com">unfinished draft</a> by David R. MacIver and by this post on <a href="https://www.benkuhn.net/writing/?ref=georgeyw.com">starting a blog</a> by Ben Kuhn. I recently re-read it for the third or fourth time and finally got off my ass to try my luck with it.</p><p>Thank you to those who volunteered to read drafts: AJ, Cyrus, Felix, Ross.</p><p>Any feedback is appreciated!</p>]]></content:encoded></item></channel></rss>