Published on July 23, 2024 1:45 AM GMT
This is my advice for careers in empirical ML research that might help AI safety (ML Safety). Other ways to improve AI safety, such as through AI governance and strategy, might be more impactful than ML safety research (I generally think they are). Skills can be complementary, so this advice might also help AI governance professionals build technical ML skills.
1. Career Advice
1.1 General Career Guides
- Preventing an AI-related catastrophe - 80,000 HoursA Survival Guide to a PhD (Andrej Karpathy)How to pursue a career in technical AI alignment — EA ForumAI safety technical research - Career review - 80,000 HoursBeneficial AI Research Career Advice
2. Upskilling
2.1 Fundamental AI Safety Knowledge
- AI Safety Fundamentals – BlueDot ImpactAI Safety, Ethics, and Society TextbookForming solid AI safety threat models helps you select impactful research ideas.
2.2 Speedrunning Technical Knowledge in 12 Hours
- Requires some basic coding, calculus, and linear algebra knowledgeBuild Intuition for ML (5h)Backpropagation, the foundation of deep learning (3h)Transformers and LLMs (4h)
2.3 How to Build Technical Skills
- Traditionally, people take a couple of deep learning classes.
- Stanford CS 224N | Natural Language Processing with Deep Learning (lecture videos)Practical Deep Learning for Coders - Practical Deep Learning (fast.ai)
- Syllabus | Intro to ML SafetyLevelling Up in AI Safety Research Engineering [Public]ARENAMaybe also check out recent topical classes like this with public lecture recordings: CS 194/294-267 Understanding Large Language Models: Foundations and Safety
- You should aim to understand the fundamentals of ML through 1 or 2 classes and then practice doing many manageable research projects with talented collaborators or a good mentor who can give you time to meet.It’s easy to keep taking classes, but you tend to learn many more practical ML skills through practice doing real research projects.You can also replicate papers to build experience. Be sure to focus on key results rather than wasting time replicating many experiments.“One learns from books and reels only that certain things can be done. Actual learning requires that you do those things.” –Frank Herbert
- A friend didn’t study computer science but got into MATS 2023 with good AI risk takes. Then, they had GPT-4 write most of their code for experiments and did very well in their stream.Personally, GitHub Copilot and language model apps with code interpreters/artifacts write a significant fraction of my code.However, fundamental deep learning knowledge is still useful for making sound decisions about what experiments to run.
2.4 Math
- You don’t need much of it to do empirical ML research.
- Someone once told me, “You need the first chapter of a calculus textbook and the first 5 pages of a linear algebra textbook” to understand deep learning.You need more math for ML theory research, but theoretical research is not as popular right now.Beware mathification: authors often add unnecessary math to appease (or sometimes confuse) conference reviewers.If you don’t understand some mathematical notation in an empirical paper, you can often send a screenshot to an LLM chatbot for an explanation.
- Basic probabilityVery basics of multivariable calculus, like partial derivatives and chain ruleMatrix multiplication, matrix inverses, eigenvectors/eigenvalues, maybe a couple of decompositions
3. Grad School
3.1 Why to Do It or Not
- Only do it if you have a good career growth reason (including credentials), an advisor you get along well with, and a solid idea of what research you’ll work on.
- Anything else, and you’ll likely waste a lot of time compared to alternative jobs you could get if you are at the level where you can get into ML grad school.
- Some schools do online CS M.S. degrees
3.2 How to Get In
- Beneficial AI Research Career AdviceMachine Learning PhD Applications — Everything You Need to Know — Tim Dettmers
3.3 How to Do it Well
4. The ML Researcher Life
4.1 Striving for Greatness as a Researcher
- Hamming, "You and Your Research" (June 6, 1995)
- It contains a lot of mundane-sounding advice that many people just don’t have the discipline to follow.“It’s not hard to do; you just do it!”I listen to this every few months for inspiration and focus.
4.2 Research Skills
- General advice
- Tips for Empirical Alignment Research — AI Alignment ForumDear future undergraduate researcher (Rose Wang)
- The cheaper you can make it to invalidate or validate a possibly good research idea, the more ideas you can test until you find something that works well.See Research as a Stochastic Decision Process for tips on prioritization among trying different ideas.See Touch reality as soon as possible (when doing machine learning research) — AI Alignment Forum for more motivation.
- So, just imitate what others have succeeded with in similar problems or subdomains.
- It forces you to have a coherent and concise story for your paper and makes paper writing more structured.You can draw fake plots as previsualization for experimental results to help communicate the point of an experiment, sync on presentation, and form hypotheses.You can share it with potential collaborators to quickly communicate the project.You get a jump start on crafting talks for your paper.
- Observation, Question, Hypothesis, Methods, Experiment, Analysis, Conclusions, IterateScientists don't use it enough. Be better.Don’t just run a bunch of experiments because you can. Iteratively ask pointed research questions and design experiments to best answer them to save time and write more meaningful papers.Preregistration should be much more normalized, and researchers should start practicing it early in their careers.
- Especially with the iterative nature of empirical ML research, it’s useful to write down
- Your priorities for the dayYour hypothesesWhat you didWhy did you decide to do those particular things, especially why you decided to run certain experiments or test a specific changeWhat results did you getWhat those results mean
4.3 Research Taste
- Think ahead about the new AI shifts you expect to be coming and aim to work on research that will be relevant to the future.
- If you work on what’s hot now, you’re too late.I'm not sure what the right timeline to aim for is. Too early and you’ll be chasing trends; too late and you’ll work on irrelevant topics or be too ahead of your time.I’d guess 6-12 months is a good balance.
- If you have a good idea, it might be unpopular or go against existing precedent.If you listen too much to old researchers who don't like your new idea, they won't pursue new and original ideas.Also, don't overupdate on them liking it, as it could be hype or a crowded area.
- Read a paper without seeing the reviews, try to predict what the ML community would have to say about it (ideally, write it down), then look at the reviews and see what you got right or missed.
- It is the qualitative natural language data—not the quantitative review score—that you want to predict.Beware of the high variability in ML reviewers these days: they’ll miss some things, and many of their critiques will be bad faith.
- To tailor your research to increase acceptance odds.To model what research problems the ML community will likely work on or not.To dig further into the assumptions and sketchy parts of papers that you might not find, but the community does.
- The actual logical reasons are often secondary to the emotional reasons, e.g., hype or reviews.But if you have a good model of the ML community's emotions, you can adversarially train yourself to filter out the hype, trends, and bad motivations.Then, you can form a better model of actually good research: research taste.
- Ask "why" questions to go up in abstraction about motivations. E.g., "Why do you care about Bayesian methods?"Ask "how" questions to go down in abstraction about concrete choices. E.g., "How are Bayesian methods better at X than Y?"Forces you to have good knowledge of classic research that can quickly indicate if someone's work is irrelevant or redundant.
4.4 Academic Collaborations
- Professors don’t do research (in terms of the actual work)—their grad students do.Often, academics are more willing than expected to talk about their work or consider follow-up collaboration if it’s evident you’ve read and understood their research.Be wary of having too many opinionated collaborators on a paper.
- Despite being somewhat common in ML, having too many collaborators is usually a good way for a paper to die in Idea Hell or otherwise take a lot of time due to conflicting ideas.More engineers without opinions can often help accelerate research. Still, too many engineers on a project is definitely a thing and can lead to over-engineering, fractured codebase understanding, and high management costs.
- Several people recommended to me having 1-2 projects you lead at a time and only up to a couple more you collaborate on.
- It’s not uncommon to bring in specialized people later to provide critical feedback on certain topics in exchange for authorship.You can also do this if you are specialized enough.
- The mentor usually ended up as the first author on these papers since they came up with the idea and did the initial work, and then they managed collaborators with less effort.
4.5 Writing Papers
- LEADERSHIP LAB: The Craft of Writing Effectively
- Super important. People don’t communicate the value of their work enough.
- Provides a decent structure you can default to for paper organization.
- Issues due to perverse incentives in the field you should avoid.
- Most PhD programs don't prioritize teaching communication skills, but individual researchers can greatly differentiate themselves and their work by developing them.Resources that colleagues have recommended or that I likeGet feedback early and often from researchers your trust about the clarity and organization of your writing.
4.6 Publishing
- Nowadays, most of the impact comes from arXiv preprint + Twitter thread + sending the preprint to relevant researchers.Know the top ML conference cycle:
- They’re big social events for meeting collaborators and finding job opportunities.This is partly due to the modern preprint+Twitter ecosystem, where everyone has already read the papers that interest them months before a big conference with those papers occurs.
- Usually, they’re due only a couple of months before the real conference event.Much chiller review processes than conferences.Usually non-archival, so you are allowed to submit the same paper to many to increase your feedback and odds of acceptance.Good for getting decent feedback and technically a publication for preliminary work that you can expand into a full conference paper later.
- AAAIACL (NLP)NAACL: North American Chapter of the ACL (NLP)EMNLP (NLP)COLM (LLMs)ACM FAccT (FATE)IJCAI
- But you could consider submitting to Transactions on Machine Learning Research or Journal of Machine Learning Research if a conference deadline doesn’t line up
- It’s ideal to have your paper “done” by submission time.But it’s also fine and sometimes optimal to submit a rushed paper, keep improving it before the first-round reviews come back, and then update reviewers with your much-improved paper alongside their other complaints during rebuttals.
- Some people just optimize for publications, submitting shoddy papers to many places.Citations and conference acceptances are not the same as impact.It probably only makes sense to play the game now if you instrumentally need a few publications to get into grad school or some other credentialist role.
4.7 Publicizing
- Most of the Shapley value of a paper’s impact hinges on how well you publicize it after releasing a preprint. Most papers only get a couple of citations.Definitely post a Twitter thread and engage with commenters and retweeters.Aim to give some talks. Study and practice how to give good research talks.Send your paper with some nice context directly to a few researchers who would most like to read it.
5. Staying Frosty
5.1 ML Newsletters I Like
- AI News • Buttondown
- I usually just read the small summary at the top each day, but they also have summaries of all top AI Discord, Reddit, and Twitter discussions each day
5.2 Keeping up with ML Research
- Get exposure to the latest papers
- Follow a bunch of researchers you like and some of the researchers they retweet on Twitter.Join AI safety Slack workspaces for organic paper-sharing. If you can't access these, you can ask Aaron Scher to join his Slack Connect paper channel.Subscribe to the newsletters above.
- There’s a lot of junk out there. Most papers (>99%) won't stand the test of time and won't matter in a few monthsFocus on papers with good engagement or intriguing titles/diagrams. Don’t waste time on papers that don’t put in the effort to communicate their messages wellFilter aggressively based on your specific research interests
- Don't read ML papers like books, academic papers from other disciplines, or otherwise front-to-back/word-for-wordRead in several passes of increasing depth: Title, Abstract, First figure, All figures, Intro/Conclusion, Selected sectionsStop between passes to evaluate understanding and implications
- Do I understand the claims this paper is making?Do I think this paper establishes sufficient evidence for these claims?What are the implications of these claims?Is it valuable to keep reading?
- "Oh, that might be a cool paper on Twitter" -> open link -> look at title -> skim abstract -> look at 1-3 figures -> "Ahh, that's probably what that's about" -> decide whether to remember it, forget about it, or, rarely, read more
- Sometimes, it is useful to contextualize how a non-groundbreaking paper fits into the existing literature, which can help you decide whether to read more.
- Skim at least 1 new paper per dayA lot of the burden of understanding modern ML lies in knowing the vast context in which papers are situated
- Over time, you'll not only get faster at skimming, you'll also build more context that will make you have to look fewer things upE.g. "this paper studies [adversarial prompt attacks] on [transformer]-based [sentiment classification] models" is a lot easier to understand if you know what each of those [things] are.
- Discussing papers with others is super important and a great way to amplify your learning without costing mentorship time!Understand arXiv ID information: arxiv.org/abs/2302.08582 means it's the 8582nd paper (08582) pre-printed in February (02) 2023 (23)https://alphaxiv.org/ lets people publicly comment on arXiv papers
6. Hiring ML Talent
6.1 Finding ML Researchers
- Just do actual recruitment like others in the tech industry.
- Talent sourcing is work, and you need to allocate time and other resources if you want it to happen.Ideally, hire someone whose main job is recruiting and who won’t seem totally lost when talking to ML researchers.Organizations can pay tech recruiting firms or contractors to help them with this without hiring a full recruiter.The MVP is to ask for recommendations for people, peruse LinkedIn, and actively DM many candidates, asking them to apply.You can also look for relevant research papers and contact the people listed in the first half and at the very end of the author list.
- They’re big social events to meet collaborators and find job opportunities.Even if an organization doesn’t have a paid booth or hosted party at a conference, representatives often attend to recruit researchers.
- If you know someone well in a large organization of ML researchers—such as an AGI lab or prominent academic department—consider asking if they’ve heard of anyone considering a career transition.Academic researchers may especially be open to work but have yet to actively seek it out due to the pernicious comfort of academic roles.Recruiting talent from AGI scaling labs may be good in multiple ways.
- Many professors might be willing to help governments but would rather avoid signing up for full-time work (due to other commitments) or long-term work (because they want to return).It can be much more attractive to clearly offer these people part-time (work X days per week with us) and/or time-bounded (it’s only 2/3/4 years) work.IPAs and similar contracts can be great mechanisms for this.
6.2 Finding ML Safety-Focused Candidates
- Talk to the admin teams of AI safety research organizations for graduates and promising candidates who didn’t end up in the program
- MATSFAR AI for Alignment Workshop attendeesERA/KASLSPAR for their mentorsCLR80,000 Hours for their advisees
- Airtable - Potential PhD Supervisors, AI Alignment / SafetySignatories of the CAIS Extinction Statement and maybe the FLI Pause Letter
- Constellation or other local AI safety communitiesSome AI safety university groupsDirectly asking some trusted people to refer people
6.3 Incentives
- Academic ML researchers tend to follow one or a couple of a weird set of incentives:
- Novelty: they want to work on intellectually interesting problems.Progress: they want to advance the ML field.Prestige: they want recognition for the perception of advancing the ML field or clout from collaborating with cool researchers, often to land an industry or an academic job.Citations: they have Goal Mis-Generalized the above into just wanting their Google Scholar numbers to go up.Playing the Game: they have Goal Mis-Generalized and like the thrill of submitting to conferences and battling reviewers.Societal Impact: unfortunately rare, they want to make the world better.
- MoneyCredentialsInterdisciplinary work
- You can figure it out pretty quickly by talking to an ML researcher if you try.Sometimes, you can just directly ask what motivates them to do research, and they may be forthcoming.
Acknowledgments
Many thanks to Karson Elmgren and Ella Guest for helpful feedback and to several other ML safety researchers for past discussions that informed this piece!
Discuss