少点错误 2024年09月16日
Why I funded PIBBSS
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

作者认为 PIBBSS 主要支持“蓝天”或“基础”研究,这些研究可能存在低回报率,但在“最坏情况”下的对齐场景中至关重要。例如,当“对齐 MVP”不起作用、“急转弯”和“智能爆炸”比预期更可能发生,或者我们拥有比预期更多的时间来应对 AGI 时。作者认为 PIBBSS 的策略在当前的生态系统和市场中是有效的,并支持 PIBBSS。

😄 PIBBSS 的研究策略:PIBBSS 主要支持“蓝天”或“基础”研究,这些研究可能存在低回报率,但在“最坏情况”下的对齐场景中至关重要。例如,当“对齐 MVP”不起作用、“急转弯”和“智能爆炸”比预期更可能发生,或者我们拥有比预期更多的时间来应对 AGI 时。作者认为,与 MATS 支持的技术研究相比,PIBBSS 的策略在当前的生态系统和市场中是有效的,并支持 PIBBSS。

😊 PIBBSS 的研究价值:PIBBSS 支持了一系列蓝天研究,这些研究在现有机构中被忽视,但在各种可能的“最坏情况”下的 AGI 场景中可能具有很高的影响力。PIBBSS 似乎在招募非计算机科学领域的经验丰富的学者方面拥有良好的记录,并帮助他们重新利用其高级科学技能,为 AI 安全开发新方法。

🤔 PIBBSS 的潜在不足:作者认为 PIBBSS 可能正在从高方差的蓝天研究转向更主流的 AI 可解释性。虽然这可能会创造更多筹资机会,但作者认为这将是一个错误。AI 安全生态系统需要一个“奇思妙想”的家园,而 PIBBSS 似乎是这个领域最信誉良好、最称职、最符合 EA 的地方。作者鼓励 PIBBSS“拥抱怪异”,同时保持高学术标准的基本研究,以最优秀的基础科学机构为榜样。

🤨 PIBBSS 的招募流程:作者尚未检查 PIBBSS 的申请人筛选流程,并且不太确定它是最佳版本,考虑到 MATS 在导师和申请人筛选方面遇到的困难以及作者对选择蓝天研究组合的难度的直觉。作者强烈鼓励 PIBBSS 公开发布并征求其申请人筛选和研究优先级流程的反馈,以便 AI 安全生态系统能够提供有益的见解。

🤩 作者对 PIBBSS 的支持:作者认为 PIBBSS 是一个非常称职且符合 EA 的组织,并希望看到它发展壮大。作者认为 PIBBSS 将从多元化资金来源中受益,因为主流 AI 安全资助者已经更多地转向应用技术研究(或更与治理相关的基础研究,例如评估)。作者认为 Manifund 的资助者非常适合认可更多投机性基础研究,但作者自己并不真正知道如何评估此类研究,因此宁愿交给专家。PIBBSS 似乎很适合提供这种专业知识。

Published on September 15, 2024 7:56 PM GMT

I just left a comment on PIBBSS' Manfund grant request (which I funded $25k) that people might find interesting. PIBBSS needs more funding!

Main points in favor of this grant

    My inside view is that PIBBSS mainly supports “blue sky” or “basic” research, some of which has a low chance of paying off, but might be critical in “worst case” alignment scenarios (e.g., where “alignment MVPs” don’t work, “sharp left turns” and “intelligence explosions” are more likely than I expect, or where we have more time before AGI than I expect). In contrast, of the technical research MATS supports, about half is basic research (e.g., interpretability, evals, agent foundations) and half is applied research (e.g., oversight + control, value alignment). I think the MATS portfolio is a better holistic strategy for furthering AI safety and reducing AI catastrophic risk. However, if one takes into account the research conducted at AI labs and supported by MATS, PIBBSS’ strategy makes a lot of sense: they are supporting a wide portfolio of blue sky research that is particularly neglected by existing institutions and might be very impactful in a range of possible “worst-case” AGI scenarios. I think this is a valid strategy in the current ecosystem/market and I support PIBBSS!In MATS’ recent post, “Talent Needs of Technical AI Safety Teams”, we detail an AI safety talent archetype we name “Connector”. Connectors bridge exploratory theory and empirical science, and sometimes instantiate new research paradigms. As we discussed in the post, finding and developing Connectors is hard, often their development time is on the order of years, and there is little demand on the AI safety job market for this role. However, Connectors can have an outsized impact on shaping the AI safety field and the few that make it are “household names” in AI safety and usually build organizations, teams, or grant infrastructure around them. I think that MATS is far from the ideal training ground for Connectors (although some do pass through!) as our program is only 10 weeks long (with an optional 4 month extension) rather than the ideal 12-24 months, we select scholars to fit established mentors’ preferences rather than on the basis of their original research ideas, and our curriculum and milestones generally focus on building object-level scientific/engineering skills rather than research ideation and “identifying gaps”. It’s thus no surprise that most MATS scholars are “Iterator” archetypes. I think there is substantial value in a program like PIBBSS existing, to support the long-term development of “Connectors” and pursue impact in a higher-variance way than MATS.PIBBSS seems to have decent track record for recruiting experienced academics in non-CS fields and helping them repurpose their advanced scientific skills to develop novel approaches to AI safety. Highlights for me include Adam Shai’s “computational mechanics” approach to interpretability and model cognition, Martín Soto’s “logical updatelessness” approach to decision theory, and Gabriel Weil’s “tort law” approach to making AI labs liable for their potential harms on the long-term future.I don’t know Lucas Teixeira (Research Director) very well, but I know and respect Dušan D. Nešić (Operations Director) a lot. I also highly endorsed Nora Ammann’s vision (albeit while endorsing a different vision for MATS). I see PIBBSS as a highly competent and EA-aligned organization, and I would be excited to see them grow!I think PIBBSS would benefit from funding from diverse sources, as mainstream AI safety funders have pivoted more towards applied technical research (or more governance-relevant basic research like evals). I think Manifund regrantors are well-positioned to endorse more speculative basic research, but I don’t really know how to evalutate such research myself, so I’d rather defer to experts. PIBBSS seems well-positioned to provide this expertise! I know that Nora had quite deep models of this while Research Director and in talking with Dusan, I have had a similar impression. I hope to talk with Lucas soon!

Donor's main reservations

    It seems that PIBBSS might be pivoting away from higher variance blue sky research to focus on more mainstream AI interpretability. While this might create more opportunities for funding, I think this would be a mistake. The AI safety ecosystem needs a home for “weird ideas” and PIBBSS seems the most reputable, competent, EA-aligned place for this! I encourage PIBBSS to “embrace the weird”, albeit while maintaining high academic standards for basic research, modelled off the best basic science institutions.I haven’t examined PIBBSS’ applicant selection process and I’m not entirely confident it is the best version it can be, given how hard MATS has found mentor and applicant selection and my intuitions around the difficulty of choosing a blue sky research portfolio. I strongly encourage PIBBSS to publicly post and seek feedback on their applicant selection and research prioritization processes, so that the AI safety ecosystem can offer useful insight. I would also be open to discussing these more with PIBBSS, though I expect this would be less useful.My donation is not very counterfactual here, given PIBBSS’ large budget and track record. However, there has been a trend in typical large AI safety funders away from agent foundations and interpretability, so I think my grant is still meaningful.

Process for deciding amount

I decided to donate the project’s minimum funding ($25k) so that other donors would have time to consider the project’s merits and potentially contribute. Given the large budget and track record of PIBBSS, I think my funds are less counterfactual here than for smaller, more speculative projects, so I only donated the minimum. I might donate significantly more to PIBBSS later if I can’t find better grants, or if PIBBSS is unsuccessful in fundraising.

Conflicts of interest

I don't believe there are any conflicts of interest to declare.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

PIBBSS AI安全 蓝天研究 基础研究 对齐
相关文章