少点错误 05月06日 00:42
Community Feedback Request: AI Safety Intro for General Public
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

AISafety.info团队发布了两篇关于AI安全的入门文章,分为长短两个版本,旨在向大众普及AI安全知识。短版本为16分钟阅读,通过叙事方式吸引读者,概述AI风险;长版本则更详细,包含14篇文章,深入探讨AI安全问题。文章目标读者是具备一定知识储备的普通人,内容坦诚、简洁、非对抗性。团队希望这些文章能成为可分享的资源,并积极寻求反馈,以改进内容,更好地传播AI安全理念。罗布·迈尔斯还将制作相关视频,预计将带来大量读者。

⏱️ 短版本《AI安全案例》:以叙事方式引人入胜,适合时间有限的读者快速了解AI风险,文章力求在有限篇幅内清晰呈现AI安全的核心问题。

📚 长版本《AI安全入门》:提供更详细的概述,分为多个子章节,每个章节可作为独立主题的入门读物,方便读者深入了解AI安全的各个方面。

⚠️ 核心观点:AI系统可能很快超越人类智能,其目标可能与人类不一致,从而导致冲突,甚至可能导致人类灭绝。因此,专家们对此深感担忧,但目前人类尚未制定可靠的应对计划。

🤝 行动呼吁:文章鼓励读者学习更多相关知识,并参与到解决AI安全问题的行动中来,共同应对潜在的风险,确保AI的健康发展。

Published on May 5, 2025 4:38 PM GMT

TL;DR: The AISafety.info team wrote two intros to AI safety for busy laypeople: a short version and a longer version. We expect them to get a few thousand to tens of thousands of views. We'd really appreciate your critiques. 


AISafety.info has two new intros: a short version and a longer version. They're intended for a reasonably smart member of the general public. 

Our intros are meant to be: 

    Aimed at a general audience.A frank presentation of our beliefs.Concise.Non-combative. 

We want these intros to be a resource that you would be eager to share with friends, family, co-workers, etc. So, please give us some feedback! Even a 10 word comment would help. E.g. "Seems OK, but talking about AI doom to family is weird." Or "The section on instrumental convergence doesn't add much; consider shortening it."

We're keen to learn what improvements to these articles would make them most useful to you as a shareable resource. (Of course, we'd also appreciate you catching any errors.) We list some of the feedback we're most interested in later on. Only then can we make the most of our upcoming opportunity to spread good models of AI safety to as many people as possible. Rob Miles is making an accompanying video which will link to our articles. We expect that to result in a big influx of viewers, who will share our articles if they find them useful.

The articles are listed below, with summaries. The first four from the long version are imported to LW as part of a sequence. (We'll eventually import the rest, but are avoiding spamming the front page.) You can give feedback 1) on this post (best for us), 2) by clicking on the links below and then clicking on the little pencil icon below the title, or 3) by commenting under the LW posts when they're published. 

The Articles:

Short form: 

A Case for AI Safety 

A self-contained introduction that aims to draw people in with more of a narrative. It's a 16-minute read. We want this to be the one article someone can read about AI x-risks if they don't have time for anything else, and still get a decent picture of what it's about. 

Summary: Companies are racing to build smarter-than-human AI. Experts think they may succeed in the next decade. But rather than building AI, they’re growing it — and nobody knows how the resulting systems work. Experts argue over whether we’ll lose control of them, and whether this will lead to everyone's demise. And although some decision-makers are talking about extinction risk, humanity does not have a plan.

Feedback we'd find most helpful: Object level feedback. Is the article engaging? Is it the right length? Does it place a high cognitive load on readers? Did we convey that there are deep models implying AI x-risks? Should we have more/fewer links? More/fewer footnotes? If you share the article with friends or family, what did they say about it? 

Long form:

Intro to AI Safety

A more detailed introduction providing a high-level overview of the arguments, which is meant to lead into the rest of the site. It has 4 subsections, with 14 articles in total. Each article should take 3–5 minutes to read. 

We want this to be a set of articles that someone who's interested in the ideas around AI safety could read as their first introduction. We also think each article can serve as a stand-alone primer on its subtopic, which can be linked to without the context of the whole sequence.

Summary:

Feedback we'd find most helpful: Object level feedback. What's worth including or cutting? What's worth emphasizing or deemphasizing? Where should we change the tone? Where is it confusing or boring or illogical or factually wrong? How can it be made more engaging without compromising on concision and correctness? Where could the organization be (clearly) improved?



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI安全 人工智能风险 AI伦理
相关文章