少点错误 2024年12月20日
?? Announcing CeSIA: The French Center for AI Safety
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

法国人工智能安全中心(CeSIA)致力于在法国和欧洲推广人工智能安全文化。其核心使命是通过教育和信息传播,降低AI风险。CeSIA的工作涵盖教育、技术研发和政策倡导三大领域。教育方面,他们创建了ML4Good训练营,并在法国顶尖大学开设AI安全课程。技术研发上,CeSIA专注于为AI安全系统建立基准,评估监控系统的稳健性。在政策倡导方面,他们积极参与欧盟通用AI行为准则的制定,并为法国政府提供AI评估实验室的建议。CeSIA通过多领域的高影响力行动,在技术、政策和教育之间形成协同效应,旨在为决策者提供具体可行的技术指导。

🏫 CeSIA通过ML4Good训练营和在ENS Ulm等顶尖大学开设的认证课程,致力于培养下一代AI研究人员,提升公众对AI安全的认知。

🔬 CeSIA的技术研发侧重于创建AI安全防护系统的评估基准,并发布了BELLS基准测试工具,用于评估LLM安全措施的稳健性,挑战了现有安全措施的有效性,并强调了监控系统的重要性。

🏛️ CeSIA积极参与AI政策的制定,包括参与欧盟通用AI行为准则的起草,并为法国国家AI评估实验室的建立提供建议,旨在通过政策倡导来促进AI安全的落地。

Published on December 20, 2024 2:17 PM GMT

Executive Summary

The Centre pour la Sécurité de l'IA[1] (CeSIA, pronounced "seez-ya") or French Center for AI Safety is a new Paris-based organization dedicated to fostering a culture of AI safety at the French and European level. Our mission is to reduce AI risks through education and information about potential risks and their solutions.

Our activities span three main areas:

Our team consists of 8 employees, 3 freelancers, and numerous volunteers.

 

Figure: An overview of how our current projects fit within a broader theory of change (in blue: CeSIA and its projects, in green: different actors and events, and what they do if we intervene on them, in red our three main focus).

By pursuing high-impact opportunities across multiple domains, we create synergistic effects between our activities. Our technical work informs our policy recommendations with practical expertise. Through education, we both identify and upskill talented individuals. Operating at the intersection of technical research and governance builds our credibility while advancing our goal of informing policy. This integrated approach allows us to provide policymakers with concrete, implementable technical guidance.

 

Figure: From field-building & education, our historical activities, to technical research and finally policy advocacy.

Mission

The future of general-purpose AI technology is uncertain, with a wide range of trajectories appearing possible even in the near future, including both very positive and very negative outcomes. But nothing about the future of AI is inevitable. It will be the decisions of societies and governments that will determine the future of AI.

– introduction of the International Scientific Report on the Safety of Advanced AI: Interim Report.

 

Our mission: Fostering a culture of AI safety by educating and informing about AI risks and solutions.

AI safety is a socio-technical problem that requires a socio-technical solution. According to the International Scientific Report on the Safety of Advanced AI, one of the most critical factors shaping the future will be society’s response to AI development. Unfortunately, many developers, policymakers, and most of the population remain completely oblivious to AI's potentially catastrophic risks.

We believe that AI safety is in need of a social approach to be correctly tackled. Technical AI safety alone is not enough, and it seems likely that solutions found through this route will be in any case bottlenecked by policy-makers' and the public's understanding of the issue.

We want to focus on such bottlenecks and maximize our counterfactual impact, targeting actions that would likely not be taken otherwise, and having clear theories of change for our project. We will be writing further articles detailing our methodology for thinking through such counterfactuals and evaluating our projects quantitatively, as we are using mixes of different approaches to ensure we are not getting trapped in the failure mode of any single one, and because the future is uncertain.        

We think that at its root, AI safety is a coordination problem, and thus, awareness and engagement on multiple fronts are going to be crucial catalysts for the adoption of best practices and effective regulation.

European and French context

As we are located in France, our immediate focus is on what we can achieve within France and Europe. Both regions are becoming increasingly crucial for AI safety, particularly as France prepares to host the next International AI Action Summit in February 2025, which follows the UK’s Bletchley summit in November 2023 and South Korea’s followup summit in May 2024.

France has positioned itself as a significant AI hub in Europe, hosting major research centers and offices of leading AI companies including Meta, Google DeepMind, Hugging Face, Mistral AI, and recently OpenAI. This concentration of AI development makes French AI policy particularly influential, as seen during the EU AI Act negotiations, where France expressed significant reservations about the legislation.

The country's strong advocacy for open-source AI development raises concerns about distributed misuse of future AIs, reduced control mechanisms, and potential risks from autonomous replication. As one of the organizations involved in drafting the EU AI Act's Code of Practice, we're working to ensure robust safety considerations for frontier AI systems are integrated into these guidelines. You can read more on France’s context here.

After the summit, it’s possible we expand our focus internationally.

Main activities

Our current activities revolve around four pillars:

Policy advocacy & Governance

Rationale: Engaging directly with key institutions and policymakers allows them to better understand the risks. This helps raise awareness, and inform decision-making.

 

R&D

Rationale: We think R&D can contribute to safety culture by creating new narratives that we can then present while discussing AI risks.

Education & field-building

Rationale: Educating the next generation of AI researchers is important. By offering university courses, bootcamps, and teaching materials, we can attract and upskill top talents.

Education & mass outreach

Rationale: By educating the general population and, in particular, the scientific communities, we aim to promote a robust culture of AI safety within public institutions, academia, and the industry. Public awareness is the foundation of societal engagement, which in turn influences institutional priorities and decision-making processes. Policymakers are not isolated from public opinion; they are often swayed by societal narratives and the concerns of their families, colleagues, and communities.

Gallery

Figure: List of organisations that participated during the Turing Dialogues, a series of round tables to discuss future challenges in AI, with key actors of the French and international scene. 

 

Figure: ML4Good 9 - Brazil, July 2024, in partnership with Condor Initiative.

 

Figure: Our field-building pipeline and different partners.

 

Figure: From the BELLS benchmark. The plot shows that every safeguard fails to detect jailbreaks effectively on at least one dataset, with detection rates dropping below 34.2%. This highlights the need for more robust or specialized safeguards to address diverse vulnerabilities.

 

Figure: The Turing Seminar, our course in ENS Paris-Saclay, arguably the most important AI Master in the country.

Team

Our starting members include:

Board, by alphabetical order:

We also collaborate with Alexandre Variengien (co-founder of CeSIA), Siméon Campos (SaferAI, co-founder of EffiSciences), Diego Dorn (the main contributor to our BELLS project, head teacher at some ML4Good), Nia Gardner (Executive Director of the ML4Good new organization), and also Jérémy Andréoletti, Amaury Lorin, Jeanne Salle, Léo Dana, Lucie Philippon, Nicolas Guillard, Michelle Nie, Simon Coumes, and Sofiane Mazière.

Funding

We have received several generous grants from the following funders (and multiple smaller ones not included here), who we cannot thank enough:

These grants fell into two primary categories: unrestricted funds for CeSIA’s general operations, and targeted support for specific project initiatives.

Next steps

We created the organization in March 2024. Most of the previous points were done during the last 6 months with the exception of our historic field-building activities, such as the ML4Good bootcamps.

It’s a lot of different types of activities! And It’s possible we will try to narrow down our focus in the future if this makes sense.

The next six months will be pretty intense, between the AI Action Summit, the drafting of the AI Act's Code of Practice, and completing ongoing projects such as the textbook.

We invite researchers, educators, and policymakers to collaborate with us as we navigate this important phase. Feel free to reach out at u>contact@securite-ia.fr</u.

Here is our Substack if you want to follow us or meet us in person in January-February for the Turing Dialogues on AI Governance, our series of roundtables, which is an official side event of the AI Action Summit.

Thanks to the numerous people who have advised, contributed, volunteered, and have beneficially influenced us.

 

  1. ^

    “Safety” usually translates to “Sécurité” in French, and “Security” translates to “Sûreté”.

  2. ^

     and in the European Union?

  3. ^

     OECD: Organisation for Economic Co-operation and Development.

  4. ^

     CNIL: National Commission on Informatics and Liberty

  5. ^

     ANSSI: National Cybersecurity Agency of France

  6. ^

     LNE: National Laboratory of Metrology and Testing

  7. ^

    For example, during ML4Good Brazil, we created and reactivated 4 different AI Safety field-building groups in Latin America: Argentina, Columbia, AI safety Brazil and ImpactRio, and ImpactUSP (University of San Paolo). Participants consistently rate us above 9/10 for recommending this bootcamp.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI安全 人工智能风险 政策倡导 技术研发 教育
相关文章