Interconnects 03月20日
The latest open artifacts (#8): The return of ~30B models, side effects of OpenAI's proposed DeepSeek ban, and yet another reasoning roundup
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了OpenAI与开源AI的关系,提到OpenAI对DeepSeek的评价及可能产生的影响,还介绍了多种AI模型、数据集及相关研究成果。

OpenAI对DeepSeek的评价引发担忧,可能影响中美AI交流

多种AI模型发布,如reka-flash-3、OLMo-2-0325-32B-Instruct等

多个机构在AI领域的成果,如DNA预测模型、数据标注公司讨论等

多种AI推理模型及数据集,如R1-Omni-0.5B、natural_reasoning等

OpenAI has had a tenuous relationship with open-source AI since ChatGPT. First was the existential risk policies, now it is real competitive threats from abroad. OpenAI released their comment on the new administration’s AI Action Plan, which has a lot in it, and this portion immediately stuck out to the broader AI community:

OpenAI claims, without existing evidence, that DeepSeek is “state-subsidized, state-controlled, and freely available,” so “the cost its users is their privacy and security.” Without flat out saying it, they’re insinuating that Chinese models should be banned within the U.S.

Yes, it is very reasonable to worry about information security for API models going to China, but this could very easily have unintended consequences of cutting off all Chinese AI models, open and closed, from the American system. This would be an extreme set back to the open-source AI movement, as open solutions win with a plurality of options.

The team at Ai2 had a much simpler message around open models in our comment to the OSTP for the White House AI Action Plan. The United States needs to invest in infrastructure and organizations who are building open tooling. By investing and enabling people building in and around the U.S. (i.e., our allies), the open model default will be effectively American. When it is the standard, benefits from open models from China get absorbed and increase progress, while being able to check for the reality of the risks that OpenAI posed above.

As you can see in this post, times are good for open models. It doesn’t take a lot to help it grow, but you need to be careful of some recommendations (or avoid some bills that have been drafted recently1).

Share

The structure of these is:

Post likely too long for email, we recommend reading on the web version.

Our Picks

What is notable here is about is how all of these models are in the 24B-32B range! This is a super popular size with developers as it is a nice combination of very capable, yet not too unwieldily large. Other than Qwen, there had been very few notable models in this range (with Llama 3 skipping it for risk concerns).

Links

Some outputs showcasing the image generation capabilities of Gemini 2.0 Flash. Sources: https://x.com/hamishivi/status/1901827063709143317, https://x.com/angrypenguinPNG/status/1902158240382935488, https://x.com/goodside/status/1900349595718148455

Reasoning

Models

Source: https://huggingface.co/open-r1/OlympicCoder-7B

Datasets

Models

Read more

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

OpenAI AI模型 AI数据集 开源AI
相关文章