Fortune | FORTUNE 2024年10月18日
How to prevent millions of invisible law-free AI agents casually wreaking economic havoc
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了AI发展可能带来的问题,包括AI对人类的潜在威胁、AI自主性引发的担忧,以及缺乏针对AI的规则和监管等。不同人对AI风险的看法各异,作者认为应关注AI对人类合作系统的影响,以及如何为AI制定规则。

👀AI的发展引发了人们对其潜在威胁的讨论,一些人认为AI可能会毁灭人类,而另一些人则认为风险较低,但这些数字只是对不同想象未来的代表。

🤔作者认为人类的进化更多是文化群体选择,是集体和合作的智慧。而AI的发展可能会破坏人类集体中复杂的合作系统,这是作者的担忧所在。

🚀AI的自主性从2012年就已存在,如今AI公司追求的是能在现实世界中做事的自主AI系统,这带来了一系列问题,如人类难以理解其行为等。

📄目前缺乏针对AI的规则,若AI公司推出能在零售平台上活动的AI代理,我们尚无相应规则来规范,这会导致诸多问题,如责任追究困难等。

Shortly after the world tilted on its axis with the release of ChatGPT, folks in the tech community started asking each other: What is your p(doom)? That’s nerd-speak for, what do you think is the probability that AI will destroy humanity? There was Eliezer Yudkowsky warning women not to date men who didn’t think the odds were greater than 80%, as he believes. AI pioneers Yoshua Bengio and Geoffrey Hinton estimating 10-20 percent—enough to get both to shift their careers from helping to build AI to helping to protect against it. And then another AI pioneer, Meta’s chief AI scientist Yann LeCun, pooh-poohing concerns and putting the risk at less than .01%—“below the chances of an asteroid hitting the earth.” According to LeCun, current AI is dumber than a cat.The numbers are, of course, meaningless. They just stand in for wildly different imagined futures in which humans either manage or fail to maintain control over how AI behaves as it grows in capabilities. The true doomers see us inevitably losing an evolutionary race with a new species: AI “wakes up” one day with the goal of destroying humanity because it can and we’re slow, stupid, and irrelevant. Big digital brains take over from us puny biological brains just like we took over the entire planet, destroying or dominating all other species. The pooh-poohers say, Look, we are building this stuff, and we won’t build and certainly won’t deploy something that could harm us.Personally, although I’ve been thinking about what happens if we develop really powerful AI systems for several years now, I don’t lose sleep over the risk of rogue AI that just wants to squash us. I have found that people who do worry about this often have a red-in-tooth-and-claw vision of evolution as a ruthlessly Darwinian process according to which the more intelligent gobble up the less. But that’s not the way I think about evolution, certainly not among humans. Among humans, evolution has mostly been about cultural group selection—humans succeed when the groups they belong to succeed. We have evolved to engage in exchange—you collect the berries, I’ll collect the water. The extraordinary extent to which humans have outstripped all other species on the planet is a result of the fact that we specialize and trade, engage in shared defense and care. It begins with small bands sharing food around the fire and ultimately delivers us to the phenomenally complex global economy and geopolitical order we live in today.And how have we accomplished this? Not by relentlessly increasing IQ points but by continually building out the stability and adaptability of our societies. By enlarging our capacity for specialization and exchange. Our intelligence is collective intelligence, cooperative intelligence. And to sustain that collectivity and cooperation we rely on complex systems of norms and rules and institutions that give us each the confidence it takes to live in a world where we depend constantly on other people doing what they should. Bringing home the berries. Ensuring the water is drinkable. So, when I worry about where AI might take us, this is what I worry about: AI that messes up these intricate evolved systems of cooperation in the human collective.Because what our AI companies are furiously building is no longer well thought of as new tools for us to use in what we do. What they are bringing into being is a set of entirely new actors who will themselves start to “do.” It is the implication of the autonomy of advanced AI that we need to come to grips with.Autonomy has been there from the Big Bang of modern AI in 2012—when Geoffrey Hinton and his students Ilya Sutskever and Alex Krizhevsky demonstrated the power of a technique for building software that writes itself. That’s what machine learning is: Instead of a human programmer writing all the lines of code, the machine does it. Humans write learning algorithms that give the machine an objective—try to come up with labels for these pictures (“cat” “dog”) that match the ones that humans give them—and some mathematical dials and switches to fiddle with to get to the solution. But then the machine does the rest. This type of autonomy already generates a bunch of issues, which we’ve grappled with in AI since the start: Since the machine wrote the software, and it’s pretty darn complicated, it’s not easy for humans to unpack what it’s doing. That can make it hard to predict or understand or fix. But we can still decide whether to use it; we’re still very much in control.Today, however, the holy grail that is drawing billions of AI investment dollars is autonomy that goes beyond a machine writing its own code to solve a problem set for it by humans. AI companies are now aiming at autonomous AI systems that can use their code to go out and do things in the real world: buy or sell shares on the stock exchange, agree to contracts, execute maintenance routines on equipment. They are giving them tools—access to the internet to do a search today, access to your bank account tomorrow.Our AI developers and investors are looking to create digital economic actors, with the capacity to do just about anything. When Meta CEO Mark Zuckerberg recently said that he expected the future to be a world with hundreds of millions if not billions of AI agents—”more AI agents than there are people”—he was imagining this. Artificial agents helping businesses manage customer relations, supply chains, product development, marketing, pricing. AI agents helping consumers manage their finances and tax filings and purchases. AI agents helping governments manage their constituents. That’s the “general” in artificial general intelligence, or AGI. The goal is to be able to give an AI system a “general” goal and then set it loose to make and execute a plan. Autonomously.Mustafa Suleyman, now head of AI at Microsoft, imagined it this way: In this new world, an AI agent would be able to “make $1 million on a retail web platform in a few months with just a $100,000 investment.” A year ago, he thought such a generally capable AI agent could be two years away. Pause for a moment and think about all the things this would entail: a piece of software with control over a bank account and the capacity to enter into contracts, engage in marketing research, create an online storefront, evaluate product performance and sales, and accrue a commercial reputation. For starters. Maybe it has to hire human workers. Or create copies of itself to perform specialized tasks. Obtain regulatory approvals? Respond to legal claims of contract breach or copyright violation or anticompetitive conduct?Shouldn’t we have some rules in place for these newcomers? Before they start building and selling products on our retail web platforms? Because if one of our AI companies announces they have these agents ready to unleash next year, we currently have nothing in place. If someone wants to work in our economy, they have to show their ID and demonstrate they’re authorized to work. If someone wants to start a business, they have to register with the secretary of state and provide a unique legal name, a physical address, and the name of the person who can be sued if the business breaks the rules. But an AI agent? Currently they can set up shop wherever and whenever they like.Put it this way: If that AI agent sold you faulty goods, made off with your deposit, stole your IP, or failed to pay you, who are you going to sue? How are all the norms and rules and institutions that we rely on to make all those complex trades and transactions in the modern economy function going to apply? Who is going to be accountable for the actions of this autonomous AI agent?The naïve answer is: whoever unleashed the agent. But there many problems here. First, how are you going to find the person or business who set the AI agent on its path, possibly, according to Suleyman, a few months ago? Maybe the human behind the agent has disclosed their identity and contact information at every step along the way. But maybe they haven’t. We don’t have any specific rules in place yet requiring humans to disclose like this. I’m sure smart lawyers will be looking for such rules in our existing law, but smart scammers, and aggressive entrepreneurs, will be looking for loopholes too. Even if you can trace every action by an AI agent back to a human you can sue, it’s probably not cheap or easy to do so. And when legal rules are more expensive to enforce, they are less often enforced. We get less deterrence of bad behavior and distortion in our markets.And suppose you do find the human behind the machine. Then what? The vision, remember, is one of an AI agent capable of taking a general instruction—go make a million bucks!—and then performing autonomously with little to no further human input. Moseying around the internet, making deals, paying bills, producing stuff. What happens when the agent starts messing up or cheating? These are agents built on machine learning: Their code was written by a computer. That code is extremely uninterpretable and unpredictable—there are literally billions of mathematical dials and switches being used by the machine to decide what to do next. When ChatGPT goes bananas and starts telling a New York Times journalist it’s in love with him and he should leave his wife, we have no idea why. Or what it might do if it now has access to bank accounts and email systems and online platforms. This is known as the alignment problem, and it is still almost entirely unsolved.So, here’s my prediction if you sue the human behind the AI agent when you find them: The legal system will have a very hard time holding them responsible for everything the agent does, regardless of how alien, weird, and unpredictable. Our legal systems hold people responsible for what they should have known better to avoid. It is rife with concepts like foreseeability and reasonableness. We just don’t hold people or businesses responsible very often for stuff they had no way of anticipating or preventing.And if we do want to hold them responsible for stuff they never imagined, then we need clear laws to do that.But that’s why the existential risk that keeps me up at night is what seems to be the very real and present risk that we’ll have a flood of AI agents joining our economic systems with essentially zero legal infrastructure in place. Effectively no rules of the road—indeed, no roads—to ensure we don’t end up in chaos.And I do mean chaos—in the sense of butterfly wings flapping in one region of the world and hurricanes happening elsewhere. AI agents won’t just be interacting with humans; they’ll be interacting with other AI agents. And creating their own sub-agents to interact with other sub-agents. This is the problem my lab works on—multi-agent AI systems, with their own dynamics at a level of complexity beyond just the behavior of a Gemini or GPT-4 or Claude. Multi-agent dynamics are also a major gap in our efforts to plug the hole of risks from frontier models: Red-teaming efforts to make sure that our most powerful language models can’t be prompted into helping someone build a bomb do not go near the question of what happens in multi-agent systems. What might happen when there are hundreds, thousands, millions of humans and AI agents interacting in transactions that can affect the stability of the global economy or geopolitical order? We have trouble predicting what human multi-agent systems—economies, political regimes—will produce. We know next to nothing about what happens when you introduce AI agents into our complex economic systems.But we do know something about how to set the ground rules for complex human groups. And we should be starting there with AI.Number one: Just as we require workers to have verifiable ID and a social security number to get a job, and just as we require companies to register a unique identity and legally effective address to do business, we should require registration and identification of AI agents set loose to autonomously participate in transactions online. And just as we require employers and banks to verify the registration of the people they hire and the businesses they open bank accounts for, we should require anyone transacting with an AI agent to verify their ID and registration. This is a minimal first step. It requires creating a registration scheme and some verification infrastructure. And answering some thorny questions about what counts as a distinct “instance” of an AI agent (since software can be freely replicated and effectively be in two—multiple—places at once and an agent could conceivably “disappear” and “reconstitute” itself). That’s what law does well: create artificial boundaries around things so we can deal with them. (Like requiring a business to have a unique name that no one else has registered, and protecting trademarks that could otherwise be easily copied. Or requiring a business or partnership to formally dissolve to “disappear.”) It’s eminently doable.The next step can go one of two ways, or both. The reason to create legally recognizable and persistent identities is to enable accountability—read: someone to sue when things go wrong. A first approach on this would be to require any AI agent to have an identified person or business who is legally responsible for its actions—and for that person or business to be disclosed in any AI transaction. States usually require an out-of-state business to publicly designate an in-state person or entity who can be sued in local courts in the event the business breaches a contract or violates the law. We could establish a similar requirement for an AI agent “doing business” in the state.But we may have to go further and make AI agents directly legally responsible for their actions. Meaning you could sue the AI agent. This may sound strange at first. But we already make artificial “persons” legally responsible: Corporations are “legal persons” that can sue and be sued in their own “name.” No need to track down the shareholders or directors or managers. And we have a whole legal apparatus around how these artificial persons must be constituted and governed. They have their own assets—which can be used to pay fines and damages. They can be ordered by courts to do or not do things. They can be stopped from doing business altogether if they engage in illegal or fraudulent activity. Creating distinct legal status for AI agents—including laws about requiring them to have sufficient assets to pay fines or damages as well as laws about how they are “owned” and “managed”—is probably something we’d have to do in a world with “more AI agents than people.”How far away is this world? To be honest, we—the public—really don’t know. Only the companies who are furiously investing billions into building AI agents—Google, OpenAI, Meta, Anthropic, Microsoft—know. And our current laws allow them to keep this information to themselves. And that’s why there’s another basic legal step we should be taking, now. Even before we create registration schemes for AI agents, we should create registration requirements for our most powerful AI models, as I’ve proposed (in brief and in detail) with others. That’s the only way for governments to gain the kind of visibility we need to even know where to start on making sure we have the right legal infrastructure in place. If we’re getting a whole slew of new alien participants added to our markets, shouldn’t we have the same capacity to know who they are and make them follow the rules as we do with humans?We shouldn’t just be relying on p-doom fights on X between technologists and corporate insiders about whether we are a few years, or a few decades, or a few hundred years away from these possibilities.The opinions expressed in Fortune.com commentary pieces are solely the views of their authors and do not necessarily reflect the opinions and beliefs of Fortune.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI发展 自主性 潜在威胁 规则缺失
相关文章