Too late. I added a 5 minute cron job for cursor AI's compose tab in agent mode that keeps replying "keep going, think of more fixes and features, random ideas as fine, do it all for me". I won't pull the plug.
This is a purely procedural question, not supporting or critiquing in any way-- other than this reads kind of like an editorial with the format of a scientific paper. The question is... are there rules about what constitutes a paper or can you just put whatever you want in there as long as you follow "scientific paper format"?
I really enjoy Margaret Mitchell‘s podcast (she is the first author on the paper), and perhaps I missed something important in the paper, but:
Shouldn’t we treat separately autonomous agent we write ourselves, or purchase to run on our own computers, on our own data and that use public APIs for data?
If Margaret is reading this thread, I am curious what her opinion is.
For autonomous agents controlled by corporations and governments, I mostly agree with the paper.
I'd recommend looking for other sources of information if you're relying on someone who co-authored the paper that introduced the most misleading and uninformed term of the LLM era: "stochastic parrot".
No not really. There's no power in the world that can restrain this in it's current form even mildly much less absolutly. Why do you think that would be even slightly possible?
For the same reason we can regulate other things? Encryption is regulated, for example. There "just" needs to be international co-operation, in the case of AI.
Despite doing a pretty decent job of containing the risk we're still on the clock until something terrible happens with nuclear war. Humanity appears to be well on track to killing millions to billions of people; rolling the dice relatively regularly waiting for a 1% chance to materialize.
If we only handle AI that well doom is probable. It has economic uses, unlike nuclear weapons, so there will be a thriving black market dodging the safety concerns.
True, so we need to make sure we don't find ourselves in a mess before it happens. Right now I don't see nearly enough concern given to risk management in industry. The safeguards companies put on their models are trivially subverted by hackers. We don't even know how to cope with an AI that would attempt to subvert its own constitution.
You will run an autonomous ai agent on your own hardware or by having your own local ai pass out commands to distributed systems online, ai, real people, or just good old fashioned programming. There is no stopping this.
It is in fact possible to stop training runs that consume billions of dollars in electricity and in GPU rental or depreciation costs. If no one does such a training run, then no one can release the weights of the model that would have been produced by the run, so you won't be able to run the model (which would never come into existence) on your own hardware. I don't care if you run DeepSeek R1 in your basement till the end of time. What my friends and I want to stop is the creation of more capable future models.
It is also quite possible for our society to decide that deep learning is too dangerous and to outlaw teaching and publishing about it, which would not completely stop the discovery of algorithmic deep-learning improvements (because some committed deep-learning enthusiasts would break the law) but would slow the discovery rate way, way down.
But it’s not actually
possible for our society to decide that. In the real world, at this moment when laws and norms are gone and a billionaire obsessed with AI has power, that will 100% not happen. It won’t happen in the next several years, and that is the time left to do what you are saying. Pretending otherwise is a waste of time.
I prefer to retain some hope that our civilization has a future and that humans or at least human values and preferences have some place in that future civilization.
And most people who think AI "progress" is so dangerous that it must be stopped before it is too late have loose confidence intervals extending for at least a couple of decades (as opposed to just a few years) as to when it definitely becomes too late.
In the incredible case that we develop fully autonomous agents capable of crippling the world, that would mean we developed fully autonomous agents capable of keeping it safe.
Unless the first one is so advanced no other can challenge it, that is.
How did you jump to that conclusion? The agent will be limited by the capabilities under its control. We have the technological ability to cripple world now and we don't have the technological means to prevent it. Give one AI control of the whole US arsenal and the objective of ending the world. Give another AI the capabilities of the rest of the world and the objective of protecting it. Would you feel safe?
> We have the technological ability to cripple world now and we don't have the technological means to prevent it
Humans have prevented it many times, but not specifically by technological ability. If Putin/Trump/Xi Ping wanted a global nuclear war, they'd better have the means to launch the nukes themselves in secret because the chain of command will challenge them.
If an out-of-control AI could discover a circuitous way to access nukes, an antagonist AI of equal capabilities should be able to figure it out too, and warn the humans in the loop.
I agree that AI development should be made responsibly, but not all people do, and it's impossible to put the cat back in the bag. The limiting factor these days is hardware, as a true AGI will likely need even more of it than our current LLMs.
Out-of-control AI is sci-fi fearmongering, it's not about worming through systems. It will be doing exactly what it was placed there to do. It will be a human failing that puts armageddon in it's hands. And since humans have NO MEANS to prevent armageddon (The predominant policy is in fact doubling down on destruction with MAD), there will be no way to place AI in command of this defense. The asymmetrical relationship between destruction and creation will mean there will never be a defense.
No one should be allowed to develop software that has bugs in it that lead to unlawful harm to others. And if they do it anyway they should be punished lawfully.
The thing with autonomous AI is that we already know it cannot be made safe in a way that satisfies lawmakers who are fully informed about how it works… unless they are bribed, I suppose.
Our analysis reveals that risks to people increase with the autonomy of a system: The more control a user cedes to an AI agent, the more risks to people arise. Particularly concerning are safety risks, which affect human life and impact further values.
I feel that these kinds of statements are more effective at promoting AI than limiting it. It reinforces the assumption that such powerful AI is behind the corner. It hypes up AI and the result is likely more money and resources being put into it.
Imagine if the A-bomb was being openly developed. What title would have contributed more to funding and research, "The A-bomb (is terribly powerful and) should not be developed" or "The A-bomb will never work"? Except the A-bomb did work and in a surprisingly short time, while autonomous AGI is still a conjecture.
This is quite concerning seeing that the authors are all affiliated with huggingface. Hopefully they won't start censoring what models you can or can't upload because they seem certain things shouldn't be developed.