Yeah, all the more reason not to have them doing autonomous behaviors.
Rules of using AI:
#1: Never use AI to think for you
#2: Never use AI to do atomonous work
That leaves using them as knowledge assistants. In time, that will be realized as their only safe application. Safe to the user's minds, and safe to the user's environment. They are idiot savants, after all, having them do atomonous work is short sighted.
Sounds good on paper, but it has a game theory problem. If your efforts can always be out-raced by someone using AI to do autonomous work, don't you end up having to use it that way just to keep up?
Game theory ideas are great on paper, but in the real world it's messy. For simple, demo and concept sized uses, sure the AI doing it autonomously will succeed. Which betrays the reality that any real application with real world complexity that includes a dynamic environment and maintenance cannot be created by AI atomonmously while at the same time existing within an organization that can maintain it. They may create it, but it will be a shit show of cascading failure over time.
You know who outpaces you 100% of the time as you walk down the stairs? The guy jumping out of the window. Just because it is faster does not mean it is the right economic strategy. E.g. which contractor would you hire for your roof, that old roofer with 20+ years of experience or some AI startup that hires the cheapest subcontractors and "plan" your roof using a LLM?
The latter may be cheaper, sure. But too cheap can become very expensive quickly.
I love your analogy.
If this paper is right, then you might at first be outraced by a competitor using autonomous AI, but only until that competitor gets stabbed in the back by its own AI.
Which unfortunately might still be long enough for them to sink your business
And their customers won't care either way it seems
Or if they do care they won't have any real ability to do anything about it anyways
Maybe. The backstabbing rate is unknown so far. If it's high enough, then autonomy will be poor strategy.
It might be the trigger :)
From an economic perspective it requires LLMs and humans to have comparable outputs. That's not possible in all domains - at least in the near future.
Maybe they’ll outpace you, or maybe they’ll end up dying in a spectacular fiery crash?
What is the reason? This is a stress test. You can tell that by reading the first sentence of the article: "We stress-tested 16 leading models from multiple developers". In a stress test you want things to fail, otherwise you have learned very little about the stress the thing you are testing can take.
For physical things that has early limitations, but not for software. I would be very confused to see an AI stress test that did not end in failure, and would always question the test instead of thinking "wow, that must mean the thing is ready for autonomous action!"
Destructive stress testing is done on materials not "people"
Apparently it's now done on "people" (which, in a very important distinction, are not actual people but software)
Good luck with that.
We have a non-insignificant amount of people doing the #1 already, and the amount of people doing the #2 is only going to increase as more and more AIs are designed to be good at autonomous agentic behavior specifically.
The ship has long sailed on "just never let AIs do anything dangerous". If that was your game plan on AI safety, you need a new plan.
We also have a huge number of failing as they beg AI to do their work for them, which is intellectually damaging them. The ship always sails early filled to the brim with short sighted thinkers, all saying "this is it! this is the ship!" as it sinks.
> Never use AI to do atomonous work
> having them do atomonous work is short sighted
I also think they shouldn’t be doing atomonous work. Maybe autonomous work, but never atomonous.
accidentally a word
0 days without a word accident
Autumn mango futures soar like terrodactyls.