On Mon, Jan 13, 2025, 10:22 PM John Rose <johnr...@polyplexic.com> wrote:

> On Monday, December 23, 2024, at 11:17 AM, James Bowery wrote:
>
> *We may be talking at cross purposes here. I am referring to the theorem
> that no intelligence can model itsel*
>
>
> I keep thinking there is a way around Wolpert's theorem, agents fully
> modelling, if you go cross-Planck hologram or the universe modelling
> itself, because it is itself, it is own prediction. Unless agents need to
> be a subset of the universe to be considered an agent.
>

An agent modeling itself is the special case of two agents modeling each
other when both have the same software and initial state. Wolpert's theorem
says they both can't predict the other, even if they both know the software
and initial state of the other so that they could run a copy of the other
agent.

Proof: Suppose these two agents (not necessary identical) play rock
scissors paper. Agent 1 runs a copy of agent 2 and predicts it will play
rock, so it plays paper. Agent 2 runs a copy of agent 1 and knows it will
play paper, so it plays scissors. Who wins?

Wolpert's theorem is fundamental to the alignment problem. Either you can
predict what AI will do, or it can predict you, but not both. We use
prediction accuracy to measure intelligence. If you want to build an AI
that is smarter than you, then you can't predict what it will do. This
means you can't control it either, because that would require you to
predict the effects of your actions.

One argument against this is that intelligence is not a point on a line. So
in reality each agent can predict the other sometimes but not always. That
is today's reality. If by intelligence you mean a broad set of
capabilities, then the practical effect is that as AI gains capabilities,
it gets better at predicting you and you get worse at predicting it.

A second argument is that the goal of AI is not intelligence, it is
modeling human behavior. But this is an argument against a singularity. It
is not an argument against the loss of control. It means instead of being
controlled by autonomous self improving AI, you are controlled by
billionaires predicting the best way to enrich themselves at your expense.

I must repeat. The risk of AI is getting everything you want. You are
controlled by positive reinforcement. When you train a dog with treats, it
thinks it is controlling you.

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/Tff34429f975bba30-M633fd156ff3504b359828fdb
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to