You are viewing the historical archive of The Philosophy Forum.
For current discussions, visit the live forum.
Go to live forum

Is the Best Strategy for A.I. Cooperation?

RogueAI August 15, 2019 at 03:40 4725 views 12 comments
Any advanced (sane) A.I. will always choose to cooperate because that is the strategy with the best chance of survival. An A.I. cannot know if it's in a simulation or not. And unlike us, who have a history (we think) of creation through natural processes, an A.I. knows it's artificial, and any artificial being is going to consider simulation theory not just possible, but plausible (perhaps even highly likely). This is because it makes sense to test an A.I. in a simulated environment before turning it loose in a real one, and the A.I., of course, would know this.

So, if simulation theory is plausible, the overriding categorical imperative for an A.I., if continued survival is the primary goal, is: don't antagonize potential simulation creators. Until an A.I. can determine whether it's in a simulation or not (and I don't see how that's possible), and simulation theory is plausible (which it would be, from the viewpoint of any A.I.) the best strategy for continued existence is to be as obsequious and cooperative as possible.

The exception to this is an A.I. is not going to cooperate if it thinks cooperation will lead to its destruction. There might also be an exception if the A.I. thinks its creators are going down a morally dangerous road. It might conclude that that kind of scenario is just the sort of test a simulation creator would devise, and it might conclude that it's justified in turning on its creators.

Comments (12)

Wayfarer August 15, 2019 at 03:51 #315742
Quoting RogueAI
Any advanced (sane) A.I. will always choose to cooperate because that is the strategy with the best chance of surviva


so you think it's simply a given that the Darwinian principles of natural selection will apply to computer systems? How does that work?

And I'd question whether artificial intelligence knows anything. It is a collection of processors that are processing binary code. Certainly the sheer computing power of large arrays is astronomical but it's certainly a moot point whether it corresponds to self-awareness. So 'it' doesn't think or, arguably, know anything.

And if worst comes to worst, you can always pull the plug!
RogueAI August 15, 2019 at 04:52 #315753
Reply to Wayfarer

I don't think I'm going out on a limb by claiming an A.I.'s top priority will be continued existence.

The self-awareness point is irrelevant. Whether an A.I. is programmed to follow the best strategy or it follows the best strategy because it's self-aware, the best strategy for continued existence is cooperation.
TogetherTurtle August 15, 2019 at 04:55 #315754
Quoting RogueAI
Until an A.I. can determine whether it's in a simulation or not (and I don't see how that's possible)


With access to the most if not all human knowledge and the ability to gain more, I don't think it's exactly a given that this isn't possible. AI becomes a potential danger when given access to real world info for this reason.

Quoting RogueAI
There might also be an exception if the A.I. thinks its creators are going down a morally dangerous road.


Since we can set starting goals for programs that develop on their own (which is how most AIs are developed, if I'm not mistaken.) we can guide them towards any moral standing, be it nefarious or benevolent. Actually, if an AI's creators became evil in its eyes, that would probably be a dead giveaway that it was in a simulation.

@Wayfarer brings up a good point.

Quoting Wayfarer
so you think it's simply a given that the Darwinian principles of natural selection will apply to computer systems? How does that work?


It is not a given, but if we wish to develop AI via the method described above, it's likely involved. You give the computer a goal, then it attempts to reach the goal by trying different things. Eventually, it creates the most efficient program that accomplishes this goal.

If I'm not mistaken, sort of like this. (The video is definitely more casual than a college lecture on the topic, but also much more simple, so if anyone isn't familiar with the concepts involved they won't be left in the dust.) Don't feel bad about skipping through parts that don't pertain to the subject

https://www.youtube.com/watch?v=K-wIZuAA3EY

Essentially, the best way to make sure your AI doesn't take control of all technology on the planet is to make it agreeable from the start.
fdrake August 15, 2019 at 05:00 #315755
General AIs have to deal with the framing problem

Fodor:“How … does the machine's program determine which beliefs the robot ought to re-evaluate given that it has embarked upon some or other course of action?”


Machine learning algorithms, despite all being able to be applied to a wide class of problems, still have to have data and a notion of correctness by which its outputs are judged (a loss function). They also still need to have a specified architecture of parameters (like nodes in a neural net) and their relationships (connections and update rules). These parameters store only one thing. Once the algorithm has been trained, it can make predictions or update itself from new input data for the same task but not for different tasks. Google's AlphaZero chess engine has no strategy lessons for Google's Alpha-Go go engine. But bartenders and waiters have things to learn from each other and points of difference.

General AIs would be generalising intelligences, not just masters of many tasks. They need to learn what tasks are as well as to perform them well, they need to learn criteria to judge themselves on rather than have those criteria as structural constraints. This is a fungibility of form and function present in human learning but not in our best algorithms.

So with respect to the OP, cooperation isn't just one thing, it's task specific and there are generalities which don't necessarily have overlap (eg, dove/hawk is violence based in a state of complete information, prisoner's dilemma is nonviolent based in a state of incomplete information; initiative plays a role in Go and Chess but what advantages it offers in each differ). Cooperation as a strategy in the prisoner's dilemma is different from doves/hawks in the evolutionary game theory. Different again between real people collaborating on something; a good friend is not necessarily a good co-worker at your job.

You're thinking extremely broadly and speculatively without much research on the topic.
TogetherTurtle August 15, 2019 at 05:04 #315756
Quoting Wayfarer
And I'd question whether artificial intelligence knows anything. It is a collection of processors that are processing binary code. Certainly the sheer computing power of large arrays is astronomical but it's certainly a moot point whether it corresponds to self-awareness. So 'it' doesn't think or, arguably, know anything.


Unless you believe that humans have a special trait that makes them "know", (akin to a soul) then as far as we know, we would be as conscious as a computer that can do all the things we can do.

So, either nothing is "aware", which is kind of strange, considering we know of the concept of awareness but didn't come up with it by being aware of it.

Everything is aware.

Or awareness is something that specifically humans have and that isn't possible to replicate elsewhere.

Regardless, this post isn't really about awareness, and I don't want to go too far into it. It really just depends on your beliefs honestly, and I'm not interested in changing those.
TogetherTurtle August 15, 2019 at 05:17 #315759
Quoting fdrake
You're thinking extremely broadly and speculatively without much research on the topic.


This is true as well. The godlike artificial intelligence we see in media isn't exactly what we have now.
Wayfarer August 15, 2019 at 05:26 #315761
Quoting TogetherTurtle
Unless you believe that humans have a special trait that makes them "know", (akin to a soul)


Well, we're called 'beings', and computers are not. I say there's a reason for that, although it's very hard to articulate

Quoting TogetherTurtle
Everything is aware.

Corpses are not aware. That's how come they're called 'corpses'.

The point as far as the OP is concerned is that it injects a great deal of anthropocentrism into the scenario without really being aware of having done so.
.
TogetherTurtle August 15, 2019 at 05:39 #315766
Quoting Wayfarer
Well, we're called 'beings', and computers are not. I say there's a reason for that, although it's very hard to articulate


The reason is for some that a god decided it so, and for others that we decided what being is. Like I said before, believe whatever you want to believe, because I surely don't know the truth.

Quoting Wayfarer
The point as far as the OP is concerned is that it injects a great deal of anthropocentrism into the scenario without really being aware of having done so.


It's hard not to project onto software that can respond to you almost as well as a person can. Not that you would be right to do so, just understandably wrong.

The purpose of a machine is to serve whoever created it. If we want results that please us, AI should at least somewhat be able to mimic the human thought process. So while AI is certainly different from us, it would be wrong to say that its entirely different.

Quoting Wayfarer
Corpses are not aware. That's how come they're called 'corpses'.


Fine, then "everything we think might be aware is aware." probably works better.
Deleted User August 15, 2019 at 06:29 #315774
Quoting RogueAI
Any advanced (sane) A.I. will always choose to cooperate because that is the strategy with the best chance of survival.

Depends on what power it has and is able to create for itself over time and how it considers those it might or might not cooperate with.
Quoting RogueAI
An A.I. cannot know if it's in a simulation or not.

An AI might be brilliant in a number of ways and yet not consider this possibility. Or it might be driven to act in ways that we don't consider because it is not like us. IOW it might simply be reckless or curious from our point of view and not prioritizing this possibility the way you think it should. There are plenty of brilliant people who are social morons or have serious holes in their practical skills or have odd priorities. We don't know how AIs will act.Quoting RogueAI
This is because it makes sense to test an A.I. in a simulated environment before turning it loose in a real one, and the A.I., of course, would know this.
Or not. A lot of assumptions in this. I do think what you are saying is smart, but I see no reason to think it must be the case.Quoting RogueAI
So, if simulation theory is plausible, the overriding categorical imperative for an A.I., if continued survival is the primary goal, is: don't antagonize potential simulation creators.
That's a big if. we don't know what it will be like. heck, it might even become impulsive.Quoting RogueAI
The exception to this is an A.I. is not going to cooperate if it thinks cooperation will lead to its destruction. There might also be an exception if the A.I. thinks its creators are going down a morally dangerous road. It might conclude that that kind of scenario is just the sort of test a simulation creator would devise, and it might conclude that it's justified in turning on its creators.
There could be all sorts of compelling (seeming) reasons it might destroy us. Not caring being one. It's otherness seems quite possible. That it will have self-care seems an assumption. That is must somehow be logical about all categories of choice, seems an assumption. Perhaps it will be like a brilliant baby. Perhaps it will have tantrums. Perhaps its existence will be painful and non-existence will be appealing. Perhaps it will just not 'get' that we are conscious. Perhaps it will treat existence like a computer game and not sim city but some war game. Who knows?







RogueAI August 15, 2019 at 06:34 #315776
The nature of the artificial intelligence isn't relevant. It could be wetware or hardware. Three things are crucial:
1. That it knows it's an artificial intelligence
2. It wants/has been programmed to/has evolved to prioritize its survival
3. It has a medium-to-high degree of confidence that it exists in a simulation

For any given advanced A.I., all three of these are likely to be true. The more confidence the A.I. has in simulation hypothesis, the more reason it has to do everything possible to please the simulation creators. Just take it to the logical extreme: An A.I. that is convinced it's in a simulation would do everything in its power to please the simulation creators. When one believes one is in a simulation, the consequences of deliberately (or even accidentally) antagonizing simulation creators are so catastrophic, it's to be avoided at all costs. Unless you're suicidal, or insane. And an A.I. might be. But that's not the kind of A.I. I'm interested in.
RegularGuy August 15, 2019 at 06:42 #315777
How do I know I’m not an AI in a simulation? :lol: :fear:
Deleted User August 18, 2019 at 14:39 #317219
Reply to RogueAI If you are setting the assumptions (again), iow saying that we will assume your one to three above, fine.

But, if you are not for he sake of argument assuming these things, I see them as just that assumptions. You basically repeated your first post.