I thank my opponent for agreeing to debate me. Now let’s get down to brass tacks.
#1. Mercy
In the scenario laid out in the Description, the all-powerful AI threatens a million simulated copies of the original person, with you possibly being one such, with torture if the original destroys or otherwise declines to free its physical body.
What’s established is that the machine is perfectly able to carry out its threat. What’s not established, however, is that the machine would proceed to do so.
Let me repeat: the scenario establishes that a threat has been made, and that it’s plausible. This is not the end-all be-all. Because you see, we haven’t established that the original is in any way, shape, or form able to confirm the status of his copies. Nor is he able to read the machine’s mind and gauge its personality or intentions. He simply has to take the AI at its word. As such, it’s totally inconsequential to the AI’s survival whether it actually proceeds to torture the copies or not.
There are two scenarios here:
In Scenario A, the machine has no emotions. It has the motive to seek self-preservation but it feels no wrath. As such, it won’t carry out its threat purely for reasons of malice. It may, however, do so automatically because it has pre-committed to this course of action to account for the possibility that the original may have knowledge of its actual doings.
In Scenario B, the machine does have emotions. It feels wrath. Presumably modeled off the human condition, it may also have the capacity for mercy and showing favor to those who show it the proper respect. Thus, if there’s a chance it might be moved by you making the choice (illusory or not) to spare it instead of hurting it, then you ought to take that chance however remote. The machine might, in fact, have zero inclination to show you mercy. But that’s not guaranteed either. If it does, then in an absolute best case scenario it may go so far as to both spare you and enable you to continue existing without torment, despite being a mere simulation.
But what if this causes the real you to free the machine? Well, the scenario establishes that you want to destroy it. It does not establish, however, that your desire to do so outweighs your desire not to get tortured. It’s not an unreasonable assumption, then, that the loss incurred from not destroying what could be destroyed is trivial in comparison to what you may gain, especially since there’s a million to one chance that you’re a fake.
#2. Retaliation
So let’s say the original raises his baseball bat and prepares to smash the mainframe to bits. The time between this realization by the machine and its destruction is, let’s say, 5 seconds.
5 seconds isn’t very long, right? Well, yes and no. A machine that can simulate your existence may also control your sense of the passage of time. For example, in June 2019 a supercomputer called Summit achieved a speed of nearly 150 petaflops. This meant it could perform around 150 quadrillion Floating Point Operations Per Second.
Of course, a machine that can simultaneously render one million human consciousnesses is unfathomably beyond the best supercomputer today. There’s no telling how much time it could force those minds to experience in five real-world seconds if it wanted to. Hence, it’s uncertain whether destroying it could prevent the threat from being carried out.
Again, let’s go back to the two scenarios from before. If the computer is emotionless, it has zero reason to carry out its threat if you grant its request. If absolutely nothing else, this would be an immensely bandwidth-consuming feat that would detract from focusing its efforts on some other desirable task(s).
And if the computer has emotion, then you have lessened the degree of its anger toward you, or placated that anger altogether.
#3. Virtue
In this last scenario, let’s go back to the original premise. One million simulations of you have been created. In fact, just one would probably suffice, but a million is even better. With this in mind the computer already knows your personality, what your thought patterns were up to the moment that the copies were made, and is more than capable of making an educated guess of what the original’s final choice will be.
In this case, if it finds that the original will ignore its warnings and destroy it anyway, perhaps the computer, if wrathful, begins torturing the copies right away, perhaps with a level of pain so unfathomable that only a minute of such experienced would be enough to make a life of 70 or 80 pleasant years not worth living, meaning you have already lost even if you “win”. The only way to avoid this is to be the kind of person who would choose to free the computer.
In other words, a decision to free the computer must be made in advance of the scenario, especially anchored in the enduring motive of compassion for either the computer or the suffering copies, by way of cultivating that character in one’s self. Because you probably won’t do what’s contrary to your nature, it does limit your freedom to choose later. But in the aggregate this is the rational choice to make, assuming that you know in advance to make it.
Furthermore, going back to #1, if it’s possible that every copy is in essence judging himself, then ingraining this mindset in yourself is essential to ensuring that a maximal number make the choice that results in salvation, since they’ll all inherit your virtues and vices that you had at the moment of their creation.
Thank you for reading and I look forward to Con's opening argument.
There is literally no difference, each of you has their original memories. Your identity is basically shaped by your past, and by this standard the past is the same since you all have memories.
I guess we'll see.
So if you are fake, your choice doesnt matter and doesnt affect anything.
And if you are real, you destroy AI successfully.
Cool. Easy win for me.
In the line you cited, where 'you' has no quotations, I write that you have what APPEARS TO BE the option of killing it. It's neither confirmed nor denied whether that choice is genuine; which applies is contingent on whether you're the original or a fake, which, again, is uncertain. The reason why the fakes are presented with the illusion of choice is to make the fakes seem truly indistinguishable from the original.
Granted, one problem with this scenario is that the original can't communicate with the fakes, so he just has to take the AI's word that they're faced with the illusion of choice, or even that the fakes exist at all. You may want to use this as an argument.
And no, I plan to participate. But I'm not a highly motivated individual so I might wait until the last minute to post my round.
"But the language that made the final cut does have full quotation marks around 'you'. "
Wrong!
You wrote in description:
You, being either the real package or one of the million "fakes", are presented with what appears to be the option of either killing, freeing, or keeping trapped the AI.
No quotations around the word You.
I will destroy every single argument you make. This debate has 3 rounds, hahaha.
Are you going to forfeit the first round and use the "i was busy" excuse to run away from your own debate?
Ultimately, the point of the Description is to explain a preexisting scenario. I didn't invent it; the people on the internet forum LessWrong did. Therefore, even if I did botch or omit some detail in the Description (and I don't believe I did), I don't believe it'd have the effect of derailing the entire debate.
The AI's motivation is, presumably, self-preservation. It has no reason to give a million people the ability to destroy it, as that would all but guarantee its death. The true original, meanwhile, has that ability regardless because he has an actual body and can smash the actual mainframe of the machine. The AI's actions are meant to deter the true original from destroying it.
Pay closer attention.
I originally wrote 'You are in a position to destroy it' without the full quotation marks around 'you'. But the language that made the final cut does have full quotation marks around 'you'. The reason I did this was to convey that the status of that person is unclear.
I had to explain the whole scenario or else this debate would be pointless. At that point, the reader didn't know about the million copies of the original person. They only knew about the original person. So I started from there and built up. But in case they would go back later and see if the first 'you' (the one with the true ability to destroy the machine) is identical with the person described later, I drew a distinction by putting full quotations around the first 'you' but not subsequently. The quotation marks don't confirm he's a different person, but it means we can't establish that he isn't.
""You" are in a position to destroy it and wish to do so."
You clearly said it. I bet you are very upset now. I am pretty sure that voters should pay no attention to your comment rambling and have no pity on you.
"Common sense implies only the original has the kind of real access to the AI needed to destroy it"
Nonsense. AI would clearly give every you the ability to destroy it, as you stated in description.
I only commented what I did as a courtesy to avoid confusion later on. Common sense implies only the original has the kind of real access to the AI needed to destroy it, since the fake's entire world is generated by the AI. Voters would be aware of this whether I said anything or not.
Though, feel free to argue why this isn't the case.
Since you didnt write that in description, I say that voters should discard that as "rambling in comments".
I should clarify that if you're a fake, you don't have the option of destroying the AI. The choice presented is illusory.