Presentation is loading. Please wait.

Presentation is loading. Please wait.

J. Blackmon. Could a synthetic thing be conscious?  A common intuition is: Of course not!  But the Neuron Replacement Thought Experiment might convince.

Similar presentations


Presentation on theme: "J. Blackmon. Could a synthetic thing be conscious?  A common intuition is: Of course not!  But the Neuron Replacement Thought Experiment might convince."— Presentation transcript:

1 J. Blackmon

2 Could a synthetic thing be conscious?  A common intuition is: Of course not!  But the Neuron Replacement Thought Experiment might convince you otherwise.

3 The Neuron Replacement Thought Experiment  Suppose bioengineers develop a synthetic neuron, a unit that is functionally identical to our own biological neurons.

4 The Neuron Replacement Thought Experiment  If one of your neurons is about to die, they can replace it with a perfect functional (although synthetic) duplicate.

5 The Neuron Replacement Thought Experiment  Imagine that this is done for one of your neurons.  You now have a synthetic neuron doing exactly what your biological neuron had been doing.

6 The Neuron Replacement Thought Experiment  Would there be any difference?  No one else would notice a difference in your behavior. After all, your behavior is driven by the activity of your neurons, and because this synthetic replacement is functionally equivalent, your behavior will be exactly the same.

7 The Neuron Replacement Thought Experiment  Suppose now that another of your neurons is replaced with a functionally equivalent synthetic neuron.  And another, and another, …

8 The Neuron Replacement Thought Experiment  What happens?  Do you gradually lose consciousness even though you continue to act exactly the same?  Do you reach some threshold at which consciousness just vanishes?  Or do you remain conscious, even when every single neuron has been replaced and your “brain” is entirely synthetic?

9 The Neuron Replacement Thought Experiment  Either you remain conscious throughout, or you lose your conscious experience at some point.  If you remain conscious throughout, then an artificial synthetic thing can be conscious.  If you lose your conscious experience at some point, then consciousness is a property that cannot be scientifically studied.

10 The Neuron Replacement Thought Experiment  If, on the other hand, you lose your conscious experience at some point, then consciousness is a property that cannot be scientifically studied at the level of behavior.  After all, your body’s behavior remains exactly the same. (And “you” continue to insist that “you” feel fine and are having conscious experience.)

11 The Neuron Replacement Thought Experiment  Either artificial consciousness is possible or consciousness cannot be detected or investigated by examining one’s behavior!  But if you think consciousness can be detected and investigated this way, then the answer to whether an artificial thing can be conscious is: Of course!

12 John Basl’s Account

13  Something has moral status if it’s worthy of our consideration in moral deliberation.  An artifact might have certain kind of moral status just because it belongs to someone else. Or because it’s instrumental to doing something morally required.  However, a being might have a different kind of moral status: moral considerability or inherent worth.

14  Something is a moral patient if it has a welfare composed of interests to be taken into account for the sake of the individual whose welfare it is.  Humans are moral patients.  Presumably, certain non-human animals are, too.  We’re going to investigate whether certain kinds of artifacts can be moral patients.

15  Something is a moral patient if it has a welfare composed of interests to be taken into account for the sake of the individual whose welfare it is.  Interests are those things which contribute to an individual’s welfare.

16  Psychological Interests: those had in virtue of having certain psychological capacities and states.  Teleo-Interests: those had in virtue of being goal- directed or teleologically organized.

17 The Easy Case  Basl: If x has certain human capacities, then x is a moral patient.  Basl considers and rejects the following alternatives:  Appeal to Religion  Appeal to Species

18 The Easy Case  Appeal to Religion: If x was chosen by God to be a moral patient, then x is a moral patient.  Problems  First, how do we know whether some x was chosen by God?  Second, how can one justify this to the governed given that some of the governed are secular?

19 The Easy Case  Appeal to Species: If x belongs to one of the special species, then x is a moral patient.  Problems  We can imagine aliens that are moral patients in the same way we are.  A subset of humans might evolve into a new species.

20 The Easy Case  So Basl rejects both the Religion and Species alternatives.  Again, if x has certain human capacities, then x is a moral patient.

21 The Harder Cases: Animals and Others

22  Basl: The capacity for attitudes toward sensory experiences is sufficient for moral patiency.  Note that Basl thinks simply having sensory experience is not enough.  Hitting a dog with a hammer is wrong not simply because the dog can experience pain, but because the dog has an aversive attitude to that sensory experience.

23 Epistemic Challenges  Problem of Other Minds: How do we truly know that other people have conscious minds like us?  Our best answer so far is that we know other humans are mentally like us because they are like us in terms of evolution, physiology, and behavior.  If x shares a common ancestor… If x is physiologically like us… If x behaves in similar ways…

24 Epistemic Challenges  Our best answer so far is that we know other humans are mentally like us because they are like us in terms of evolution, physiology, and behavior.  But with machines, we do not have these obvious connections.

25

26 How can non-sentient things have interests?  Basl builds from work in environmental ethics.  Why is acid bad for the maple tree, while water is good for it?  The answer cannot rely on the maple tree’s welfare unless it can be established that the tree has welfare.  However, the tree cannot suffer or be happy.

27 How can non-sentient things have interests?  Some people appeal to teleology.  A maple tree has the end/goal of survival and reproduction.  But if it cannot grow leaves, then it cannot meet its ends.  So its interests/ends are frustrated.

28 Teleo-Interests  Non-sentient beings can have interests. Thus they can have welfare and so be moral patients.  If non-sentient beings can have interests, why can’t (non-sentient) machines have them?  Basl considers one objection, the Objection from Derivativeness.

29 Objection from Derivativeness:  The teleological organization of artifacts is derivative of our interests. (Not so with organisms.)  Therefore, the only sense in which mere machines can have interests is derivative.  But derivative interests are insufficient for welfare and moral considerability/patiency.

30 Basl’s Reply to the Objection from Derivativeness  There are two kinds of derivativeness.  Use-Derivativeness: Machines which exist only for our use because of our needs and desires may have use-derivative interests. Their very existence derives from our intentions and interests.  Explanatory Derivativeness: The ends or teleo- organization of a mere machine can only be explained by reference to the intentions or ends of conscious beings. The explanation derives from our intentions.

31 Basl’s Reply to the Objection from Derivativeness  Now, consider organisms such as crops or pets. Yes, these things exist because of our interest. However, they clearly have interests of their own.  So use-derivativeness does not disqualify a thing for having interests of its own.

32 Basl’s Reply to the Objection from Derivativeness  Similarly, consider the case in which I play a significant role in the life and career choice of a child. That child’s preferences might not be explainable without reference to mine; however, the child still has interests of his or her own.  Thus explanatory derivativeness does not disqualify a thing for having interests of its own.

33 Basl says the Objection from Derivativeness is the best objection, but it fails. His alternative is the Comparable Welfare Thesis.

34

35 Comparable Welfare Thesis (CTW): If non- sentient organisms have teleo-interests, then mere machines have teleo-interests.

36 Comparable Welfare Thesis (CWT): If non-sentient organisms have teleo-interests, then mere machines have teleo-interests.  Note that the CWT, of course, does not say that mere machines have teleo-interests.  It just says that if non-sentient organisms have them, then so do mere machines.  The idea is that whatever justifies acknowledging teleo-interests in organisms also justifies acknowledging them in mere machines.

37 Comparable Welfare Thesis (CWT): If non- sentient organisms have teleo-interests, then mere machines have teleo-interests. But consider:  Basl has argued that derivativeness can’t rob a thing of its teleo-interests. So he has rejected the Objection from Derivativeness.  He has not refuted the position that teleo- interests must be acquired “naturally”.

38 Thus the environmental ethicist might take the following stance:  Agreed, an organism can’t be excluded from teleo-interests simply because someone created and molded it to suit their own ends. (The cases of the crops, pets, and child can all be accepted.)  But this doesn’t mean that mere machines have teleo-interests.

39 Thus the environmental ethicist might take the following stance (cont’d):  How can crops, pets, and children have teleo- interests while mere machines do not?  The environmental ethicist might try to argue that the crop, the pet, and the child are all members of natural kinds (plants, animals), while mere machines are not. This view faces a line-drawing problem. (Consider genetically modified organisms.)

40 Basl argues that teleo-interests are nonetheless unimportant. “You may recycle your computers with impunity.”

41

42  So long as machines are not conscious, we may do with them largely as we please.  But once we have something approaching artificial consciousness, we must consider whether these things are conscious and, crucially, whether they have attitudes.  We must then be careful with the epistemic uncertainties so as to avoid ignoring a possible person.


Download ppt "J. Blackmon. Could a synthetic thing be conscious?  A common intuition is: Of course not!  But the Neuron Replacement Thought Experiment might convince."

Similar presentations


Ads by Google