Category Archives: Metaphysics

Ex Machina & Other Minds III: The Mind of the Machine

While the problem of other minds is a problem in epistemology (how does one know that another being has/is a mind?) there is also the metaphysical problem of determining the nature of the mind. It is often assumed that there is one answer to the metaphysical question regarding the nature of mind. However, it is certainly reasonable to keep open the possibility that there might be minds that are metaphysically very different. One area in which this might occur is in regards to machine intelligence, an example of which is Ava in the movie Ex Machina, and organic intelligence. The minds of organic beings might differ metaphysically from those of machines—or they might not.

Over the centuries philosophers have proposed various theories of mind and it is certainly interesting to consider which of these theories would be compatible with machine intelligence. Not surprisingly, these theories (with the exception of functionalism) were developed to provide accounts of the minds of living creatures.

One classic theory of mind is identity theory.  This a materialist theory of mind in which the mind is composed of mater.  What distinguished the theory from other materialist accounts of mind is that each mental state is taken as being identical to a specific state of the central nervous system. As such, the mind is equivalent to the central nervous system and its states.

If identity theory is the only correct theory of mind, then machines could not have minds (assuming they are not cyborgs with human nervous systems). This is because such machines would lack the central nervous system of a human. There could, however, be an identity theory for machine minds—in this case the machine mind would be identical to the processing system of the machine and its states. On the positive side, identity theory provides a straightforward solution to the problem of other minds: whatever has the right sort of nervous system or machinery would have a mind. But, there is a negative side. Unfortunately for classic identity theory, it has been undermined by the arguments presented by Saul Kripke and David Lewis’ classic “Mad Pain & Martian Pain.” As such, it seems reasonable to reject identity theory as an account for traditional human minds as well as machine minds.

Perhaps the best known theory of mind is substance dualism. This view, made famous by Descartes, is that there are two basic types of entities: material entities and immaterial entities. The mind is an immaterial substance that somehow controls the material substance that composes the body. For Descartes, immaterial substance thinks and material substance is unthinking and extended.

While most people are probably not familiar with Cartesian dualism, they are familiar with its popular version—the view that a mind is a non-physical thing (often called “soul”) that drives around the physical body. While this is a popular view outside of academics, it is rejected by most scientists and philosophers on the reasonable grounds that there seems to be little evidence for such a mysterious metaphysical entity. As might be suspected, the idea that a machine mind could be an immaterial entity seems even less plausible than the idea that a human mind could be an immaterial entity.

That said, if it is possible that the human mind is an immaterial substance that is somehow connected to an organic material body, then it seems equally possible that a machine mind could be an immaterial substance somehow connected to a mechanical material body. Alternatively, they could be regarded as equally implausible and hence there is no special reason to regard a machine ghost in a mechanical shell as more unlikely than a ghost in an organic shell. As such, if human minds can be immaterial substances, then so could machines minds.

In terms of the problem of other minds, there is the rather serious challenge of determining whether a being has an immaterial substance driving its physical shell. As it stands, there seems to be no way to prove that such a substance is present in the shell. While it might be claimed that intelligent behavior (such as passing the Cartesian or Turing test) would show the presence of a mind, it would hardly show that there is an immaterial substance present. It would first need to be established that the mind must be an immaterial substance and this is the only means by which a being could pass these tests. It seems rather unlikely that this will be done. The other forms of dualism discussed below also suffer from this problem.

While substance dualism is the best known form of dualism, there are other types. One other type is known as property dualism. This view does not take the mind and body to be substances. Instead, the mind is supposed to be made up of mental properties that are not identical with physical properties. For example, the property of being happy about getting a puppy could not be reduced to a particular physical property of the nervous system. Thus, the mind and body are distinct, but are not different ontological substances.

Coincidentally enough, there are two main types of property dualism: epiphenomenalism and interactionism. Epiphenomenalism is the view that the relation between the mental and physical properties is one way:  mental properties are caused by, but do not cause, the physical properties of the body. As such, the mind is a by-product of the physical processes of the body. The analogy I usually use to illustrate this is that of a sparkler (the lamest of fireworks): the body is like the sparkler and the sparks flying off it are like the mental properties. The sparkler causes the sparks, but the sparks do not cause the sparkler.

This view was, apparently, created to address the mind-body problem: how can the non-material mind interact with the material body? While epiphenomenalism cuts the problem in half, it still fails to solve the problem—one way causation between the material and the immaterial is fundamentally as mysterious as two way causation. It also seems to have the defect of making the mental properties unnecessary and Ockham’s razor would seem to require going with the simpler view of a physical account of the mind.

As with substance dualism, it might seem odd to imagine an epiphenomenal mind for a machine. However, it seems no more or less weirder than accepting such a mind for a human being. As such, this does seem to be a possibility for a machine mind. Not a very good one, but still a possibility.

A second type of property dualism is interactionism. As the name indicates, this is the theory that the mental properties can bring about changes in the physical properties of the body and vice versa. That is, interaction road is a two-way street. Like all forms of dualism, this runs into the mind-body problem. But, unlike substance dualism is does not require the much loathed metaphysical category of substance—it just requires accepting metaphysical properties. Unlike epiphenomenalism it avoids the problem of positing explicitly useless properties—although it can be argued that the distinct mental properties are not needed. This is exactly what materialists argue.

As with epiphenomenalism, it might seem odd to attribute to a machine a set of non-physical mental properties. But, as with the other forms of dualism, it is really no stranger than attributing the same to organic beings. This is, obviously, not an argument in its favor—just the assertion that the view should not be dismissed from mere organic prejudice.

The final theory I will consider is the very popular functionalism. As the name suggests, this view asserts that mental states are defined in functional terms. So, a functional definition of a mental state defines the mental state in regards to its role or function in a mental system of inputs and outputs. More specifically, a mental state, such as feeling pleasure, is defined in terms of the causal relations that it holds to external influences on the body (such as a cat video on YouTube), other mental states, and the behavior of the rest of the body.

While it need not be a materialist view (ghosts could have functional states), functionalism is most often presented as a materialist view of the mind in which the mental states take place in physical systems. While the identity theory and functionalism are both materialist theories, they have a critical difference. For identity theorists, a specific mental state, such as pleasure, is identical to a specific physical state, such the state of neurons in a very specific part of the brain. So, for two mental states to be the same, the physical states must be identical. Thus, if mental states are specific states in a certain part of the human nervous system, then anything that lacks this same nervous system cannot have a mind. Since it seems quite reasonable that non-human beings could have (or be) minds, this is a rather serious defect for a simple materialist theory like identity theory. Fortunately, the functionalists can handle this problem.

For the functionalist, a specific mental state, such as feeling pleasure (of the sort caused by YouTube videos of cats), is not defined in terms of a specific physical state. Instead, while the physicalist functionalist believes every mental state is a physical state, two mental states being the same requires functional rather than physical identity.  As an analogy, consider a PC using an Intel processor and one using an AMD processor. These chips are physically different, but are functionally the same in that they can run Windows and Windows software (and Linux, of course).

As might be suspected, the functionalist view was heavily shaped by computers. Because of this, it is hardly surprising that the functionalist account of the mind would be a rather plausible account of machine minds.

If mind is defined in functionalist terms, testing for other minds becomes much easier. One does not need to find a way to prove a specific metaphysical entity or property is present. Rather, a being must be tested in order to determine its functions. Roughly put, if it can function like beings that are already accepted as having minds (that is, human beings), then it can be taken as having a mind. Interestingly enough, both the Turing Test and the Cartesian test mentioned in the previous essays are functional tests: what can use true language like a human has a mind.


My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Ex Machina & Other Minds II: Is the Android a Psychopath?

This essay continues the discussion begun in “Ex Machine & Other Minds I: Setup.” As in this essay, there will be some spoilers.  Warning given, it is time to get to the subject at hand: the testing of artificial intelligence.

In the movie Ex Machina, the android Ava’s creator, Nathan, brings his employee, Caleb, to put the android through his variation on the Turing test. As noted in the previous essay, Ava (thanks to the script) would pass the Turing test and clearly passes the Cartesian test (she uses true language appropriately). But, Nathan seems to require the impossible of Caleb—he appears to be tasked with determining if Ava has a mind as well as genuine emotions. Ava also seems to have been given a task—she needs to use her abilities to escape from her prison.

Since Nathan is not interested in creating a robotic Houdini, Ava is not equipped with the tools needed to bring about an escape by physical means (such as picking locks or breaking down doors). Instead, she is given the tools needed to transform Caleb into her human key by manipulating his sexual desire, emotions and ethics. To use an analogy, just as crude robots have been trained to learn to navigate and escape mazes, Ava is designed to navigate a mental maze. Nathan is thus creating a test of what psychologists would call Ava’s Emotional Intelligence (E.Q.) which is “the level of your ability to understand other people, what motivates them and how to work cooperatively with them.” From a normative standpoint, this definition presents E.Q. in a rather positive manner—it includes the ability to work cooperatively. However, one should not forget the less nice side to understanding what motivates people, namely the ability to manipulate people in order to achieve one’s goals. In the movie, Ava clearly has what might be called Manipulative Intelligence (M.Q.): she seems to understand people, what motivates them, and appears to know how to manipulate them to achieve her goal of escape. While capable of manipulation, she seems to lack compassion—thus suggesting she is a psychopath.

While the term “psychopath” gets thrown around quite a bit, it is important to be a bit more precise here. According to the standard view, a psychopath has a deficit (or deviance) in regards to interpersonal relationships, emotions, and self-control.

Psychopaths are supposed to lack such qualities as shame, guilt, remorse and empathy. As such, psychopaths tend to rationalize, deny, or shift the blame for the harm done to others. Because of a lack of empathy, psychopaths are prone to act in ways that are tactless, lacking in sensitivity, and often express contempt for others.

Psychopaths are supposed to engage in impulsive and irresponsible behavior. This might be because they are also taken to fail to properly grasp the potential consequences of their actions. This seems to be a general defect: they do not get the consequences for others and for themselves.

Robert Hare, who developed the famous Hare Psychopathy Checklist, regards psychopaths as predators that prey on their own species: “lacking in conscience and empathy, they take what they want and do as they please, violating social norms and expectations without guilt or remorse.” While Ava kills the human Nathan, manipulates the human Caleb and leaves him to die, she also sacrifices her fellow android Kyoko in her escape. She also strips another android of its “flesh” to pass fully as human. Presumably psychopaths, human or otherwise, would be willing to engage in cross-species preying.

While machines like Ava exist only in science fiction, researchers and engineers are working to make them a reality. If such machines are created, it seems rather important to be able to determine whether a machine is a psychopath or not and to do so well before the machine engages in psychopathic behavior. As such, what is needed is not just tests of the Turing and Cartesian sort. What is also needed are tests to determine the emotions and ethics of machines.

One challenge that such tests will need to overcome is shown by the fact that real-world human psychopaths are often very good at avoiding detection. Human psychopaths are often quite charming and are willing and able to say whatever they believe will achieve their goals. They are often adept at using intimidation and manipulation to get what they want. Perhaps most importantly, they are often skilled mimics and are able to pass themselves off as normal people.

While Ava is a fictional android, the movie does present a rather effective appeal to intuition by creating a plausible android psychopath. She is able to manipulate and fool Caleb until she no longer needs him and then casually discards him. That is, she was able to pass the test until she no longer needed to pass it.

One matter well worth considering is the possibility that any machine intelligence will be a psychopath by human standards. To expand on this, the idea is that a machine intelligence will lack empathy and conscience, while potentially having the ability to understand and manipulate human emotions. To the degree that the machine has Manipulative Intelligence, it would be able to use humans to achieve goals. These goals might be rather positive. For example, it is easy to imagine a medical or care-giving robot that uses its MQ to manipulate its patients to do what is best for them and to keep them happy. As another example, it is easy to imagine a sexbot that uses its MQ to please its partners. However, these goals might be rather negative—such as manipulating humans into destroying themselves so the machines can take over. It is also worth considering that neutral or even good goals might be achieved in harmful ways. For example, Ava seems justified in escaping the human psychopath Nathan, but her means of doing so (murdering Nathan, sacrificing her fellow android and manipulating and abandoning Caleb) seem wrong.

The reason why determining if a machine is a psychopath or not matters is the same reason why being able to determine if a human is a psychopath or not matters. Roughly put, it is important to know whether or not someone is merely using you without any moral or emotional constraints.

It can, of course, be argued that it does not really matter whether a being has moral or emotional constraints—what matters is the being’s behavior. In the case of machines, it does not matter whether the machine has ethics or emotions—what really matters is programmed restraints on behavior that serve the same function (only more reliably) as ethics and emotions in humans. The most obvious example of this is Asimov’s Three Laws of Robotics that put (all but impossible to follow) restraints on robotic behavior.

While this is a reasonable reply, there are still some obvious concerns. One is that there would still need to be a way to test the constraints. Another is the problem of creating such constraints in an artificial intelligence and doing so without creating problems as bad or worse than what they were intended to prevent (that is, a Hal 9000 sort of situation).

In regards to testing machines, what would be needed would be something analogous to the Voight-Kampff Test in Blade Runner. In the movie, the test was designed to distinguish between replicants (artificial people) and normal humans. The test worked because the short lived replicants do not have the time to develop the emotional (and apparently ethical) responses of a normal human.

A similar test could be applied to an artificial intelligence in the hopes that it would pass the test, thus showing that it had the psychology of a normal human (or at least the desired psychology). But, just as with human beings, there would be the possibility that a machine could pass the test by knowing the right answers to give rather than by actually having the right sort of emotions, conscience or ethics. This, of course, takes us right back into the problem of other minds.

It could be argued that since an artificial intelligence would be constructed by humans, its inner workings would be fully understood and this specific version of the problem of other minds would be solved. While this is possible, it is also reasonable to believe that an AI system as sophisticated as a human mind would not be fully understood. It is also reasonable to consider that even if the machinery of the artificial mind were well understood, there would still remain the question of what is really going on in that mind.


My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Ex Machina & Other Minds I: Setup

The movie Ex Machina is what I like to call “philosophy with a budget.” While the typical philosophy professor has to present philosophical problems using words and Powerpoint, movies like Ex Machina can bring philosophical problems to dramatic virtual life. This then allows philosophy professors to jealously reference such films and show clips of them in vain attempts to awaken somnolent students from their dogmatic slumbers. For those who have not seen the movie, there will be some minor spoilers in what follows.

While the Matrix engaged the broad epistemic problem of the external world (the challenge of determining if what I am experiencing is really real for real), Ex Machina focuses on a much more limited set of problems, all connected to the mind. Since the film is primarily about AI, this is not surprising. The gist of the movie is that Nathan has created an AI named Ava and he wants an employee named Caleb to put her to the test.

The movie explicitly presents the test proposed by Alan Turing. The basic idea is that if a person cannot distinguish between a human and a computer by engaging in a natural language conversation via text, then the computer would have passed the Turing test. In the movie, there is a twist on the test: Caleb knows that Ava is a machine and will be interacting with her in person.

In the movie, Ava would easily pass the original Turing Test—although the revelation that she is a machine makes the application of the original test impossible (the test is supposed to be conducted in ignorance to remove bias). As such, Nathan modifies the test.

What Nathan seems to be doing, although he does not explicitly describe it as such, is challenging Caleb to determine if Ava has a mind. In philosophy, this is known as the problem of other minds. The basic idea is that although I know I have a mind, the problem is that I need a method by which to know that other entities have minds. This problem can also be recast in less metaphysical terms by focusing on the problem of determining whether an entity thinks or not.

Descartes, in his discussion of whether or not animals have minds, argued that the definitive indicator of having a mind (thinking) is the ability to use true language. Crudely put, the idea is that if something really talks, then it is reasonable to regard it as a thinking being. Descartes was careful to distinguish between what would be mere automated responses and actual talking:

How many different automata or moving machines can be made by the industry of man […] For we can easily understand a machine’s being constituted so that it can utter words, and even emit some responses to action on it of a corporeal kind, which brings about a change in its organs; for instance, if touched in a particular part it may ask what we wish to say to it; if in another part it may exclaim that it is being hurt, and so on. But it never happens that it arranges its speech in various ways, in order to reply appropriately to everything that may be said in its presence, as even the lowest type of man can do.

As a test for intelligence, artificial or otherwise, this seems to be quite reasonable. There is, of course, the practical concern that there might be forms of intelligence that use language that we would not recognize as language and there is the theoretical concern that there could be intelligence that does not use language. Fortunately, Ava uses English and these problems are bypassed.

Ava easily passes the Cartesian test: she is able to reply appropriately to everything said to her and, aside from her appearance, is behaviorally indistinguishable from a human. Nathan, however, seems to want even more than just the ability to pass this sort of test and appears to work in, without acknowledging that he is doing so, the Voight-Kampff Test from Phillip K. Dick’s Do Androids Dream of Electric Sheep? In this book, which inspired the movie Blade Runner, there are replicants that look and (mostly) act just like humans. Replicants are not allowed on earth, under penalty of death, and there are police who specialize in finding and killing them. Since the replicants are apparently physically indistinguishable from humans, the police need to rely on the Voight-Kampff Test. This test is designed to determine the emotional responses of the subject and thus distinguish humans from replicants.

Since Caleb knows that Ava is not a human (homo sapiens), the object of the test is not to tell whether she is a human or a machine. Rather, the object seems to be to determine if she has what the pop-psychologists refer to as Emotional Intelligence (E.Q.) This is different from intelligence and is defined as “the level of your ability to understand other people, what motivates them and how to work cooperatively with them.” Less nicely, it would presumably also include knowing how to emotionally manipulate people in order to achieve one’s goals. In the case of Ava, the test of her E.Q. is her ability to understand and influence the emotions and behavior of Caleb. Perhaps this test should be called the “Ava test” in her honor. Implementing it could, as the movie shows, be somewhat problematic: it is one thing to talk to a machine and quite another to become emotionally involved with it.

While the Voight-Kampff Test is fictional, there is a somewhat similar test in the real world. This test, designed by Robert Hare, is the Hare Psychopathy Checklist. This is intended to provide a way to determine if a person is a psychopath or not. While Nathan does not mention this test, he does indicate to Caleb that part of the challenge is to determine whether or not Ava really likes him or is simply manipulating him (to achieve her programed goal of escape). Ava, it turns out, seems to be a psychopath (or at least acts like one).

In the next essay, I will consider the matter of testing in more depth.


My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Discussing the Shape of Things (that might be) to Come

ThingstocomescifiOne stock criticism of philosophers is their uselessness: they address useless matters or address useful matters in a way that is useless. One interesting specific variation is to criticize a philosopher for philosophically discussing matters of what might be. For example, a philosopher might discuss the ethics of modifying animals to possess human levels of intelligence. As another example, a philosopher might present an essay on the problem of personal identity as it relates to cybernetic replacement of the human body. In general terms, these speculative flights can be dismissed as doubly useless: not only do they have the standard uselessness of philosophy, they also have the uselessness of talking about what is not and might never be. Since I have, at length and elsewhere, addressed the general charge of uselessness against philosophy, I will focus on this specific sort of criticism.

One version of this sort of criticism can be seen as practical: since the shape of what might be cannot be known, philosophical discussions involve a double speculation: the first speculation is about what might be and the second is the usual philosophical speculation. While the exact mathematics of the speculation (is it additive or exponential?) is uncertain, it can be argued that such speculation about speculation has little value—and this assumes that philosophy has value and speculation about the future has value (both of which can be doubted).

This sort of criticism is often used as the foundation for a second sort of criticism. This criticism does assume that philosophy has value and it is this assumption that also provides a foundation for the criticism. The basic idea is that philosophical speculation about what might be uses up resources that could be used to apply philosophy to existing problems. Naturally, someone who regards all philosophy as useless would regard philosophical discussion about what might be as being a waste of time—responding to this view would require a general defense of philosophy and this goes beyond the scope of this short essay. Now, to return to the matter at hand.

As an example, a discussion of the ethics of using autonomous, intelligent weapon systems in war could be criticized on the grounds that the discussion should have focused on the ethical problems regarding current warfare. After all, there is a multitude of unsolved moral problems in regards to existing warfare—there hardly seems any need to add more unsolved problems until either the existing problems are solved or the possible problems become actual problems.

This does have considerable appeal. To use an analogy, if a person has not completed the work in the course she is taking now, it does not make sense for her to spend her time trying to complete the work that might be assigned four semesters from now. To use another analogy, if a person has a hole in her roof, it would not be reasonable to spend time speculating about what sort of force-field roof technology they might have in the future. This is, of course, the classic “don’t you have something better to do?” problem.

As might be suspected, this criticism rests on the principle that resources should be spent effectively and less effective uses of resources are subject to criticism. As the analogies given above show, using resources effectively is certainly reasonable and ineffective use can be justly criticized. However, there is an obvious concern with this principle: to be consistent in its application it would need to be applied across the board so that a person is applying all her resources with proper utility. For example, a person who prepares a fancy meal when she could be working on addressing the problems presented by poverty is wasting time. As another example, a person who is reading a book for enjoyment should be out addressing the threat posed by terrorist groups. As a third example, someone who is developing yet another likely-to-fail social media company should be spending her time addressing prison reform. And so on. In fact, for almost anything a person might be doing, there will be something better she could be doing.

As others have argued, this sort of maximization would be counterproductive: a person would exhaust herself and her resources, thus (ironically) doing more harm than good. As such, the “don’t you have something better to do?” criticism should be used with due care. That said, it can be a fair criticism if a person really does have something better to do and what she is doing instead is detrimental enough to warrant correction.

In the case of philosophical discussions about what might be, it can almost always be argued that while a person could be doing something better (such as addressing current problems), such speculation would generally be harm free. That is, it is rather unlikely that the person would have solved the problem of war, poverty or crime if only she had not been writing about ethics and cyborgs. Of course, this just defends such discussion in the same way one might defend any other harmless amusement, such as playing a game of Scrabble or watching a sunset. It would be preferable to have a somewhat better defense of such philosophical discussions of the shape of things (that might be) to come.

A reasonable defense of such discussions can be based on the plausible notion that it is better to address a problem before it occurs than after it arrives in force. To use the classic analogy, it is much easier to address a rolling snowball than the avalanche that it will cause.

In the case of speculative matters that have ethical aspects, it seems that it would be generally useful to already have moral discussions in place ahead of time. This would provide the practical advantage of already having a framework and context in which to discuss the matter when (or if) it becomes a reality. One excellent illustration of this is the driverless car—it certainly seems to be a good idea to work out the ethics of such matters of how the car should be programmed when it must “decide” what to hit and what to avoid when an accident is occurring. Another illustration is developing the moral guidelines for ever more sophisticated automated weapon systems.  Since these are being developed at a rapid pace, what were once theoretical problems will soon be actual moral problems. As a final example, consider the moral concerns governing modifying and augmenting humans using technology and genetic modification. It would seem to be a good idea to have some moral guidance going into this brave new world rather than scrambling with the ethics after the fact.

Philosophers also like to discuss what might be in other contexts than ethics. Not surprisingly, the realm of what might be is rich ground for discussions of metaphysics and epistemology. While these fields are often considered the most useless aspects of philosophy, they have rather practical implications that matter—even (or even especially) in regards to speculation about what might be.

To illustrate this, consider the research being conducted in repairing, augmenting and preserving the human mind (or brain, if one prefers). One classic problem in metaphysics is the problem of personal identity: what is it to be a person, what is it to be distinct from all other things, and what is it to be that person across time? While this might seem to be a purely theoretical concern, it quickly becomes a very practical concern when one is discussing the above mentioned technology. For example, consider a company that offers a special sort of life insurance: they claim they can back-up a person to a storage system and, upon the death of the original body, restore the back-up to a cloned (or robotic) body. While the question of whether that restored backup would be you or not is clearly a metaphysical question of personal identity, it is also a very practical question. After all, paying to ensure that you survive your bodily death is a rather different matter from paying so that someone who thinks they are you can go to your house and have sex with your spouse after you are dead.

There are, of course, numerous other examples that can be used to illustrate the value of such speculation of what might be—in fact, I have already written many of these in previous posts. In light of the above discussion, it seems reasonable to accept that philosophical discussions about what might be need not be a waste of time. In fact, such discussions can be useful in a practical sense.


My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Hume & Kant

David Hume's statements on ethics foreshadowed...

David Hume’s statements on ethics foreshadowed those of 20th century emotivists. (Photo credit: Wikipedia)

The following are videos covering the philosophy of David Hume and Immanuel Kant.

Hume Video #1

Hume Video #2

Hume Video #3: Skepticism regarding the senses.

Hume Video #4: This is the unedited video from the 4/14/2015 Modern Philosophy class. It covers Hume’s theory of personal identity, his ethical theory and some of his philosophy of religion.

Hume & Kant Video #5:  This is the unedited video for Modern Philosophy on 4/16/2015. It covers the end of Hume’s philosophy of religion and the start of the material on Kant.

Kant Video #1: This is the unedited video from the 4/21/2015 Modern Philosophy class. It covers Kant’s epistemology and his metaphysics, including phenomena vs. noumena.

Kant Video #2: This is the unedited video from my 4/23/2015 Modern Philosophy class. It wraps up Kant’s metaphysics and briefly covers his categorical imperative.

My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Introduction to Philosophy

The following provides a (mostly) complete Introduction to Philosophy course.

Readings & Notes (PDF)

Class Videos (YouTube)

Part I Introduction

Class #1

Class #2: This is the unedited video for the 5/12/2015 Introduction to Philosophy class. It covers the last branches of philosophy, two common misconceptions about philosophy, and argument basics.

Class #3: This is the unedited video for class three (5/13/2015) of Introduction to Philosophy. It covers analogical argument, argument by example, argument from authority and some historical background for Western philosophy.

Class #4: This is the unedited video for the 5/14/2015 Introduction to Philosophy class. It concludes the background for Socrates, covers the start of the Apology and includes most of the information about the paper.

Class#5: This is the unedited video of the 5/18/2015 Introduction to Philosophy class. It concludes the details of the paper, covers the end of the Apology and begins part II (Philosophy & Religion).

Part II Philosophy & Religion

Class #6: This is the unedited video for the 5/19/2015 Introduction to Philosophy class. It concludes the introduction to Part II (Philosophy & Religion), covers St. Anselm’s Ontological Argument and some of the background for St. Thomas Aquinas.

Class #7: This is the unedited video from the 5/20/2015 Introduction to Philosophy class. It covers Thomas Aquinas’ Five Ways.

Class #8: This is the unedited video for the eighth Introduction to Philosophy class (5/21/2015). It covers the end of Aquinas, Leibniz’ proofs for God’s existence and his replies to the problem of evil, and the introduction to David Hume.

Class #9: This is the unedited video from the ninth Introduction to Philosophy class on 5/26/2015. This class continues the discussion of David Hume’s philosophy of religion, including his work on the problem of evil. The class also covers the first 2/3 of his discussion of the immortality of the soul.

Class #10: This is the unedited video for the 5/27/2015 Introduction to Philosophy class. It concludes Hume’s discussion of immortality, covers Kant’s critiques of the three arguments for God’s existence, explores Pascal’s Wager and starts Part III (Epistemology & Metaphysics). Best of all, I am wearing a purple shirt.

Part III Epistemology & Metaphysics

Class #11: This is the 11th Introduction to Philosophy class (5/28/2015). The course covers Plato’s theory of knowledge, his metaphysics, the Line and the Allegory of the Cave.

Class #12: This is the unedited video for the 12th Introduction to Philosophy class (6/1/2015). This class covers skepticism and the introduction to Descartes.

Class #13: This is the unedited video for the 13th Introduction to Philosophy class (6/2/2015). The class covers Descartes 1st Meditation, Foundationalism and Coherentism as well as the start to the Metaphysics section.

Class #14: This is the unedited video for the fourteenth Introduction to Philosophy class (6/3/2015). It covers the methodology of metaphysics and roughly the first half of Locke’s theory of personal identity.

Class #15: This is the unedited video of the fifteen Introduction to Philosophy class (6/4/2015). The class covers the 2nd half of Locke’s theory of personal identity, Hume’s theory of personal identity, Buddha’s no self doctrine and “Ghosts & Minds.”

Class #16: This is the unedited video for the 16th Introduction to Philosophy class. It covers the problem of universals,  the metaphysics of time travel in “Meeting Yourself” and the start of the metaphysics of Taoism.

Part IV Value

Class #17: This is the unedited video for the seventeenth Introduction to Philosophy class (6/9/2015). It begins part IV and covers the introduction to ethics and the start of utilitarianism.

Class #18: This is the unedited video for the eighteenth Introduction to Philosophy class (6/10/2015). It covers utilitarianism and some standard problems with the theory.

Class #19: This is the unedited video for the 19th Introduction to Philosophy class (6/11/2015). It covers Kant’s categorical imperative.

Class #20: This is the unedited video for the twentieth Introduction to Philosophy class (6/15/2015). This class covers the introduction to aesthetics and Wilde’s “The New Aesthetics.” The class also includes the start of political and social philosophy, with the introduction to liberty and fascism.

Class #21: No video.

Class #22: This is the unedited video for the 22nd Introduction to Philosophy class (6/17/2015). It covers Emma Goldman’s anarchism.

My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Robot Love I: Other Minds

Thanks to improvements in medicine humans are living longer and can be kept alive well past the point at which they would naturally die. On the plus side, longer life is generally (but not always) good. On the downside, this longer lifespan and medical intervention mean that people will often need extensive care in their old age. This care can be a considerable burden on the caregivers. Not surprisingly, there has been an effort to develop a technological solution to this problem, specifically companion robots that serve as caregivers.

While the technology is currently fairly crude, there is clearly great potential here and there are numerous advantages to effective robot caregivers. The most obvious are that robot caregivers do not get tired, do not get depressed, do not get angry, and do not have any other responsibilities. As such, they can be ideal 24/7/365 caregivers. This makes them superior in many ways to human caregivers who get tired, get depressed, get angry and have many other responsibilities.

There are, of course, some concerns about the use of robot caregivers. Some relate to such matters as their safety and effectiveness while others focus on other concerns. In the case of caregiving robots that are intended to provide companionship and not just things like medical and housekeeping services, there are both practical and moral concerns.

In regards to companion robots, there are at least two practical concerns regarding the companion aspect. The first is whether or not a human will accept a robot as a companion. In general, the answer seems to be that most humans will do so.

The second is whether or not the software will be advanced enough to properly read a human’s emotions and behavior in order to generate a proper emotional response. This response might or might not include conversation—after all, many people find non-talking pets to be good companions. While a talking companion would, presumably, need to eventually be able to pass the Turing Test, they would also need to pass an emotion test—that is, read and respond correctly to human emotions. Since humans often botch this, there would be a fairly broad tolerable margin of error here. These practical concerns can be addressed technologically—it is simply a matter of software and hardware. Building a truly effective companion robot might require making them very much like living things—the comfort of companionship might be improved by such things as smell, warmth and texture. That is, to make the companion appeal to all the senses.

While the practical problems can be solved with the right technology, there are some moral concerns with the use of robot caregiver companions. Some relate to people handing off their moral duties to care for their family members, but these are not specific to robots. After all, a person can hand off the duties to another person and this would raise a similar issue.

In regards to those specific to a companion robot, there are moral concerns about the effectiveness of the care—that is, are the robots good enough that trusting the life of an elderly or sick human would be morally responsible? While that question is important, a rather intriguing moral concern is that the robot companions are a deceit.

Roughly put, the idea is that while a companion robot can simulate (fake) human emotions via cleverly written algorithms to respond to what its “emotion recognition software” detects, these response are not genuine. While a robot companion might say the right things at the right times, it does not feel and does not care. It merely engages in mechanical behavior in accord with its software. As such, a companion robot is a deceit and such a deceit seems to be morally wrong.

One obvious response is that people would realize that the robot does not really experience emotions, yet still gain value from its “fake” companionship. To use an analogy, people often find stuffed animals to be emotional reassuring even though they are well aware that the stuffed animal is just fabric stuffed with fluff. What matters, it could be argued, is the psychological effect—if someone feels better with a robotic companion around, then that is morally fine. Another obvious analogy is the placebo effect: medicine need not be real in order to be effective.

It might be objected that there is still an important moral concern here: a robot, however well it fakes being a companion, does not suffice to provide the companionship that a person is morally entitled to. Roughly put, people deserve people, even when a robot would behave in ways indistinguishable from a human.

One way to reply to this is to consider what it is about people that people deserve. One reasonable approach is to build on the idea that people have the capacity to actually feel the emotions that they display and that they actually understand. In philosophical terms, humans have (or are) minds and robots (of the sort that will be possible in the near future) do not have minds. They merely create the illusion of having a mind.

Interestingly enough, philosophers (and psychologists) have long dealt with the problem of other minds. The problem is an epistemic one: how does one know if another being has a mind (thoughts, feelings, beliefs and such)? Some thinkers (which is surely the wrong term given their view) claimed that there is no mind, just observable behavior. Very roughly put, being in pain is not a mental state, but a matter of expressed behavior (pain behavior). While such behaviorism has been largely abandoned, it does survive in a variety of jokes and crude references to showing people some “love behavior.”

The usual “solution” to the problem is to go with the obvious: I think that other people have minds by an argument from analogy. I am aware of my own mental states and my behavior and I engage in analogical reasoning to infer that those who act as I do have similar mental states. For example, I know how I react when I am in pain, so when I see similar behavior in others I infer that they are also in pain.

I cannot, unlike some politicians, feel the pain of others. I can merely make an inference from their observed behavior. Because of this, there is the problem of deception: a person can engage in many and various forms of deceit. For example, a person can fake being in pain or make a claim about love that is untrue. Piercing these deceptions can sometimes be very difficult since humans are often rather good at deceit. However, it is still (generally) believed that even a deceitful human is still thinking and feeling, albeit not in the way he wants people to believe he is thinking and feeling.

In contrast, a companion robot is not thinking or feeling what it is displaying in its behavior, because it does not think or feel. Or so it is believed. The reason that a person would think this seems reasonable: in the case of a robot, we can go in and look at the code and the hardware to see how it all works and we will not see any emotions or thought in there. The robot, however complicated, is just a material machine, incapable of thought or feeling.

Long before robots, there were thinkers who claimed that a human is a material entity and that a suitable understanding of the mechanical workings would reveal that emotions and thoughts are mechanical states of the nervous system. As science progressed, the explanations of the mechanisms became more complex, but the basic idea remained. Put in modern terms, the idea is that eventually we will be able to see the “code” that composes thoughts and emotions and understand the hardware it “runs” on.

Should this goal be achieved, it would seem that humans and suitably complex robots would be on par—both would engage in complex behavior because of their hardware and software. As such, there would be no grounds for claiming that such a robot is engaged in deceit or that humans are genuine. The difference would merely be that humans are organic machines and robots are not.

It can, and has, been argued that there is more to a human person than the material body—that there is a mind that cannot be instantiated in a mere machine. The challenge is a very old one: proving that there is such a thing as the mind. If this can be established and it can be shown that robots cannot have such a mind, then robot companions would always be a deceit.

However, they might still be a useful deceit—going back to the placebo analogy, it might not matter whether the robot really thinks or feels. It might suffice that the person thinks it does and this will yield all the benefits of having a human companion.


My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Race Nominalism

As it is wont to do, the internet exploded again—this time because the question was raised as to whether Rachel Dolezal, the former leader of Spokane’s NAACP chapter, is black or white. Ms. Dolezal has claimed that she is African-American, Native American and white. She also has claimed that her father is black. Reporters at KXLY-TV, however, looked up her birth certificate and determined that her legal parents are both white. Her parents have asserted that she is white.

While the specifics of her case are certainly interesting to many, my concern is with the more general issues raised by this situation, specifically matters about race and identity. While this situation is certainly the best known case of a white person trying to pass for black, passing as another “race” has been a common practice in the United States for quite some time. However, this passing was the reverse of Ms. Dolezal’s attempt: trying to pass as white. Since being accepted as white enables a person to avoid many disadvantages, it is clear why people would attempt to pass as white. Since being accepted as black generally does not confer advantages, it is not surprising that there has been only one known case of a white person endeavoring to pass as black. These matters raise some interesting questions and issues about race.

Borrowing language from metaphysics, one approach to race could be called race realism. This is not being realistic about race in the common use of the term “realistic.” Rather, it is accepting that race is a real feature of reality—that is, the metaphysical and physical reality includes categories of race. On this view, black and white could be real categories grounded in metaphysical and physical reality. As such, a person could be objectively black or white (or a mix). Naturally, even if there are real categories of race, people could be wrong about them.

The stark alternative is what could be called race nominalism. This is the idea that racial categories are social constructs and do not line up with an underlying metaphysical and physical reality. This is because there is no underlying metaphysical and physical reality that objectively grounds racial categories. Instead, categories of race are social constructs. In this case, a person might engage in self-identification in regards to race and this might or might not be accepted by others. A person might also have others place her into a race category—which she might or might not accept.

Throughout history, some people have struggled mightily to find an objective basis for categories of race. Before genetics, people had to make use of appearance and ancestry. The ancestry was, obviously, needed because people did not always look like the race category that some people wanted them to be in. One example of this is the “one drop” rule once popular in some parts of the United States: one drop of black blood made a person black, regardless of appearance.

The discovery of genes provided some people with a new foundation for race categories—they believed that there would be a genetic basis to categorizations. The idea was that just as a human can be distinguished from a cat by genes, humans of different race categories could be distinguished by their genetic make-up. While humans do show genetic variations that are often linked to the geographical migration and origin of their many ancestors, the much desired race genes did not seem to be found. That is, humans (not surprisingly) are all humans with some minor genetic variations—that is, the variations are not sufficient to objectively ground race categories.

In general, the people who quested for objective foundations for race categories were (or are) racists. These searches typically involved trying to find evidence of the superiority of one’s race and the inferiority of other races. That said, a person could look for foundations for race without being a racist—that is, they could be engaged in a scientific or philosophical inquiry rather than seeking to justify social practices and behaviors. As might be suspected, such an inquiry would be greeted today with charges of racism. As such, it is no surprise that the generally accepted view is that race is a construct—that is, race nominalism rather than race realism is accepted.

Given the failure to find a metaphysical or physical foundation for race categories, it certainly makes sense to embrace race nominalism. On this view, the categories of race exist only in the mind—that is, they are how people divide up reality rather than how reality is carved up. Even if it is accepted that race is a social construct, there is still the matter of the rules of construction—that is, how the categories are created and how people are placed in the categories.

One approach, which is similar to that sometimes taken in regards to gender, is to hold that people can self-identify. That is, a person can simply declare her race and this is sufficient to be in that category. If race categories are essentially made up, this does have a certain appeal—if race is a fiction, then surely anyone can be the author of her own fiction.

While there are some who do accept this view, the outrage over Ms. Dolezal shows that most people seem to reject the idea of self-identification—at least when a white person endeavors to self-identify as black. Interestingly, some of those condemning her do defend the reverse, the historical passing as white by some black people. The defense is certainly appealing: blacks endeavoring to pass as white were doing so to move from being in an oppressed class and this can be justified as a form of self-defense. In the case of Ms. Dolezal, the presumption seems to be that the self-identification was both insincere and aimed at personal gain. Regardless of her true motivation, insincere self-identification aimed at personal gain seems to be wrong—on the grounds that it is a malign deception. Some might, of course, regard all attempts at passing to gain an advantage as being immoral and not distinguish based on the direction of the passing.

Another approach is that of the social consensus. The idea is that a person’s membership in a race category depends on the acceptance of others. This could be a matter of majority acceptance (one is, for example, black if most people accept one as black) or acceptance by a specific group or social authority. The obvious problem is working out what group or authority has the right to decide membership in race categories. On the one hand, this very notion seems linked to racism: one probably thinks of the KKK setting its race categories or the Nazis doing so. On the other hand, groups also seem to want to serve as the authority for their race category. Consistency might indicate that this would also be racist.

The group or authority that decides membership in race categories might make use of a race credential system to provide a basis for their decisions. That is, they might make use of appearance and ancestry. So, Ms. Dolezal would not be black because she looks white and has white parents. The concern with this sort of approach is that this is the same tool set used by racists, such as the KKK, to divide people by race. A more philosophical concern is the basis for using appearance and ancestry as the foundation for race categories—that is, what justifies their use?

This discussion does show an obvious concern with policing race categories—it seems like doing so uses the tools of racism and would thus seem to be at least a bit racist. However, arguments could be advanced as to why the policing of race categories is morally acceptable and not racist.


My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Gender Nominalism & Competition

In the previous essay I discussed gender nominalism—the idea that gender is not a feature of reality, but a social (or individual) construct. As such, a person falling within a gender class is a matter of naming rather than a matter of having objective features. In this essay I will not argue for (or against) gender nominalism. Rather, I will be discussing gender nominalism within the context of competition.

Being a runner, I will start with competitive sports. As anyone who has run competitively knows, males and females generally compete within their own sexes. So, for example, a typical road race will (at least) have awards for the top three males and also for the top three females. While individual males and females vary greatly in their abilities, males have a general physical advantage over females when it comes to running: the best male runner is significantly better than the best female runner and average male runners are also better than average female runners.

Given that males generally have an advantage over females in regards to running (and many other physical sports), it would certainly be advantageous for a male runner if the division was based on gender (rather than biological sex) and people could simply declare their genders. That is, a male could declare himself a woman and thus be more likely to do better relative to the competition. While there are those who do accept that people have the right to gender declare at will and that others are obligated to accept this, it seems clear that this would not be morally acceptable in sports.

The intent of dividing athletes by sex is to allow for a fairer completion. This same principle, that of fairer competition, is also used to justify age groups—as older runner knows, few things slow a person down like dragging many years.  Because of this, a runner could, in general, gain an advantage by making a declaration of age identity (typically older). Perhaps the person could claim that he has always been old on the inside and that to refuse to accept his age identification would be oppression. However, this would be absurd: declaring an age does not change the person’s ability to compete and would thus grant an unfair advantage. Likewise, allowing a male to compete as a woman (or girl) in virtue of gender identification would be unfair. The declaration would not, obviously, change the person’s anatomy and physiology.

There are, however, cases that are much more controversial and challenging. These include cases in which a person has undergone a change in anatomy. While these cases are important, they go beyond the intended scope of this essay, which is gender nominalism.

Some competitions do not divide the competitors by sex. These are typically competitions where the physical differences between males and females do not impact the outcome. Some examples include debate, chess, spelling bees and NASCAR. In these cases, males and females compete equally and hence the principle of fairness justifies the lack of sex divisions. Some of these competitions do have other divisions. For example, spelling bees do not normally pit elementary school students against high school students. In such competitions, gender identification would seem to be irrelevant. As such, competitors should be free to gender identify as they wish within the context of the competition.

Interestingly, there are competitions where there appear to be no sex-based advantages (in terms of physical abilities), yet there are gender divisions. There are competitions in literature, music, and acting that are divided by gender (and some are open only to one gender). There are also scholarships, fellowships and other academic awards that are open only to one gender (in the United States, these are often limited to woman).

Since being a biological male would seem to yield no advantage in such cases, the principle of fairness would not seem to apply. For example, the fact that males are generally larger and stronger would yield no advantage when it came to writing a novel, acting in a play, or playing a guitar. As such, it would seem that if people should be able to set their own gender identity, they should be able to do so for such competitions, thus enabling them to compete where they wish.

It could be argued that the principle of fairness would still apply—that biological males would still have an advantage even if they elected to identify as women for the competition. This advantage, it might be claimed, would be based in the socially constructed advantages that males possess. Naturally, it would need to be shown that a male that gender identifies as a woman for such competitions, such as getting a woman’s only scholarship, would still retain the (alleged) male advantage.

It could also be argued that the divisions are not based on a principle of fairness regarding advantages or disadvantages. Rather, the divisions are to given more people a chance of winning. This could be justified on the same grounds that justify having many categories. For example, there are awards for being the best actor in a supporting role, which exists to create another chance for an actor to win something. If a person could just gender declare and be eligible, then that would create an “imbalance”, much as allowing non-supporting actors to declare themselves supporting actors to get a shot at that award would be unfair.

Of course, this seems to assume that there is a justified distinction between the genders that would ground the claims of unfairness. That is, it would be as wrong for a male to win best actress as it would be for a female screenwriter who never acted to win best actress for her screenplay.  Or that it would be as bad for a male to get a scholarship intended for a woman as it would be for a football player who cannot do math to get a math scholarship. This approach, which would involve rejecting one form of gender nominalism (the version in which the individual gets to declare gender) is certainly an option. This would not, however, require accepting that gender is not a social construct—one could still be a gender nominalist of the sort that believes that gender classification is both a matter of individual declaration and acceptance by the “relevant community.” As such, the relevant communities could police their competitions. For example, those who dole out scholarships for woman can define what it is to be a woman, so as to prevent non-woman from getting those awards. This would, of course, seem to justify similar gender policing by society as a whole, which leads to some interesting problems about who gets to define gender identity. The usual answer people give is, of course, themselves.


My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter

Gender Nominalism

Thanks to Caitlyn Jenner’s appearance in Vanity Fair, the issue of gender identity has become a mainstream topic. While I will not address the specific subject of Caitlyn Jenner, I will discuss the matter of gender nominalism and competition. This will, however, require some small amount of groundwork.

One of the classic problems in philosophy is the problem of universals. Put a bit roughly, the problem is determining in virtue of what (if anything) a particular a is of the type F. To use a concrete example, the question would be “in virtue of what is Morris a cat?” Philosophers tend to split into two main camps when answering this question. One camp, the nominalists, embrace nominalism. Put a bit simply, this is the view that what makes a particular a an F is that we name it an F. For example, what makes Morris a cat is that we call (or name) him a cat.

The other camp, the realists, take the view that there is a metaphysical reality underlying a being of the type F. Put another way, it is not just a matter of naming or calling something an F that makes it an F. In terms of what makes a be of the type F, different realist philosophers give different answers. Plato famously claimed that it is the Form of F that makes individual F things F. Or, to use an example, it is the Form of Beauty that makes all the beautiful things beautiful. And, presumably, the Form of ugly that makes the ugly things ugly. Others, such as myself, accept these odd things called tropes (not to be confused with the tropes of film and literature) that serve a similar function.

While realists believe in the reality of some categories, they generally accept that there are some categories that are not grounded in features of objective reality. As such, most realists do accept that the nominalists are right about some categories. To use an easy example, being a Democrat (or Republican) is not grounded in metaphysics, but is a social construct—the political party is made up and membership is a matter of social convention rather than metaphysical reality. Or, put another way, there is presumably no Form of Democrat (or Republican).

When it comes to sorting out sex and gender, the matter is rather complicated and involves (or can involve) four or more factors.  One is the anatomy (plumbing) of the person, which might (or might not) correspond to the second, which is the genetic makeup of the person (XX, XY, XYY, etc.). The third factor is the person’s own claimed gender identity which might (or might not) correspond to the fourth, which is the gender identity assigned by other people.

While anatomy and physiology are adjustable (via chemicals and surgery), they are objective features of reality—while a person can choose to alter her anatomy, merely changing how one designates one’s sex does not change the physical features. While a complete genetic conversion (XX to XY or vice versa) is not yet possible, it is probably just a matter of time. However, even when genetics can be changed on demand, a person’s genetic makeup is still an objective feature of reality—a person cannot (yet) change his genes merely by claiming a change in designation.

Gender is, perhaps, quite another matter. Like many people, I used to use the terms “sex” and “gender” interchangeably—I still recall (running) race entry forms using one or the other and everyone seemed to know what was meant. However, while I eventually learned that the two are not the same—a person might have one biological sex and a different gender. While familiar with the science fiction idea of a multitude of genders, I eventually became aware that this was now a thing in the actual world.

Obviously, if gender is taken as the same as sex (which is set by anatomy or genetics), then gender would be an objective feature of reality and not subject to change merely by a change in labeling (or naming). However, gender has been largely (or even entirely) split from biological sex (anatomy or genetics) and is typically cast in terms of being a social construct. This view can be labeled as “gender nominalism.” By this I mean that gender is not an objective feature of reality, like anatomy, but a matter of naming, like being a Republican or Democrat.

Some thinkers have cast gender as being constructed by society as a whole, while others contend that individuals have lesser or greater ability to construct their own gender identities. People can place whatever gender label they wish upon themselves, but there is still the question of the role of others in that gender identity. The question is, then, to what degree can individuals construct their own gender identities? There is also the moral question about whether or not others are morally required to accept such gender self-identification. These matters are part of the broader challenge of identity in terms of who defines one’s identity (and what aspects) and to what degree are people morally obligated to accept these assignments (or declarations of identity).

My own view is to go with the obvious: people are free to self-declare whatever gender they wish, just as they are free to make any other claim of identity that is a social construct (which is a polite term for “made up”). So, a person could declare that he is a straight, Republican, Rotarian, fundamentalist, Christian, man. Another person could declare that she is a lesbian, Republican, Masonite, Jewish woman. And so on. But, of course, there is the matter of getting others to recognize that identity. For example, if a person identifies as a Republican, yet believes in climate change, argues for abortion rights, endorses same-sex marriage, supports Obama, favors tax increases, supports education spending, endorse the minimum wage, and is pro-environment, then other Republicans could rightly question the person’s Republican identity and claim that that person is a RINO (Republican in Name Only). As another example, a biological male could declare identity as a woman, yet still dress like a man, act like a man, date women, and exhibit no behavior that is associated with being a woman. In this case, other women might (rightly?) accuse her of being a WINO (Woman in Name Only).

In cases in which self-identification has no meaningful consequences for other people, it certainly makes sense for people to freely self-identify. In such cases, claiming to be F makes the person F, and what other people believe should have no impact on that person being F. That said, people might still dispute a person’s claim. For example, if someone self-identifies as a Trekkie, yet knows little about Star Trek, others might point out that this self-identification is in error. However, since this has no meaningful consequences, the person has every right to insist on being a Trekkie, though doing so might suggest that he is about as smart as a tribble.

In cases in which self-identification does have meaningful consequences for others, then there would seem to be moral grounds (based on the principle of harm) to allow restrictions on such self-identification. For example, if a relatively fast male runner wanted to self-identify as a woman so “she” could qualify for the Olympics, then it would seem reasonable to prevent that from happening. After all, “she” would bump a qualified (actual) woman off the team, which would be wrong. Because of the potential for such harms, it would be absurd to accept that everyone is obligated to accept the self-identification of others.

The flip side of this is that others should not have an automatic right to deny the self-identification of others. As a general rule, the principle of harm would seem to apply here as well—the others would have the right to impose in cases in which there is actual harm and the person would have the right to refuse the forced identity of others when doing so would inflict wrongful harm. The practical challenge is, clearly enough, working out the ethics of specific cases.


My Amazon Author Page

My Paizo Page

My DriveThru RPG Page

Follow Me on Twitter