How do I create a brief description of the social problem inclusion for individuals with developmental disabilities? Then, explain one ethical dilemm
How do I create a brief description of the social problem inclusion for individuals with developmental disabilities? Then, explain one ethical dilemma and one moral dilemma that could arise when solving the problem as well as the power differentials that may be present. Finally, explain how you would prevent the ethical and moral dilemmas and address the power differentials in your role as an advanced human services professional practitioner. Be sure to incorporate the Ethical Standards for Human Services Professionals into your explanation.
When working with people in your community to solve complex social problems, it is important to be prepared to prevent ethical and moral issues that may arise. An ethical issue is any issue that violates the Ethical Standards for Human Services Professionals. Examples include imposing your own values or biases into discussions about how to solve a problem, failing to incorporate multicultural considerations into the solution, and engaging in dual relationships with service users who are adversely affected by the problem. A moral issue is any issue that conflicts with your own personal values or beliefs. For example, if the social problem is a lack of access to quality healthcare in a rural community, some practitioners may have a moral issue with doctors who refuse to accept Medicaid.
Practitioners must also be prepared to address power differentials that may be present when addressing a social problem. Power differentials often occur when there is a concentration of power among people in formal, high-level positions (e.g., public officials, legislators, executive directors of non-profits) and a lack of power among those who are most adversely affected by the problem (e.g., community residents, service users, etc.).
Consider how to prevent ethical and moral issues that may arise when solving the social problem of inclusion for individuals with developmental disabilities. You will also examine how to address power differentials that may be present.
- Review the Learning Resources on power. When solving the social problem that you selected, reflect on who has the power to make decisions and influence the outcome and who lacks the power to do so. Then, consider how you could minimize those power differentials.
References:
Greene, J. & Haidt, J. (2002). How (and where) does moral judgement work?Links to an external site. Trends in Cognitive Sciences, 6(12), 517–523. https://go.openathens.net/redirector/waldenu.edu?url=https://doi.org/10.1016/S1364-6613(02)02011-9
National Community Development Institute. (n.d.). Sources of power.Links to an external site. http://www.buildingmovement.org/wp-content/uploads/2020/03/Sources_of_Power.pdf
National Organization for Human Services. (2015). Ethical standards for human services professionals.Links to an external site. https://www.nationalhumanservices.org/ethical-standards-for-hs-professionals
Stroh, D. P. (2015). Systems thinking for social change: A practical guide to solving complex problems, avoiding unintended consequences, and achieving lasting results. Chelsea Green Publishing.
- Chapter 3, “Telling Systems Stories” (pp. 29–44)
- Chapter 5, “An Overview of the Four-Stage Change Process” (pp. 71–78)
TRENDS in Cognitive Sciences Vol.6 No.12 December 2002
http://tics.trends.com 1364-6613/02/$ – see front matter © 2002 Elsevier Science Ltd. All rights reserved. PII: S1364-6613(02)02011-9
517Review
Joshua Greene
Dept of Psychology, Green Hall, Princeton University, Princeton, NJ 08544-1010, USA. e-mail: [email protected]
Jonathan Haidt
Dept of Psychology, University of Virginia, P.O. Box 400400, Charlottesville, VA 22904- 4400, USA. e-mail: [email protected]
Why do we care so strongly about what other people do, even when their actions won’t affect us? And how do we decide that someone else has done something wrong? These questions are at the heart of moral psychology, and psychologists’answers to these questions have changed with intellectual fashion. Historically, psychologists have disagreed about whether moral judgments are primarily products of emotional and non-rational processes (such as Freudian internalization or behaviorist reinforcement) or of reasoning and ‘higher’cognition (as in Piaget’s and Kohlberg’s post-conventional reasoning). Recently, however, findings from several areas of cognitive neuroscience have begun to converge on an answer: emotions and reasoning both matter, but automatic emotional processes tend to dominate.
Trends in moral psychology
During the cognitive revolution of the 1950s and 1960s, behaviorist and Freudian theories gave way to mental models and information processing as the preferred framework in psychology. In the moral domain, Lawrence Kohlberg was a part of this revolution. He built on the earlier work of Jean Piaget [1] to develop a six-stage model of the development of moral reasoning [2]. According to Kohlberg, moral growth is driven not by simple brain maturation but rather by experience in ‘role taking’, or looking at a problem from multiple perspectives. Role taking improves moral reasoning, and moral reasoning, Kohlberg thought, drives moral judgment.
But as the cognitive revolution matured in the 1980s, many researchers began calling for a complementary ‘affective revolution’. Kohlberg’s focus on moral reasoning seemed to ignore the importance of moral emotions. At the same time, new findings in evolutionary psychology [3,4] and primatology [5] began to point to the origins of human morality in a set of emotions (linked to expanding cognitive abilities) that make individuals care about the
welfare of others (e.g. kin altruism, including feelings of sympathy), and about cooperation, cheating, and norm-following (e.g. reciprocal altruism, including feelings of shame, gratitude and vengeance).
Integrating affect and reasoning
In the 1990s the affective revolution was reinforced by a new focus on ‘automaticity’– the mind’s ability to solve many problems, including high-level social ones, unconsciously and automatically [6]. A recent comprehensive model, the social intuitionist model [7], brings together research on automaticity with findings in neuroscience and theory in evolutionary psychology. This model suggests that moral judgment is much like aesthetic judgment: we see an action or hear a story and we have an instant feeling of approval or disapproval. These feelings are best thought of as affect-laden intuitions, as they appear suddenly and effortlessly in consciousness, with an affective valence (good or bad), but without any feeling of having gone through steps of searching, weighing evidence, or inferring a conclusion. These intuitions – for example, about reciprocity, loyalty, purity, suffering – are shaped by natural selection, as well as by cultural forces. People certainly do engage in moral reasoning, but, as suggested by studies of informal reasoning [8], these processes are typically one-sided efforts in support of pre-ordained conclusions. (As William James said, ‘A great many people think they are thinking when they are merely rearranging their prejudices.’) Moral reasoning matters, but it matters primarily in social contexts in which people try to influence each other and reach consensus with friends and allies.
This emphasis on quick, automatic affective reactions is supported by recent findings in social psychology, such as: (1) that people evaluate others and apply morally laden stereotypes automatically [9]; (2) that motivations to maintain relationships and defend against threatening ideas bias judgments and motivate subsequent reasoning [10,11]; and (3) that people can very easily construct post-hoc reasons to justify their actions and judgments [12–14].
Somatic markers and decision-making
In keeping with this affective trend, Antonio Damasio and colleagues have generated widespread interest in the affective neural bases of social judgment through their ongoing study of patients with damage to the ventral and medial portions of the frontal lobes [15,16]. To varying degrees, these patients resemble
Moral psychology has long focused on reasoning, but recent evidence suggests
that moral judgment is more a matter of emotion and affective intuition than
deliberate reasoning. Here we discuss recent findings in psychology and
cognitive neuroscience, including several studies that specifically investigate
moral judgment. These findings indicate the importance of affect, although they
allow that reasoning can play a restricted but significant role in moral judgment.
They also point towards a preliminary account of the functional neuroanatomy
of moral judgment, according to which many brain areas make important
contributions to moral judgment although none is devoted specifically to it.
How (and where) does moral
judgment work?
Joshua Greene and Jonathan Haidt
Phineas Gage, the 19th century railroad foreman who made neurological history after an accidental explosion sent a tamping iron through his medial prefrontal cortex, robbing him of his sound judgment and remarkably little else [15,16]. Damasio and colleagues argue that contemporary patients with Gage-like damage (such as patient EVR [17]), have emotional deficits and, more specifically, an inability to generate and effectively use ‘somatic markers’, neural representations of body states that imbue behavioral options with affective significance and thus guide on-line decision-making. These patients’deficits are revealed in their abnormal skin-conductance responses (an index of autonomic arousal) and poor performance on the Iowa Gambling Task, which simulates real-life decision-making [17]. These deficits exist to a surprising extent against a background of preserved ‘cognitive function’as indexed by IQ tests and other measures. Moreover, such patients exhibit preserved abstract social knowledge in spite of their disastrous real-life judgment. Their affective deficits render them unable to feel their way through life, which suggests that normal decision-making is more emotional and less reasoned than many have believed [15].
Frontal damage and anti-social behavior
Frontal patients like EVR (see above) are more likely to hurt themselves than other people, but a recent study by Anderson et al. of two patients with prefrontal damage acquired during early childhood reports behavior that is more flagrantly immoral [18,19]. These patients lie, steal, have neglected their children, and at times have been physically aggressive – all without apparent remorse. Both patients perform reasonably well on IQ tests and other standard cognitive measures and perform poorly on the Iowa Gambling Task, but unlike adult-onset patients their knowledge of social/moral norms is deficient. Their moral reasoning appears to be, in the terminology of Kohlberg, ‘preconventional’, conducted from an egocentric perspective in which the purpose is to avoid punishment. Other tests show that they have a limited understanding of the social and emotional implications of decisions and fail to identify primary issues and generate appropriate responses to hypothetical social situations. Thus, it appears that the brain regions compromised in these patients (ventral, medial, and polar aspects of the prefrontal cortex) include structures that are crucial not only for on-line decision- making but also for the acquisition of social knowledge and dispositions towards normal social behavior.
Other studies have documented impaired social behavior resulting from frontal dysfunction [20]. In a study of 279 Vietnam War veterans, Grafman and colleagues found that patients with ventromedial frontal damage tend towards violence and aggression [21]. Raine and colleagues have found that individuals diagnosed with anti-social personality disorder have reduced prefrontal gray matter and exhibit reduced autonomic responses to the performance of a socially
stressful task [22], and a recent popular account documenting a large number of case studies attributes violent criminal behavior to a combination of childhood abuse and frontal damage [23].
Psychopaths exhibit extreme anti-social behavior in the absence of observed brain damage. However, a recent neuroimaging study has shown that psychopaths exhibit less emotion-related neural activity than control subjects while responding to emotionally valenced words [24]. Blair and others have conducted several studies that characterize psychopathy as an affective disorder involving a reduction in empathy [25] and consequent deficits in moral judgment both in and out of the laboratory [26]. Blair observes that psychopaths, unlike frontal patients, are marked by their tendency towards instrumental rather than reactive aggression [27].
Neuroimaging
Responses to moral sentences and pictures A handful of recent studies have used functional neuroimaging to study moral psychology. In an fMRI study, Jorge Moll and colleagues [28] presented subjects with simple claims, some with moral content (‘They hung an innocent’) and others without moral content (‘Stones are made of water’). Judgments in response to claims with moral content produced increased activity bilaterally in the frontal pole, in the medial frontal gyrus, right cerebellum, right temporal pole, superior temporal sulcus (STS), left orbitofrontal cortex (OFC), left precuneus, and the posterior globus pallidus. A more recent study by Moll and colleagues [29] compared judgments in response to simple moral claims with judgments in response to unpleasantly valenced non-moral claims with social content, many of which evoke disgust (e.g. ‘He licked the dirty toilet’, ‘Pregnant women often throw up’). A direct comparison of these two conditions revealed greater activity in the left medial OFC for the moral condition and greater activation in the left lateral OFC as well as the left amygdala for the non-moral/social condition. These results suggest a functional dissociation between neural networks within the OFC and associated structures that specialize in processing different kinds of social/emotional information relevant (in varying degrees) to moral judgment. A third study by Moll and colleagues [30] found similar neural responses to pictures with moral content (e.g. physical assaults, poor abandoned children). The medial frontal and posterior cingulate regions were also activated in an fMRI study of empathy and forgiveness [31]. (See also Table 1.)
Emotional engagement in ‘personal’ versus ‘impersonal’ moral judgments Whereas Moll and colleagues have investigated moral cognition by distinguishing the effects of moral versus non-moral phenomena, Greene and colleagues [32] have drawn a distinction within the moral domain between ‘personal’and ‘impersonal’moral judgments (see Box 1). Greene and colleagues scanned subjects
TRENDS in Cognitive Sciences Vol.6 No.12 December 2002
http://tics.trends.com
518 Review
using fMRI while they responded to a series of personal and impersonal moral dilemmas as well as non-moral dilemmas, all of which involved complex narratives. They found that responding to personal moral dilemmas, as compared with impersonal and non-moral dilemmas, produced increased activity in areas associated with social/emotional processing: medial frontal gyrus, posterior cingulate gyrus, and bilateral STS (originally labeled ‘angular gyrus’). By contrast, impersonal and non-moral dilemmas as compared with personal dilemmas produced increased activity in areas associated with working memory: dorsolateral prefrontal and parietal areas (see Fig. 1). They found comparatively little difference between the impersonal-moral and non-moral conditions, suggesting that impersonal moral judgment has less in common with personal moral judgment than with certain kinds of non-moral practical judgment.
Greene et al. carried out an analysis of subjects’ reaction times to link these imaging data to behavior. Subjects were slow to approve of personal violations but relatively quick to condemn them. By contrast, approvals and disapprovals took equally long for impersonal moral and non-moral judgments. This pattern is explained by subjects’having to overcome their negative emotional responses when approving of personal moral violations as compared with other, less emotionally charged actions (this might be likened to the pattern of interference observed in the Stroop task).
The neuroanatomy of moral judgment
The functional neuroimaging boom has provided a wealth of information about the neuroanatomy of emotion, social cognition, and other neural processes. These data, combined with the lesion and pathology data above, allow us to interpret the results of the imaging studies described in the previous section and thus broaden and refine our understanding of the ‘moral brain’.
TRENDS in Cognitive Sciences Vol.6 No.12 December 2002
http://tics.trends.com
519Review
Suppose a runaway trolley is about to run over and kill five people. Suppose further that you can hit a switch that will divert the trolley onto a different set of tracks where it will kill only one person instead of five. Is it okay to hit the switch? Now, what if the only way to save the five people were to push a large person (larger than yourself) in front of the trolley, killing him but saving the others? Would that be okay?
Most people say ‘yes’ to the first case and ‘no’ to the second in spite of the fact that these cases are so similar [a]. Although it is easy to generate (and surprisingly difficult to defend) hypotheses about why one ought to treat these cases differently [b], Greene et al. have attempted to explain how people do in fact arrive at this puzzling pair of conclusions [a]. To explain the difference, they posit a distinction between what they believe are two fundamentally different kinds of moral thinking, drawing on capacities that emerged at different stages of human evolution. On the one hand, moral thinking is driven largely by social-emotional dispositions built on those we inherited from our primate ancestors [c,d]. At the same time, humans have a unique capacity for sophisticated abstract reasoning that can be applied to any subject matter. One might suppose, then, that human moral thinking is not one kind of process, but rather a complex interplay between (at least) two distinct types of processes: domain-specific, social-emotional responses and domain-neutral reasoning processes applied in moral contexts.
With this in mind, Greene and colleagues distinguished between ‘personal’ and ‘impersonal’ moral violations and judgments. A moral violation is personal if it is: (i) likely to cause serious bodily harm, (ii) to a particular person, (iii) in such a way that the harm does not result from the deflection of an existing threat onto a different party. A moral violation is impersonal if it fails to meet these criteria. One can think of these criteria for personal harm in terms of ‘ME HURT YOU’, and as delineating roughly those violations that a chimpanzee can appreciate. The ‘HURT’ condition picks out roughly the kinds of harm that a chimp can understand (e.g. assault, as opposed to, say, tax evasion). The ‘YOU’ condition requires that the victim be vividly represented as an individual. Finally, the ‘ME’ condition captures the notion of ‘agency’ [b], the idea that the action must spring in a vivid way from an agent’s will, must be ‘authored’ rather than merely ‘edited’ by an agent.
Pushing someone in front of a trolley meets all three criteria and is therefore personal, whereas diverting a trolley involves merely deflecting an existing threat, removing the crucial sense of agency and therefore making this violation impersonal.
References
a Greene, J.D. et al. (2001) An fMRI investigation of emotional engagement in moral judgment. Science 293, 2105–2108
b Thomson, J.J. (1986) Rights, Restitution, and Risk: Essays in Moral Theory, Harvard University Press
c Flack, J.C. and de Waal, F.B.M. (2000) ‘Any animal whatever’: Darwinian building blocks of morality in monkeys and apes. In Evolutionary Origins of Morality (Katz, L.D., ed.), pp. 1–29, Imprint Academic
d Haidt, J. (2001) The emotional dog and its rational tail: a social intuitionist approach to moral judgment. Psychol. Rev. 108, 814–834
Box 1. Two kinds of moral thinking: personal and impersonal
Fig. 1. Brain areas (indicated by Brodmann’s area (BA)) exhibiting differences in activity in response to personal moral dilemmas as compared with impersonal and non-moral dilemmas [32]. Areas exhibiting greater activity for personal moral dilemmas (as compared with impersonal and non-moral): medial frontal gyrus (BA 9/10); posterior cingulate gyrus (BA 31); superior temporal sulcus,
inferior parietal lobe (BA 39). Areas exhibiting greater activity for impersonal moral dilemmas (as compared with personal): dorsolateral prefrontal cortex (BA 46); parietal lobe (BA 7/40). Images are reversed left to right according to radiologic convention. (Reprinted with permission from Greene et al. [32]. Copyright 2001 American Association for the Advancement of Science.)
The ventral and medial prefrontal cortices Table 1 and Fig. 2 summarize results relevant to the neuroanatomy of moral judgment. One brain area of great interest is the medial frontal gyrus, around the border of Brodmann areas (BA) 9 and 10, which
probably serves in the integration of emotion into decision-making and planning [33,34] and might also play a role in theory of mind (ToM, the capacity to represent others’mental states) and other specifically social functions relevant to moral judgment [35,36]. This
TRENDS in Cognitive Sciences Vol.6 No.12 December 2002
http://tics.trends.com
520 Review
Table 1. The moral brain (see also Fig. 2, color-coded to this figure.) The first column lists eight brain areas (Brodmann’s areas in
parentheses) implicated in moral cognition by neuroimaging studies. Subsequent columns provide additional information about
their functions.
Brain region
(with BA)
Associated moral tasks Other associated tasks Social pathology from
damage
Likely functions
1. Medial Personal moral Attributing intentionality to moving Poor practical Integration of emotion frontal gyrus judgments shapes and cartoon characters* judgment [15,16], into decision-making (BA 9/10) Impersonal moral Theory of mind (ToM) stories and cartoons* Reactive aggression and planning [15,16],
judgments Representing a historical figure's [27] and (primarily in esp. for conscious (relative to non- mental states* [36] developmental processes [33] moral) [32] Viewing angry/sad faces [47] cases) diminished ToM [36] Simple moral pleasant pictures, negative pictures empathy and social judgments* [28] [48] (with emotional report [49]) knowledge [18] Viewing moral Reward [37] pictures [30] Viewing and/or recall of happy, sad, and Forgivability disgusting films [50] judgments [31]* Emotional autobiographical recall [51] (*also lateral Emotional planning [34] frontopolar) ‘Rest’ [42]
*(focus in the paracingulate sulcus)
2. Posterior Personal moral Hearing affective autobiographical episodes Impaired recognition Integration of emotion, cingulate, judgments [52], threat words [38] memory for faces imagery (esp. precuneus precuneus, Impersonal moral Reading coherent stories, esp. ToM stories [53] Capgras delusion? [55] [39]), and memory [38], retrosplenial judgments (relative to Viewing ToM cartoons [54], familiar faces [55], esp. for coherent social cortex (BA 31/7) non-moral) [32] disgusted faces, sad faces, snake video, narratives
Simple moral judgments previously experienced robbery video, [28] combat pictures (and imagery) Forgivability judgments Sad autobiographical recall (men) [38] [31] Recognizing neutral words from negative Moral pictures [30] context [56]
Emotional planning [34] Recall of happy personal life episodes [57], imaginable word pairs [39] ‘Rest’ [42]
3. Superior Personal moral Viewing biological motion (hands, faces, eyes, Impaired judgment from Supporting representations temporal sulcus, judgments [32] body) [40]; sad faces [47]; happy, sad, and eye gaze (monkeys) [40] of socially significant inferior parietal Simple moral judgments disgusting films [50,51]; ToM cartoons, Capgras delusion? [41] movements [40], and lobe [28,29] reading coherent stories with self- possibly complex (BA 39) Moral pictures [30] perspective and with characters , esp ToM representations of
attributing intentionality to moving shapes ‘personhood’ [41] Representing a historical figure's mental ToM [36] states [36] Recognizing neutral words from negative context [56] Recall of imaginable word pairs [39] Judgment of indoor/outdoor vs. subjective response to (un)pleasant pictures [49] Emotional film viewing vs. recall [51] ‘Rest’ [42]
4. Orbitofrontal/ Simple moral judgments Reward/punishment [37] Poor practical judgment Representation of ventromedial [28,29] Sad autobiographical recall [57] [15,16] reward/punishment value frontal cortex Moral pictures [30] Recognizing words from positive context [56] Reactive aggression [27] [15,16,37] (BA 10/11) Viewing angry faces [47] and (primarily in control of inappropriate/
‘Rest’ [42] developmental cases) disadvantageous (Note: absent in many PET studies diminished empathy and behavior [15,27] of emotion [34,48–50]) social knowledge [18] ‘hot’ ToM [58]
Difficulty with advanced ToM tasks [58]
area should be distinguished from other medial frontal areas, such as the anterior cingulate cortex and a more superior medial frontal region (BA 8/9) that has been implicated in ToM and self-referential tasks [36], but not (yet) in specifically moral tasks. This region should also be distinguished from orbitofrontal/ventromedial areas that appear to be primarily involved in the on-line representation of reward and punishment value [15,17,37]. Although undoubtedly crucial to social behavior and development, the contributions made by these areas may not be specifically social (but see [29]). Rather these regions might perform a more general regulative function in which affective information guides approach and avoidance behavior in both social and non-social contexts. It is worth noting that many of the lesion patients who exhibit acquired sociopathic behavior, such as Gage [16] and EVR [17], have medial damage extending dorsally into BA 9/10. It is also worth noting that many studies that use complex emotional stimuli and PET (which, unlike fMRI, is well-suited to imaging ventral frontal regions) find no activation in orbitofrontal/ventromedial regions [33].
The posterior cingulate and STS Another region implicated in moral judgment is the posterior cingulate/retrosplenial cortex region centered around BA 31 and the neighboring precuneus area
TRENDS in Cognitive Sciences Vol.6 No.12 December 2002
http://tics.trends.com
521Review
Table 1. continued
Brain region
(with BA)
Associated moral tasks Other associated tasks Social pathology from
damage
Likely functions
5. Temporal pole Simple moral judgments Reading coherent stories (with characters), Impaired autobiographical Imparting affective tone to (BA 38) [28,29] ToM stories memory [59] experience and memory
Attributing intentionality to moving shapes [59] and cartoon characters, ToM [36] Representing a historical figure's mental states Recall of familiar faces and scenes [36] Hearing affective autobiographical episodes [52] Recognition of emotional pictures [59] Viewing emotional pictures (with subjective report) [49], angry/sad faces [47] Viewing and recall of happy, sad (viewing only), and disgusting film [50] Emotional autobiographical recall [51]
6. Amygdala Moral pictures [30] Recognition of emotional pictures [59] Poor social judgment from Rapid assessment of Viewing emotional film [51], sad faces [47] faces and movement [43] reward/punishment Viewing racial outgroup faces [44,45] value, esp. visual
and negative [43]
7. Dorsolateral Impersonal moral Working memory and other ‘cognitive’ tasks Working memory and prefrontal cortex judgment [32] [46] other ‘cognitive’ (BA 9/10/46); functions [46] 8. Parietal lobe (BA 7/40)
Fig. 2. The moral brain (see also Table 1, color-coded to this figure.) Brain areas implicated in moral cognition by neuroimaging studies (Brodmann’s areas in parentheses): 1. medial frontal gyrus (9/10); 2. posterior cingulate, precuneus, retrosplenial cortex (31/7); 3. superior temporal sulcus, inferior parietal lobe (39); 4. orbitofrontal, ventromedial frontal cortex (10/11); 5. temporal pole (38); 6. Amygdala; 7. dorsolateral prefrontal cortex (9/10/46); 8. parietal lobe (7/40). (Adapted with permission from Adolphs (in press) [60].)
extending caudally and dorsally into BA 7. The posterior cingulate/retrosplenial cortex is one of the most commonly activated areas in neuroimaging studies of emotion [38] and one of the areas exhibiting decreased activity in psychopaths during the processing of emotionally valenced words [24]. It appears to be involved in memory and processes involving imagery, especially affective imagery in the context of coherent narrative. The precuneus also appears to be a crucial area for imagery [39].
The posterior STS/inferior parietal region has been an area of intense interest in the emerging field of social cognitive neuroscience for some time. Its primary function appears to be the perception and representation of socially significant information, particularly from biological motion cues, which are crucial for making inferences about the beliefs and intentions of others [36,40]. Brothers and Ring [41] have suggested that more basic social representations of movement (voices, faces, etc.) might serve as building blocks for more complex representations of ‘persons’. They note the connection between temporal lobe damage and Capgras delusion, in which patients report that loved ones have been replaced by identical looking and sounding imposters. To the Capgras patient, the faces and voices of loved ones are immediately recognizable, but somehow unfamiliar. What is missing, it seems, is the activation of a high-level, affectively significant representation of the person in question. The suggestion that the posterior STS/inferior parietal region supports representations of ‘personhood’dovetails nicely with the finding that this region responds strongly to personal moral dilemmas, and no more strongly to impersonal moral dilemmas than to non-moral dilemmas. (In other words, this region may be responsible for putting the ‘YOU’ in ‘ME HURT YOU’– see Box 1).
The moral brain and the resting brain In a remarkable convergence, the three brain regions highlighted above coincide with regions identified in a meta-analysis of the resting brain’s activity [42]. These regions appear to be components in a tonically active neural system, the activity of which is attenuated when people are engaged in ‘goal-directed actions’. Gusnard and Raichle propose that this system’s function is the ongoing evaluation of the environment and one’s relation to it [42]. We tentatively propose that the key element behind this convergence is introspection; the high-level social-emotional processing involved in moral judgment may be a ‘turbo-charged’version of the
personal ruminations in which we engage when otherwise unengaged.
The amygdala and dorsolateral ‘cognitive’ areas The amygdala plays a crucial role in social-emotional processing [43] and is known to exhibit increased activity in response to racial outgroup faces [44,45] and moral pictures [30]. It appears, however, to be relatively stimulus-driven and biased towards the visual [43], suggesting that its influence on moral judgment is likely to be rather crude. Less crude influences might come from classically ‘cognitive’areas in the dorsolateral prefrontal cortex and parietal lobes [46], which have been implicated in impersonal moral judgment [32]. These activations may represent the application of domain-neutral reasoning to moral judgment (see Box 1).
Conclusions
Neuroima
Collepals.com Plagiarism Free Papers
Are you looking for custom essay writing service or even dissertation writing services? Just request for our write my paper service, and we'll match you with the best essay writer in your subject! With an exceptional team of professional academic experts in a wide range of subjects, we can guarantee you an unrivaled quality of custom-written papers.
Get ZERO PLAGIARISM, HUMAN WRITTEN ESSAYS
Why Hire Collepals.com writers to do your paper?
Quality- We are experienced and have access to ample research materials.
We write plagiarism Free Content
Confidential- We never share or sell your personal information to third parties.
Support-Chat with us today! We are always waiting to answer all your questions.