chapter
stringlengths 1.97k
1.53M
| path
stringlengths 47
241
|
---|---|
Thus far, we have described the main characteristics of neurons, including how their processes come in close contact with one another to form synapses. In this section, we consider the conduction of communication within a neuron and how this signal is transmitted to the next neuron. There are two stages of this electrochemical action in neurons. The first stage is the electrical conduction of dendritic input to the initiation of an action potential within a neuron. The second stage is a chemical transmission across the synaptic gap between the presynaptic neuron and the postsynaptic neuron of the synapse. To understand these processes, we first need to consider what occurs within a neuron when it is at a steady state, called resting membrane potential.
Resting Membrane Potential
The intracellular (inside the cell) fluid and extracellular (out- side the cell) fluid of neurons is composed of a combination of ions (electrically charged molecules; see Figure \(1\)). Cations are positively charged ions, and anions are negatively charged ions. The composition of intracellular and extracellular fluid is similar to salt water, containing sodium (Na+), potassium (K+), chloride (Cl−), and anions (A−).
The cell membrane, which is composed of a lipid bilayer of fat molecules, separates the cell from the surrounding extracellular fluid. There are proteins that span the membrane, forming ion channels that allow particular ions to pass between the intracellular and extracellular fluid (see Figure \(1\)). These ions are in different concentrations inside the cell relative to outside the cell, and the ions have different electrical charges. Due to this difference in concentration and charge, two forces act to maintain a steady state when the cell is at rest: diffusion and electrostatic pressure. Diffusion is the force on molecules to move from areas of high concentration to areas of low concentration. Electrostatic pressure is the force on two ions with similar charge to repel each other and the force of two ions with opposite charge to attract to one another. Remember the saying, opposites attract?
Regardless of the ion, there exists a membrane potential at which the force of diffusion is equal and opposite of the force of electrostatic pressure. This voltage, called the equilibrium potential, is the voltage at which no ions flow. Since there are several ions that can permeate the cell’s membrane, the base- line electrical charge inside the cell compared with outside the cell, referred to as resting membrane potential, is based on the collective drive of force on several ions. Relative to the extracellular fluid, the membrane potential of a neuron at rest is negatively charged at approximately −70 mV (see Figure \(2\)). These are very small voltages compared with the voltages of batteries and electrical outlets, which we encounter daily, that range from 1.5 to 240 V.
Let us see how these two forces, diffusion and electrostatic pressure, act on the four groups of ions mentioned above.
1. Anions (A): Anions are highly concentrated inside the cell and contribute to the negative charge of the resting membrane potential. Diffusion and electrostatic pressure are not forces that determine A− concentration because A− is impermeable to the cell membrane. There are no ion channels that allow for A− to move between the intracellular and extracellular fluid.
2. Potassium(K+):The cell membrane is very permeable to potassium at rest, but potassium remains in high concentrations inside the cell. Diffusion pushes K+ outside the cell because it is in high concentration inside the cell. However, electrostatic pressure pushes K+ inside the cell because the positive charge of K+ is attracted to the negative charge inside the cell. In combination, these forces oppose one another with respect to K+.
3. Chloride (Cl): The cell membrane is also very permeable to chloride at rest, but chloride remains in high concentration outside the cell. Diffusion pushes Cl− inside the cell because it is in high concentration outside the cell. However, electrostatic pressure pushes Cl− outside the cell because the negative charge of Cl− is attracted to the positive charge outside the cell. Similar to K+, these forces oppose one another with respect to Cl−.
4. Sodium (Na+): The cell membrane is not very permeable to sodium at rest. Diffusion pushes Na+ inside the cell because it is in high concentration outside the cell. Electrostatic pressure also pushes Na+ inside the cell because the positive charge of Na+ is attracted to the negative charge inside the cell. Both of these forces push Na+ inside the cell; however, Na+ cannot permeate the cell membrane and remains in high concentration outside the cell. The small amounts of Na+ inside the cell are removed by a sodium-potassium pump, which uses the neuron’s energy (adenosine tri- phosphate, ATP) to pump 3 Na+ ions out of the cell in exchange for bringing 2 K+ ions inside the cell.
Action Potential
Now that we have considered what occurs in a neuron at rest, let us consider what changes occur to the resting membrane potential when a neuron receives input, or information, from the presynaptic terminal button of another neuron. Our understanding of the electrical signals or potentials that occur within a neuron results from the seminal work of Hodgkin and Huxley that began in the 1930s at a well-known marine biology lab in Woodshole, Massachusetts. Their work, for which they won the Nobel Prize in Medicine in 1963, has resulted in the general model of electrochemical transduction that is described here (Hodgkin & Huxley, 1952). Hodgkin and Huxley studied a very large axon in the squid, a common species for that region of the United States. The giant axon of the squid is roughly 100 times larger than that of axons in the mammalian brain, making it much easier to see. Activation of the giant axon is responsible for a withdrawal response the squid uses when trying to escape from a predator, such as large fish, birds, sharks, and even humans. When was the last time you had calamari? The large axon size is no mistake in nature’s design; it allows for very rapid transmission of an electrical signal, enabling a swift escape motion in the squid from its predators.
While studying this species, Hodgkin and Huxley noticed that if they applied an electrical stimulus to the axon, a large, transient electrical current conducted down the axon. This transient electrical current is known as an action potential (see Figure \(2\)). An action potential is an all-or-nothing response that occurs when there is a change in the charge or potential of the cell from its resting membrane potential (−70 mV) in a more positive direction, which is a depolarization (see Figure \(2\)). What is meant by an all-or-nothing response? I find that this concept is best compared to the binary code used in computers, where there are only two possibilities, 0 or 1. There is no halfway or in-between these possible values; for example, 0.5 does not exist in binary code. There are only two possibilities, either the value of 0 or the value of 1. The action potential is the same in this respect. There is no halfway; it occurs, or it does not occur. There is a specific membrane potential that the neuron must reach to initiate an action potential. This membrane potential, called the threshold of excitation, is typically around −50 mV. If the threshold of excitation is reached, then an action potential is triggered.
How is an action potential initiated? At any one time, each neuron is receiving hundreds of inputs from the cells that synapse with it. These inputs can cause several types of fluctuations in the neuron’s membrane potentials (see Figure \(2\)):
1. Excitatory post synaptic potentials(EPSPs): a depolarizing current that causes the membrane potential to become more positive and closer to the threshold of excitation; or
2. Inhibitory post synaptic potentials (IPSPs): a hyper- polarizing current that causes the membrane potential to become more negative and further away from the threshold of excitation.
These post synaptic potentials, EPSPs and IPSPs, summate or add together in time and space. The IPSPs make the membrane potential more negative, but how much so depends on the strength of the IPSPs. The EPSPs make the membrane potential more positive; again, how much more positive depends on the strength of the EPSPs. If you have two small EPSPs at the same time and the same synapse then the result will be a large EPSP. If you have a small EPSP and a small IPSP at the same time and the same synapse then they will cancel each other out. Unlike the action potential, which is an all- or-nothing response, IPSPs and EPSPs are smaller and graded potentials, varying in strength. The change in voltage during an action potential is approximately 100 mV. In comparison, EPSPs and IPSPs are changes in voltage between 0.1 to 40 mV. They can be different strengths, or gradients, and they are measured by how far the membrane potentials diverge from the resting membrane potential.
I know the concept of summation can be confusing. As a child, I use to play a game in elementary school with a very large parachute where you would try to knock balls out of the center of the parachute. This game illustrates the properties of summation rather well. In this game, a group of children next to one another would work in unison to produce waves in the parachute in order to cause a wave large enough to knock the ball out of the parachute. The children would initiate the waves at the same time and in the same direction. The additive result was a larger wave in the parachute, and the balls would bounce out of the parachute. However, if the waves they initiated occurred in the opposite direction or with the wrong timing, the waves would cancel each other out, and the balls would remain in the center of the parachute. EPSPs or IPSPs in a neuron work in the same fashion to the properties of the waves in the parachute; they either add or cancel each other out. If you have two EPSPs, then they sum together and become a larger depolarization. Similarly, if two IPSPs come into the cell at the same time, they will sum and become a larger hyper polarization in membrane potential. However, if two inputs were opposing one another, moving the potential in opposite directions, such as an EPSP and an IPSP, their sum would cancel each other out.
At any moment in time, each cell is receiving mixed messages, both EPSPs and IPSPs. If the summation of EPSPs is strong enough to depolarize the membrane potential to reach the threshold of excitation, then it initiates an action potential. The action potential then travels down the axon, away from the soma, until it reaches the ends of the axon (the terminal button). In the terminal button, the action potential triggers the release of neurotransmitters from the presynaptic terminal button into the synaptic gap. These neurotransmitters, in turn, cause EPSPs and IPSPs in the post synaptic dendritic spines of the next cell (see Figure \(1\) and Figure \(3\)). The neurotransmitter released from the presynaptic terminal button binds with ionotropic receptors in a lock-and- key fashion on the post-synaptic dendritic spine. Ionotropic receptors are receptors on ion channels that open, allowing some ions to enter or exit the cell, depending upon the presence of a particular neurotransmitter. The type of neurotransmitter and the permeability of the ion channel it activates will determine if an EPSP or IPSP occurs in the dendrite of the post-synaptic cell. These EPSPs and IPSPs summate in the same fashion described above, and the entire process occurs again in another cell.
The Change in Membrane Potential During an Action Potential
We discussed previously which ions are involved in maintaining the resting membrane potential. Not surprisingly, some of these same ions are involved in the action potential. When the cell becomes depolarized (more positively charged) and reaches the threshold of excitation, this causes a voltage- dependent Na+ channel to open. A voltage-dependent ion channel is a channel that opens, allowing some ions to enter or exit the cell, depending upon when the cell reaches a particular membrane potential. When the cell is at resting membrane potential, these voltage-dependent Na+ channels are closed. As we learned earlier, both diffusion and electrostatic pressure are pushing Na+ inside the cells. However, Na+ can- not permeate the membrane when the cell is at rest. Now that these channels are open, Na+ rushes inside the cell, causing the cell to become very positively charged relative to the out- side of the cell. This is responsible for the rising or depolarizing phase of the action potential (see Figure \(2\)). The inside of the cell becomes very positively charged, +40 mV. At this point, the Na+ channels close and become refractory. This means the Na+ channels cannot reopen again until after the cell returns to the resting membrane potential. Thus, a new action potential cannot occur during the refractory period. The refractory period also ensures the action potential can only move in one direction down the axon, away from the soma. As the cell becomes more depolarized, a second type of voltage-dependent channel opens; this channel is permeable to K+. With the cell very positive relative to the outside of the cell (depolarized) and the high concentration of K+ within the cell, both the force of diffusion and the force of electrostatic pressure drive K+ outside of the cell. The movement of K+ out of the cell causes the cell potential to return back to the resting membrane potential, the falling or hyperpolarizing phase of the action potential (see Figure \(2\)). A short hyperpolarization occurs partially due to the gradual closing of the K+ channels. With the Na+ closed, electrostatic pressure continues to push K+ out of the cell. In addition, the sodium-potassium pump is pushing Na+ out of the cell. The cell returns to the resting membrane potential, and the excess extracellular K+ diffuses away. This exchange of Na+ and K+ ions happens very rapidly, in less than 1 msec. The action potential occurs in a wave-like motion down the axon until it reaches the terminal button. Only the ion channels in very close proximity to the action potential are affected.
Earlier you learned that axons are covered in myelin. Let us consider how myelin speeds up the process of the action potential. There are gaps in the myelin sheaths called nodes of Ranvier. The myelin insulates the axon and does not allow any fluid to exist between the myelin and cell membrane. Under the myelin, when the Na+ and K+ channels open, no ions flow between the intracellular and extracellular fluid. This saves the cell from having to expend the energy necessary to rectify or regain the resting membrane potential. (Remember, the pumps need ATP to run.) Under the myelin, the action potential degrades some, but is still large enough in potential to trigger a new action potential at the next node of Ranvier. Thus, the action potential actively jumps from node to node; this process is known as saltatory conduction.
In the presynaptic terminal button, the action potential triggers the release of neurotransmitters (see Figure 6.2.2). Neurotransmitters cross the synaptic gap and open sub- types of receptors in a lock-and-key fashion (see Figure 6.2.2). Depending on the type of neurotransmitter, an EPSP or IPSP occurs in the dendrite of the post synaptic cell. Neurotransmitters that open Na+ or calcium (Ca+) channels cause an EPSP; an example is the NMDA receptors, which are activated by glutamate (the main excitatory neurotransmitter in the brain). In contrast, neurotransmitters that open Cl− or K+ channels cause an IPSP; an example is gamma-amino butryric acid (GABA) receptors, which are activated by GABA, the main inhibitory neurotransmitter in the brain. Once the EPSPs and IPSPs occur in the postsynaptic site, the process of communication within and between neurons cycles on (see Figure \(3\)). A neurotransmitter that does not bind to recep- tors is broken down and inactivated by enzymes or glial cells, or it is taken back into the presynaptic terminal button in a process called reuptake, which will be discussed further in the module on psychopharmacology.
REFERENCES
De Carlos, J. A., & Borrell, J. (2007). A historical reflection of the contributions of Cajal and Golgi to the foundations of neuroscience. Brain Research Reviews, 55(1), 8–16. doi.org/10.1016/j. brainresrev.2007.03.010
Furtak, S. C., Moyer, J. R., Jr., & Brown, T. H. (2007). Morphology and ontogeny of rat perirhinal cortical neurons. Journal of Comparative Neurology, 505(5), 493–510. doi.org/10.1002/cne.21516
Grant, G. (2007). How the 1906 Nobel Prize in Physiology or Medicine was shared between Golgi and Cajal. Brain Research Reviews, 55(2), 490–498. https://doi.org/10.1016/j.brainresrev.2006.11.004
Hodgkin, A. L., & Huxley, A. F. (1952). A quantitative description of membrane current and its application to conduction and excitation in nerve. Journal of Physiology, 117(4), 500–544. doi.org/10.1113/ jphysiol.1952.sp004764
López-Muñoz, F., Boya, J., & Alamo, C. (2006). Neuron theory, the cornerstone of neuroscience, on the centenary of the Nobel Prize award to Santiago Ramón y Cajal. Brain Research Bulletin, 70(4–6), 391–405. https://doi.org/10.1016/j.brainresbull.2006.07.010
Pasternak, J. F., & Woolsey, T. A. (1975). On the “selectivity” of the Golgi-Cox method. Journal of Comparative Neurology, 160(3), 307–312. doi.org/10.1002/cne.901600304
Ramón y Cajal, S. (1911). Histology of the nervous system of man and vertebrates. Oxford University Press.
Simons, M., & Trotter, J. (2007). Wrapping it up: The cell biology of myelination. Current Opinion in Neurobiology, 17(5), 533–540. https:// doi.org/10.1016/j.conb.2007.08.003
Smit, G. J., & Colon, E. J. (1969). Quantitative analysis of the cerebral cortex. I. Aselectivity of the Golgi-Cox staining technique. Brain Research, 13(3), 485–510. doi.org/10.1016/ 0006-8993(69)90262-5
Williams, R. W., & Herrup, K. (1988). The control of neuron number. Annual Review of Neuroscience, 11, 423–453. doi.org/10.1146/ annurev.ne.11.030188.002231 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/06%3A_Neurons/6.03%3A_Communication_Within_And_Between_Neurons.txt |
If you were someone who understood brain anatomy and were to look at the brain of an animal that you had never seen before, you would nevertheless be able to deduce the likely capacities of the animal. This is because the brains of all animals are very similar in overall form. In each animal the brain is layered, and the basic structures of the brain are similar (see Figure \(1\)). The innermost structures of the brain— the parts nearest the spinal cord—are the oldest part of the brain, and these areas carry out the same the functions they did for our distant ancestors. The “old brain” regulates basic survival functions, such as breathing, moving, resting, and feeding, and creates our experiences of emotion. Mammals, including humans, have developed further brain layers that provide more advanced functions—for instance, better memory, more sophisticated social interactions, and the ability to experience emotions. Humans have a very large and highly developed outer layer known as the cerebral cortex (see Figure \(2\)), which makes us particularly adept at these processes.
7.02: The Old Brain- Wired For Survival
The brain stem is the oldest and innermost region of the brain. It’s designed to control the most basic functions of life,
including breathing, attention, and motor responses (Figure \(1\)). The brain stem begins where the spinal cord enters the skull and forms the medulla, the area of the brain stem that controls heart rate and breathing. In many cases the medulla alone is sufficient to maintain life—animals that have the remainder of their brains above the medulla severed are still able to eat, breathe, and even move. The spherical shape above the medulla is the pons, a structure in the brain stem that helps control the movements of the body, playing a particularly important role in balance and walking.
Running through the medulla and the pons is a long, narrow network of neurons known as the reticular formation. The job of the reticular formation is to filter out some of the stimuli that are coming into the brain from the spinal cord and to relay the remainder of the signals to other areas of the brain. The reticular formation also plays important roles in walking, eating, sexual activity, and sleeping. When electrical stimulation is applied to the reticular formation of an animal, it immediately becomes fully awake, and when the reticular formation is severed from the higher brain regions, the animal falls into a deep coma.
Above the brain stem are other parts of the old brain that also are involved in the processing of behavior and emotions (see Figure \(2\)). The thalamus is the egg-shaped structure above the brain stem that applies still more filtering to the sensory information that is coming up from the spinal cord and through the reticular formation, and it relays some of these remaining signals to the higher brain levels (Guillery & Sherman, 2002). The thalamus also receives some of the higher brain’s replies, forwarding them to the medulla and the cerebellum. The thalamus is also important in sleep because it shuts off incoming signals from the senses, allowing us to rest.
The cerebellum (literally, “little brain”) consists of two wrinkled ovals behind the brain stem. It functions to coordinate voluntary movement. People who have damage to the cerebellum have difficulty walking, keeping their balance, and holding their hands steady. Consuming alcohol influences the cerebellum, which is why people who are drunk have more difficulty walking in a straight line. Also, the cerebellum contributes to emotional responses, helps us discriminate between different sounds and textures, and is important in learning (Bower & Parsons, 2003).
Whereas the primary function of the brain stem is to regulate the most basic aspects of life, including motor functions, the limbic system is largely responsible for memory and emotions, including our responses to reward and punishment. The limbic system is a brain area, located between the brain stem and the two cerebral hemispheres, that governs emotion and memory. It includes the amygdala, the hypothalamus, and the hippocampus.
The amygdala consists of two almond-shaped clusters (amygdala comes from the Latin word for “almond”) and is primarily responsible for regulating our perceptions of, and reactions to, aggression and fear. The amygdala has connections to other bodily systems related to fear, including the sympathetic nervous system (which we will see later is important in fear responses), facial responses (which perceive and express emotions), the processing of smells, and the release of neurotransmitters related to stress and aggression (Best, 2009). In one early study, Klüver and Bucy (1939) damaged the amygdala of an aggressive rhesus monkey. They found that the once-angry animal immediately became passive and no longer responded to fearful situations with aggressive behavior. Electrical stimulation of the amygdala in other animals also influences aggression. In addition to helping us experience fear, the amygdala also helps us learn from situations that create fear. When we experience events that are dangerous, the amygdala stimulates the brain to remember the details of the situation so that we learn to avoid it in the future (Sigurdsson et al., 2007).
Located just under the thalamus (hence its name) the hypothalamus is a brain structure that contains a number of small areas that perform a variety of functions, including the important role of linking the nervous system to the endocrine system via the pituitary gland. Through its many interactions with other parts of the brain, the hypothalamus helps regulate body temperature, hunger, thirst, and sex, and responds to the satisfaction of these needs by creating feelings of plea- sure. Olds and Milner (1954) discovered these reward centers accidentally after they had momentarily stimulated the hypo- thalamus of a rat. The researchers noticed that after being stimulated, the rat continued to move to the exact spot in its cage where the stimulation had occurred, as if it were trying to re-create the circumstances surrounding its original experience. Upon further research into these reward centers, Olds (1958) discovered that animals would do almost anything to re-create enjoyable stimulation, including crossing a painful electrified grid to receive it. In one experiment a rat was given the opportunity to electrically stimulate its own hypothalamus by pressing a pedal. The rat enjoyed the experience so much that it pressed the pedal more than 7,000 times per hour until it collapsed from sheer exhaustion.
The hippocampus consists of two “horns” that curve back from the amygdala. The hippocampus is important in storing information in long-term memory. If the hippocampus is damaged, a person cannot build new memories, living instead in a strange world where everything he or she experiences just fades away, even while older memories from the time before the damage are untouched. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/07%3A_Our_Brains_Control_Our_Thoughts_Feelings_and_Behavior/7.01%3A_Introduction.txt |
All animals have adapted to their environments by developing abilities that help them survive. Some animals have hard shells, others run extremely fast, and some have acute hearing. Human beings do not have any of these particular characteristics, but we do have one big advantage over other animals—we are very, very smart.
You might think that we should be able to determine the intelligence of an animal by looking at the ratio of the animal’s brain weight to the weight of its entire body. But this does not really work. The elephant’s brain is one thousandth of its weight, but the whale’s brain is only one ten-thousandth of its body weight. On the other hand, although the human brain is one sixtieth of its body weight, the mouse’s brain represents one fortieth of its body weight. Despite these comparisons, elephants do not seem 10 times smarter than whales, and humans definitely seem smarter than mice.
The key to the advanced intelligence of humans is not found in the size of our brains. What sets humans apart from other animals is our larger cerebral cortex—the outer bark- like layer of our brain that allows us to so successfully use language, acquire complex skills, create tools, and live in social groups (Gibson, 2002). In humans, the cerebral cortex is wrinkled and folded, rather than smooth as it is in most other animals. This creates a much greater surface area and size, and allows increased capacities for learning, remembering, and thinking. The folding of the cerebral cortex is referred to as corticalization.
Although the cortex is only about one tenth of an inch thick, it makes up more than 80% of the brain’s weight. The cortex contains about 20 billion nerve cells and 300 trillion synaptic connections (de Courten-Myers, 1999). Supporting all these neurons are billions more glial cells (glia), cells that surround and link to the neurons, protecting them, providing them with nutrients, and absorbing unused neurotransmitters. The glia come in different forms and have different functions. For instance, the myelin sheath surrounding the axon of many neurons is a type of glial cell. The glia are essential partners of neurons, without which the neurons could not survive or function (Miller, 2005).
The cerebral cortex is divided into two hemispheres, and each hemisphere is divided into four lobes, each separated by folds known as fissures. If we look at the cortex starting at the front of the brain and moving over the top (see Figure \(1\)), we see first the frontal lobe (behind the forehead), which is responsible primarily for thinking, planning, memory, and judgment. Following the frontal lobe is the parietal lobe, which extends from the middle to the back of the skull and which is responsible primarily for processing information about touch. Then comes the occipital lobe, at the very back of the skull, which processes visual information. Finally, in front of the occipital lobe (pretty much between the ears) is the temporal lobe, responsible primarily for hearing and language.
Fritsch and Hitzig also found that the movement that followed the brain stimulation only occurred when they stimulated a specific arch-shaped region that runs across the top of the brain from ear to ear, just at the front of the parietal lobe (see Figure \(2\)). Fritsch and Hitzig had discovered the motor cortex, the part of the cortex that controls and executes movements of the body by sending signals to the cerebellum and the spinal cord. More recent research has mapped the motor cortex even more fully, by providing mild electronic stimulation to different areas of the motor cortex in fully conscious patients while observing their bodily responses (because the brain has no sensory receptors, these patients feel no pain). As you can see in Figure \(2\), this research has revealed that the motor cortex is specialized for providing control over the body, in the sense that the parts of the body that require more precise and finer movements, such as the face and the hands, also are allotted the greatest amount of cortical space.
Just as the motor cortex sends out messages to the specific parts of the body, the somatosensory cortex, an area just behind and parallel to the motor cortex at the back of the frontal lobe, receives information from the skin’s sensory receptors and the movements of different body parts. Again, the more sensitive the body region, the more area is dedicated to it in the sensory cortex. Our sensitive lips, for example, occupy a large area in the sensory cortex, as do our fingers and genitals.
Other areas of the cortex process other types of sensory information. The visual cortex is the area located in the occipital lobe (at the very back of the brain) that processes visual information (see Figure \(1\)). If you were stimulated in the visual cortex, you would see flashes of light or color, and perhaps you remember having had the experience of “seeing stars” when you were hit in, or fell on, the back of your head. The temporal lobe, located on the lower side of each hemisphere, contains the auditory cortex, which is responsible for hearing and language (see Figure \(1\)). The temporal lobe also processes some visual information, providing us with the ability to name the objects around us (Martin, 2007).
As you can see in Figure \(2\), the motor and sensory areas of the cortex account for a relatively small part of the total cortex. The remainder of the cortex is made up of association areas in which sensory and motor information is combined and associated with our stored knowledge. These association areas are the places in the brain that are responsible for most of the things that make human beings seem human. The association areas are involved in higher mental functions, such as learning, thinking, planning, judging, moral reflecting, figuring, and spatial reasoning. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/07%3A_Our_Brains_Control_Our_Thoughts_Feelings_and_Behavior/7.03%3A_The_Cerebral_Cortex_Creates_Consciousness_And_Thinking.txt |
The control of some specific bodily functions, such as movement, vision, and hearing, is performed in specified areas of the cortex, and if these areas are damaged, the individual will likely lose the ability to perform the corresponding function. For instance, if an infant suffers damage to facial recognition areas in the temporal lobe, it is likely that he or she will never be able to recognize faces (Farah et al., 2000). On the other hand, the brain is not divided up in an entirely rigid way. The brain’s neurons have a remarkable capacity to reorganize and extend themselves to carry out particular functions in response to the needs of the organism, and to repair damage. As a result, the brain constantly creates new neural communication routes and rewires existing ones. Neuroplasticity refers to the brain’s ability to change its structure and function in response to experience or damage. Neuroplasticity enables us to learn and remember new things and adjust to new experiences.
Our brains are the most “plastic” when we are young children, as it is during this time that we learn the most about our environment. On the other hand, neuroplasticity continues to be observed even in adults (Kolb & Fantie, 1989). The principles of neuroplasticity help us understand how our brains develop to reflect our experiences. For instance, accomplished musicians have a larger auditory cortex compared with the general population (Bengtsson et al., 2005) and also require less neural activity to move their fingers over musical instruments than do novices (Münte et al., 2002). These observations reflect the changes in the brain that follow our experiences.
Plasticity is also observed when there is damage to the brain or to parts of the body that are represented in the motor and sensory cortexes. When a tumor in the left hemisphere of the brain impairs language, the right hemisphere will begin to compensate to help the person recover the ability to speak (Thiel et al., 2006). And if a person loses a finger, the area of the sensory cortex that previously received information from the missing finger will begin to receive input from adjacent fingers, causing the remaining digits to become more sensitive to touch (Fox, 1984).
Although neurons cannot repair or regenerate themselves as skin or blood vessels can, new evidence suggests that the brain can engage in neurogenesis, the forming of new neurons (Van Praag et al., 2004). These new neurons originate deep in the brain and may then migrate to other brain areas where they form new connections with other neurons (Gould, 2007). This leaves open the possibility that someday scientists might be able to “rebuild” damaged brains by creating drugs that help grow neurons.
RESEARCH FOCUS
identifying the Unique Functions of the Left and Right Hemispheres Using Split-Brain Patients
We have seen that the left hemisphere of the brain primarily senses and controls the motor movements on the right side of the body, and vice versa. This fact provides an interesting way to study brain lateralization—the idea that the left and the right hemispheres of the brain are specialized to perform different functions. Gazzaniga et al. (1965) studied a patient, known as W. J., who had undergone an operation to relieve severe seizures. In this surgery the region that normally connects the two halves of the brain and supports communication between the hemispheres, known as the corpus callosum, is severed. As a result, the patient essentially becomes a person with two separate brains. Because the left and right hemispheres are separated, each hemisphere develops a mind of its own, with its own sensations, concepts, and motivations (Gazzaniga, 2005).
In their research, Gazzaniga and his colleagues (1965) tested the ability of W. J. to recognize and respond to objects and written passages that were presented to only the left or to only the right brain hemispheres (see Figure \(1\)). The researchers had W. J. look straight ahead and then flashed, for a fraction of a second, a picture of a geometrical shape to the left of where he was looking. By doing so, they assured that—because the two hemispheres had been separated—the image of the shape was experienced only in the right brain hemisphere (remember that sensory input from the left side of the body is sent to the right side of the brain). Gazzaniga and his colleagues found that W. J. was able to identify what he had been shown when he was asked to pick the object from a series of shapes, using his left hand, but that he could not do this when the object was shown in the right visual field. On the other hand, W. J. could easily read written material presented in the right visual field (and thus experienced in the left hemisphere) but not when it was presented in the left visual field.
This research, and many other studies following it, has demonstrated that the two brain hemispheres specialize in different abilities. In most people the ability to speak, write, and understand language is located in the left hemisphere. This is why W. J. could read passages that were presented on the right side and thus transmitted to the left hemisphere, but could not read passages that were only experienced in the right brain hemisphere. The left hemisphere is also better at math and at judging time and rhythm. It is also superior in coordinating the order of complex movements— for example, lip movements needed for speech. The right hemisphere, on the other hand, has only very limited verbal abilities, and yet it excels in perceptual skills. The right hemisphere is able to recognize objects, including faces, patterns, and melodies, and it can put a puzzle together or draw a picture. This is why W. J. could pick out the image when he saw it on the left, but not the right, visual field.
Although Gazzaniga’s research demonstrated that the brain is in fact lateralized, such that the two hemispheres specialize in different activities, this does not mean that when people behave in a certain way or perform a certain activity they are only using one hemisphere of their brains at a time. That would be drastically oversimplifying the concept of brain differences. We normally use both hemispheres at the same time, and the difference between the abilities of the two hemispheres is not absolute (Soroker et al., 2005). ■
PSYCHOLOGY IN EVERYDAY LIFE
Why Are Some People Left-Handed
Across cultures and ethnic groups, about 90% of people are mainly right-handed, whereas only 10% are primarily left-handed (Peters et al., 2006). This fact is puzzling, in part because the number of left-handers is so low, and in part because other animals, including our closest primate relatives, do not show any type of handedness. The existence of right-handers and left-handers provides an interesting example of the relationship among evolution, biology, and social factors and how the same phenomenon can be understood at different levels of analysis (Harris, 1990; McManus, 2002).
At least some handedness is determined by genetics. Ultrasound scans show that 9 out of 10 fetuses suck the thumb of their right hand, suggesting that the preference is determined before birth (Hepper et al., 2005), and the mechanism of transmission has been linked to a gene on the X chromosome (Jones & Martin, 2000). It has also been observed that left-handed people are likely to have fewer children, and this may be in part because the mothers of left-handers are more prone to miscarriages and other prenatal problems (Mckeever et al., 2000).
But culture also plays a role. In the past, left-handed children were forced to write with their right hands in many countries, and this practice continues, particularly in collectivistic cultures, such as India and Japan, where left-handedness is viewed negatively as compared with individualistic societies, such as the United States. For example, India has about half as many left-handers as the United States (Ida & Mandal, 2003).
There are both advantages and disadvantages to being left-handed in a world where most people are right-handed. One problem for lefties is that the world is designed for right- handers. Automatic teller machines (ATMs), classroom desks, scissors, microscopes, drill presses, and table saws are just some examples of everyday machinery that is designed with the most important controls on the right side. This may explain in part why left-handers suffer somewhat more accidents than do right-handers (Dutta & Mandal, 2006).
Despite the potential difficulty living and working in a world designed for right-handers, there seem to be some advantages to being left-handed. Throughout history, a number of prominent artists have been left-handed, including leonardo da vinci, Michelangelo, Pablo Picasso, and Max Escher. Because the right hemisphere is superior in imaging and visual abilities, there may be some advantage to using the left hand for drawing or painting (Springer & Deutsch, 1998). left-handed people are also better at envisioning three-dimensional objects, which may explain why there is such a high number of left-handed architects, artists,
and chess players in proportion to their numbers (Coren, 1992). However, there are also more left-handers among those with reading disabilities, allergies, and migraine headaches (Geschwind & Behan, 2007), perhaps due to the fact that a small minority of left-handers owe their handedness to a birth trauma, such as being born prematurely (Betancur et al., 1990).
In sports in which handedness may matter, such as tennis, boxing, fencing, or judo, left-handers may have an advantage. They play many games against right-handers and learn how to best handle their styles. Right-handers, however, play very few games against left-handers, which may make them more vulnerable. This explains why a disproportionately high number of left-handers are found in sports where direct one- on-one action predominates. In other sports, such as golf, there are fewer left-handed players because the handedness of one player has no effect on the competition.
The fact that left-handers excel in some sports suggests the possibility that they may have also had an evolutionary advantage because their ancestors may have been more successful in important skills such as hand-to-hand combat (Bodmer & Mckie, 1994). At this point, however, this idea remains only a hypothesis, and determinants of human handedness are yet to be fully understood. ■ | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/07%3A_Our_Brains_Control_Our_Thoughts_Feelings_and_Behavior/7.04%3A_The_Brain_Is_Flexible-_Neuroplasticity.txt |
• The old brain—including the brain stem, medulla, pons, reticular formation, thalamus, cerebellum, amygdala, hypothalamus, and hippocampus—regulates basic survival functions, such as breathing, moving, resting, feeding, emotions, and memory.
• The cerebral cortex, made up of billions of neurons and glial cells, is divided into the right and left hemispheres and into four lobes.
• The frontal lobe is primarily responsible for thinking, planning, memory, and judgment. The parietal lobe is primarily responsible for bodily sensations and touch. The temporal lobe is primarily responsible for hearing and language. The occipital lobe is primarily responsible for vision. Other areas of the cortex act as association areas, responsible for integrating information.
• The brain changes as a function of experience and potential dam- age in a process known as plasticity. The brain can generate new neurons through neurogenesis.
• The motor cortex controls voluntary movements. Body parts requiring the most control and dexterity take up the most space in the motor cortex.
• The sensory cortex receives and processes bodily sensations. Body parts that are the most sensitive occupy the greatest amount of space in the sensory cortex.
• The left cerebral hemisphere is primarily responsible for language and speech in most people, whereas the right hemisphere specializes in spatial and perceptual skills, visualization, and the recognition of patterns, faces, and melodies.
• The severing of the corpus callosum, which connects the two hemispheres, creates a “split-brain patient,” with the effect of creating two separate minds operating in one person.
• Studies with split-brain patients as research participants have been used to study brain lateralization.
• Neuroplasticity allows the brain to adapt and change as a function of experience or damage.
EXERCISES AND CRITICAL THINKING
1. Do you think that animals experience emotion? What aspects of brain structure might lead you to believe that they do or do not?
2. Consider your own experiences and speculate on which parts of your brain might be particularly well developed as a result of these experiences.
3. Which brain hemisphere are you likely to be using when you search for a fork in the silverware drawer? Which brain hemisphere are you most likely to be using when you struggle to remember the name of an old friend?
4. Do you think that encouraging left-handed children to use their right hands is a good idea? Why or why not?
REFERENCES
Bengtsson, S. L., Nagy, Z., Skare, S., Forsman, L., Forssberg, H., & Ullén, F. (2005). Extensive piano practicing has regionally specific effects on white matter development. Nature Neuroscience, 8(9), 1148–1150. doi.org/10.1038/nn1516
Best, B. (2009). The amygdala and the emotions. In The anatomical basis of mind (Chap. 9). https://www.benbest.com/science/anatmind/ anatmind.html
Betancur, C., Vélez, A., Cabanieu, G., Le Moal, M., & Neveu, P. J. (1990). Association between left-handedness and allergy: A reappraisal. Neuropsychologia, 28(2), 223–227. doi.org/10.1016/ 0028-3932(90)90104-V
Bodmer, W., & McKie, R. (1994). The book of man: The quest to discover our genetic heritage. Little, Brown and Company.
Bower, J. M., & Parsons, J. M. (2003). Rethinking the lesser brain. Scientific American, 289, 50–57. doi.org/10.1038/ scientificamerican0803-48
Camazine, S. (n.d.). The human brain. Scott Camazine Photography and Digital Imagery. http://www.scottcamazine.com/Camazine_iWeb/ Intro.html
Coren, S. (1992). The left-hander syndrome: The causes and consequences of left-handedness. Free Press.
de Courten-Myers, G. M. (1999). The human cerebral cortex: Gender differences in structure and function. Journal of Neuropathology and Experimental Neurology, 58, 217–226. doi.org/10.1097/ 00005072-199903000-00001
Dutta, T., & Mandal, M. K. (2006). Hand preference and accidents in India. Laterality: Asymmetries of Body, Brain, and Cognition, 11, 368–372. https://doi.org/10.1080/13576500600624239
Farah, M. J., Rabinowitz, C., Quinn, G. E., & Liu, G. T. (2000). Early commitment of neural substrates for face recognition. Cognitive Neuropsychology, 17(1–3), 117–123. doi.org/10.1080/ 026432900380526
Fox, J. L. (1984). The brain’s dynamic way of keeping in touch. Science, 225(4664), 820–821. https://doi.org/10.1126/science.6474155
Fritsch, G., & Hitzig, E. (2009). Electric excitability of the cerebrum (Über die elektrische Erregbarkeit des Grosshirns). Epilepsy & Behavior, 15(2), 123–130. (Original work published 1870). https://doi.org/ 10.1016/j.yebeh.2009.03.001
Gazzaniga, M. S. (2005). Forty-five years of split-brain research and still going strong. Nature Reviews Neuroscience, 6(8), 653–659. https:// doi.org/10.1038/nrn1723
Gazzaniga, M. S., Bogen, J. E., & Sperry, R. W. (1965). Observations on visual perception after disconnexion of the cerebral hemispheres in man. Brain, 88(2), 221–236. https://doi.org/10.1093/brain/88.2.221
Geschwind, N., & Behan, P. (2007). Left-handedness: Association with immune disease, migraine, and developmental learning disorder. MIT Press.
Gibson, K. R. (2002). Evolution of human intelligence: The roles of brain size and mental construction. Brain Behavior and Evolution 59, 10–20. https://doi.org/10.1159/000063730
Gould, E. (2007). How widespread is adult neurogenesis in mammals? Nature Reviews Neuroscience 8, 481–488. https://doi.org/10.1038/ nrn2147
Guillery, R. W., & Sherman, S. M. (2002). Thalamic relay functions and their role in corticocortical communication: Generalizations from the visual system. Neuron, 33(2), 163–175. doi.org/10.1016/ s0896-6273(01)00582-7
Harris, L. J. (1990). Cultural influences on handedness: Historical and contemporary theory and evidence. In S. Coren (Ed.), Left- handedness: Behavioral implications and anomalies. Elsevier.
Hepper, P. G., Wells, D. L., & Lynch, C. (2005). Prenatal thumb sucking is related to postnatal handedness. Neuropsychologia, 43, 313–315. https://doi.org/10.1016/j.neuropsych...ia.2004.08.009
Ida, Y., & Mandal, M. K. (2003). Cultural differences in side bias: Evidence from Japan and India. Laterality: Asymmetries of Body, Brain, and Cognition, 8(2), 121–133. https://doi.org/10.1080/713754478
Jones, G. V., & Martin, M. (2000). A note on Corballis (1997) and the genetics and evolution of handedness: Developing a unified distributional model from the sex-chromosomes gene hypothesis. Psychological Review, 107(1), 213–218. doi.org/10.1037/ 0033-295X.107.1.213
Klüver, H., & Bucy, P. C. (1939). Preliminary analysis of functions of the temporal lobes in monkeys. Archives of Neurology & Psychiatry, 42, 979–1000. doi.org/10.1001/archneurpsyc.1939. 02270240017001
Kolb, B., & Fantie, B. (1989). Development of the child’s brain and behavior. In C. R. Reynolds & E. Fletcher-Janzen (Eds.), Handbook of clinical child neuropsychology (pp. 17–39). Plenum Press.
Martin, A. (2007). The representation of object concepts in the brain. Annual Review of Psychology, 58, 25–45. doi.org/10.1146/ annurev.psych.57.102904.190143
McKeever, W. F., Cerone, L. J., Suter, P. J., & Wu, S. M. (2000). Family size, miscarriage-proneness, and handedness: Tests of hypotheses of the developmental instability theory of handedness. Laterality: Asymmetries of Body, Brain, and Cognition, 5(2), 111–120. doi. org/10.1080/713754367
McManus, I. C. (2002). Right hand, left hand: The origins of asymmetry in brains, bodies, atoms, and cultures. Harvard University Press.
Miller, G. (2005). Neuroscience: The dark side of glia. Science, 308(5723), 778–781. https://doi.org/10.1126/science.308.5723.778
Münte, T. F., Altenmüller, E., & Jäncke, L. (2002). The musician’s brain as a model of neuroplasticity. Nature Reviews Neuroscience, 3(6), 473–478. https://doi.org/10.1038/nrn843
Olds, J. (1958). Self-stimulation of the brain: Its use to study local effects of hunger, sex, and drugs. Science, 127, 315–324. https://doi.org/ 10.1126/science.127.3294.315
Olds, J., & Milner, P. (1954). Positive reinforcement produced by electrical stimulation of septal area and other regions of rat brain. Journal of Comparative and Physiological Psychology, 47, 419–427. doi. org/10.1037/h0058775
Peters, M., Reimers, S., & Manning, J. T. (2006). Hand preference for writing and associations with selected demographic and behavioral variables in 255,100 subjects: The BBC Internet study. Brain and Cognition, 62(2), 177–189. doi.org/10.1016/j.bandc.2006. 04.005
Sherman, S. M., & Guillery, R. W. (2006). Exploring the thalamus and its role in cortical function (2nd ed.). MIT Press.
Sigurdsson, T., Doyère, V., Cain, C. K., & LeDoux, J. E. (2007). Long-term potentiation in the amygdala: A cellular mechanism of fear learning and memory. Neuropharmacology, 52(1), 215–227. https://doi.org/ 10.1016/j.neuropharm.2006.06.022
Soroker, N., Kasher, A., Giora, R., Batori, G., Corn, C., Gil, M., & Zaidel, E. (2005). Processing of basic speech acts following localized brain damage: A new light on the neuroanatomy of language. Brain and Cognition, 57(2), 214–217. doi.org/10.1016/j.bandc.2004. 08.047
Springer, S. P., & Deutsch, G. (1998). Left brain, right brain: Perspectives from cognitive neuroscience (5th ed.). W. H. Freeman/Times Books/ Henry Holt & Co.
Thiel, A., Habedank, B., Herholz, K., Kessler, J., Winhuisen, L., Haupt,
W. F., & Heiss, W. D. (2006). From the left to the right: How the brain compensates progressive loss of language function. Brain and Language, 98(1), 57–65. https://doi.org/10.1016/j.bandl.2006.01.007
Van Praag, H., Zhao, X., Gage, F. H., & Gazzaniga, M. S. (2004). Neurogenesis in the adult mammalian brain. In M. S. Gazzaniga (Ed.), The cognitive neurosciences (3rd ed., pp. 127–137). MIT Press.
Wikia Education. (n.d.). Cerebral cortex. Psychology Wiki. https:// psychology.wikia.org/wiki/Cerebral_cortex | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/07%3A_Our_Brains_Control_Our_Thoughts_Feelings_and_Behavior/7.05%3A_Key_Takeways.txt |
As a generally noninvasive subset of neuroscience methods, psychophysiological methods are used across a variety of disciplines in order to answer diverse questions about psychology, both mental events and behavior. Many different techniques are classified as psychophysiological. Each technique has its strengths and weaknesses, and knowing them allows researchers to decide what each offers for a particular question. Additionally, this knowledge allows research consumers to evaluate the meaning of the results in a particular experiment.
HISTORY
In the mid-nineteenth century, a railroad worker named Phineas Gage was in charge of setting explosive charges for blasting through rock in order to prepare a path for railroad tracks. He would lay the charge in a hole drilled into the rock, place a fuse and sand on top of the charge, and pack it all down using a tamping iron (a solid iron rod approximately one yard long and a little over an inch in diameter). On a September afternoon when Gage was performing this task, his tamping iron caused a spark that set off the explosive prematurely, sending the tamping iron flying through the air.
Unfortunately for Gage, his head was above the hole and the tamping iron entered the side of his face, passed behind his left eye, and exited out of the top of his head, eventually landing 80 feet away. Gage lost a portion of his left frontal lobe in the accident, but survived and lived for another 12 years. What is most interesting from a psychological perspective is that Gage’s personality changed as a result of this accident. He became more impulsive, he had trouble carrying out plans, and, at times, he engaged in vulgar profanity, which was out of character. This case study leads one to believe that there are specific areas of the brain that are associated with certain psychological phenomena. When studying psychology, the brain is indeed an interesting source of information. Although it would be impossible to replicate the type of damage done to Gage in the name of research, methods have developed over the years that are able to safely measure different aspects of nervous system activity in order to help researchers better understand psychology as well as the relationship between psychology and biology.
INTRODUCTION
Psychophysiology is defined as any research in which the dependent variable (what the researcher measures) is a physiological measure, and the independent variable (what the researcher manipulates) is behavioral or mental. In most cases the work is done noninvasively with awake human participants. Physiological measures take many forms and range from blood flow or neural activity in the brain to heart rate variability and eye movements. These measures can provide information about processes including emotion, cognition, and the interactions between them. In these ways, physiological measures offer a very flexible set of tools for researchers to answer questions about behavior, cognition, and health.
Psychophysiological methods are a subset of the very large domain of neuroscience methods. Many neuroscience methods are invasive, such as involving lesions of neural tissue, injection of neutrally active chemicals, or manipulation of neural activity via electrical stimulation. The present survey emphasizes noninvasive methods widely used with human subjects.
Crucially, in examining the relationship between physiology and overt behavior or mental events, psychophysiology does not attempt to replace the latter with the former. As an example, happiness is a state of pleasurable contentment and is associated with various physiological measures, but one would not say that those physiological measures are happiness. We can make inferences about someone’s cognitive or emotional state based on his or her self-report, physiology, or overt behavior. Sometimes our interest is primarily in inferences about internal events and sometimes primarily in the physiology itself. Psychophysiology addresses both kinds of goals. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/08%3A_Psychophysiological_Methods_in_Neuroscience/8.01%3A_Introduction.txt |
This module provides an overview of several popular psycho-physiological methods, though it is far from exhaustive. Each method can draw from a broad range of data-analysis strategies to provide an even more expansive set of tools. The psychophysiological methods discussed below focus on the central nervous system. Structural magnetic resonance imaging (sMRI) is a noninvasive technique that allows researchers and clinicians to view anatomical structures within a human. The participant is placed in a magnetic field that may be 66,000 times greater than the Earth’s magnetic field, which causes a small portion of the atoms in his or her body to line up in the same direction. The body is then pulsed with low-energy radio frequencies that are absorbed by the atoms in the body, causing them to tip over. As these atoms return to their aligned state, they give off energy in the form of harmless electromagnetic radiation, which is measured by the machine. The machine then transforms the measured energy into a three-dimensional picture of the tissue within the body. In psychophysiology research, this image may be used to compare the size of structures in different groups of people (e.g., are areas associated with pleasure smaller in individuals with depression?) or to increase the accuracy of spatial locations as measured with functional magnetic resonance imaging.
Functional magnetic resonance imaging (fMRI) is a method that is used to assess changes in activity of tissue, such as measuring changes in neural activity in different areas of the brain during thought. This technique builds on the principles of sMRI and also uses the property that, when neurons fire, they use energy, which must be replenished. Glucose and oxygen, two key components for energy production, are supplied to the brain from the blood stream as needed. Oxygen is transported through the blood using hemoglobin, which contains binding sites for oxygen. When these sites are saturated with oxygen, it is referred to as oxygenated hemoglobin. When the oxygen molecules have all been released from a hemoglobin molecule, it is known as deoxygenated hemoglobin. As a set of neurons begins firing, oxygen in the blood surrounding those neurons is consumed, leading to a reduction in oxygenated hemoglobin. The body then compensates and provides an abundance of oxygenated hemoglobin in the blood surrounding that activated neural tissue. When activity in that neural tissue declines, the level of oxygenated hemoglobin slowly returns to its original level, which typically takes several seconds.
fMRI measures the change in the concentration of oxygenated hemoglobin, which is known as the blood-oxygen- level-dependent (BOLD) signal. This leads to two import- ant facts about fMRI. First, fMRI measures blood volume and blood flow, and from this we infer neural activity; fMRI does not measure neural activity directly. Second, fMRI data typically have poor temporal resolution (the precision of measurement with respect to time); however, when combined with sMRI, fMRI provides excellent spatial resolution (the ability to distinguish one object from another in space). Temporal resolution for fMRI is typically on the order of seconds, whereas its spatial resolution is on the order of millimeters. Under most conditions there is an inverse relationship between temporal and spatial resolution—one can increase temporal resolution at the expense of spatial resolution and vice versa.
This method is valuable for identifying specific areas of the brain that are associated with different physical or psycho- logical tasks. Clinically, fMRI may be used before neurosurgery in order to identify areas that are associated with language so that the surgeon can avoid those areas during the operation. fMRI allows researchers to identify differential or convergent patterns of activation associated with tasks. For example, if participants are shown words on a screen and are expected to indicate the color of the letters, are the same brain areas recruited for this task if the words have emotional con- tent or not? Does this relationship change in psychological disorders such as anxiety or depression? Is there a different pattern of activation even in the absence of overt performance differences? fMRI is an excellent tool for comparing brain activation in different tasks and/or populations. Figure \(2\) provides an example of results from fMRI analyses overlaid on an sMRI image. The blue and orange shapes represent areas with significant changes in the BOLD signal, thus changes in neural activation.
Electroencephalography (EEG) is another technique for studying brain activation. This technique uses at least two and sometimes up to 256 electrodes to measure the difference in electrical charge (the voltage) between pairs of points on the head. These electrodes are typically fastened to a flexible cap (similar to a swimming cap) that is placed on the participant’s head. From the scalp, the electrodes measure the electrical activity that is naturally occurring within the brain. They do not introduce any new electrical activity. In contrast to fMRI, EEG measures neural activity directly, rather than a correlate of that activity.
Electrodes used in EEG can also be placed within the skull, resting directly on the brain itself. This application, called electrocorticography (ECoG), is typically used before medical procedures for localizing activity, such as the origin of epileptic seizures. This invasive procedure allows for more precise localization of neural activity, which is essential in medical applications. However, it is generally not justifiable to open a person’s skull solely for research purposes, and instead electrodes are placed on the participant’s scalp, resulting in a noninvasive technique for measuring neural activity.
Given that this electrical activity must travel through the skull and scalp before reaching the electrodes, localization of activity is less precise when measuring from the scalp, but it can still be within several millimeters when localizing activity that is near the scalp. One major advantage of EEG is its temporal resolution. Data can be recorded thousands of times per second, allowing researchers to document events that happen in less than a millisecond. EEG analyses typically investigate the change in amplitude or frequency components of the recorded EEG on an ongoing basis or averaged over dozens of trials (see Figure \(3\)).
Magnetoencephalography (MEG) is another technique for noninvasively measuring neural activity. The flow of electrical charge (the current) associated with neural activity produces very weak magnetic fields that can be detected by sensors placed near the participant’s scalp. The number of sensors used varies from a few to several hundred. Due to the fact that the magnetic fields of interest are so small, special rooms that are shielded from magnetic fields in the environment are needed in order to avoid contamination of the signal being measured. MEG has the same excellent temporal resolution as EEG. Additionally, MEG is not as susceptible to distortions from the skull and scalp. Magnetic fields are able to pass through the hard and soft tissue relatively unchanged, thus providing better spatial resolution than EEG. MEG analytic strategies are nearly identical to those used in EEG. However, the MEG recording apparatus is much more expensive than EEG, so MEG is much less widely available.
EEG and MEG are both excellent for elucidating the temporal dynamics of neural processes. For example, if someone is reading a sentence that ends with an unexpected word (e.g., Michelle is going outside to water the book), how long after he or she reads the unexpected word does he or she recognize this as unexpected? In addition to these types of questions, EEG and MEG methods allow researchers to investigate the degree to which different parts of the brain “talk” to each other. This allows for a better understanding of brain net- works, such as their role in different tasks and how they may function abnormally in psychopathology.
Positron emission tomography (PET) is a medical imaging technique that is used to measure processes in the body, including the brain. This method relies on a positron-emitting tracer atom that is introduced into the blood stream in a biologically active molecule, such as glucose, water, or ammonia. A positron is a particle much like an electron but with a positive charge. One example of a biologically active molecule is fludeoxyglucose, which acts similarly to glucose in the body.
Fludeoxyglucose will concentrate in areas where glucose is needed—commonly areas with higher metabolic needs. Over time, this tracer molecule emits positrons, which are detected by a sensor. The spatial location of the tracer molecule in the brain can be determined based on the emitted positrons. This allows researchers to construct a three-dimensional image of the areas of the brain that have the highest metabolic needs, typically those that are most active. Images resulting from PET usually represent neural activity that has occurred over tens of minutes, which is very poor temporal resolution for some purposes. PET images are often combined with computed tomography (CT) images to improve spatial resolution, as fine as several millimeters. Tracers can also be incorporated into molecules that bind to neurotransmitter receptors, which allows researchers to answer some unique questions about the action of neurotransmitters. Unfortunately, very few research centers have the equipment required to obtain the images or the special equipment needed to create the positron-emitting tracer molecules, which typically need to be produced on site.
Transcranial magnetic stimulation (TMS) is a noninvasive method that causes depolarization or hyperpolarization in neurons near the scalp. This method is not considered psychophysiological because the independent variable is physiological, rather than the dependent. However, it does qualify as a neuroscience method because it deals with the function of the nervous system, and it can readily be combined with conventional psychophysiological methods. In TMS, a coil of wire is placed just above the participant’s scalp. When electricity flows through the coil, it produces a magnetic field. This magnetic field travels through the skull and scalp and affects neurons near the surface of the brain. When the magnetic field is rapidly turned on and off, a current is induced in the neurons, leading to depolarization or hyper- polarization, depending on the number of magnetic field pulses. Single- or paired-pulse TMS depolarizes site-specific neurons in the cortex, causing them to fire. If this method is used over primary motor cortex, it can produce or block muscle activity, such as inducing a finger twitch or preventing someone from pressing a button. If used over primary visual cortex, it can produce sensations of flashes of light or impair visual processes. This has proved to be a valuable tool in studying the function and timing of specific processes such as the recognition of visual stimuli. Repetitive TMS produces effects that last longer than the initial stimulation. Depending on the intensity, coil orientation, and frequency, neural activity in the stimulated area may be either attenuated or amplified. Used in this manner, TMS is able to explore neural plasticity, which is the ability of connections between neurons to change. This has implications for treating psychological disorders as well as understanding long-term changes in neuronal excitability. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/08%3A_Psychophysiological_Methods_in_Neuroscience/8.02%3A_Central_Nervous_System_%28CNS%29.txt |
The psychophysiological methods discussed above focus on the central nervous system. Considerable research has also focused on the peripheral nervous system. These methods include skin conductance, cardiovascular responses, muscle activity, pupil diameter, eye blinks, and eye movements. Skin conductance, for example, measures the electrical conductance (the inverse of resistance) between two points on the skin, which varies with the level of moisture. Sweat glands are responsible for this moisture and are controlled by the sympathetic nervous system (SNS). Increases in skin conductance can be associated with changes in psychological activity. For example, studying skin conductance allows a researcher to investigate whether psychopaths react to fearful pictures in a normal way. Skin conductance provides relatively poor temporal resolution, with the entire response typically taking several seconds to emerge and resolve. However, it is an easy way to measure SNS response to a variety of stimuli.
Cardiovascular measures include heart rate, heart rate variability, and blood pressure. The heart is innervated by the parasympathetic nervous system (PNS) and SNS. Input from the PNS decreases heart rate and contractile strength, whereas input from the SNS increases heart rate and contractile strength. Heart rate can easily be monitored using a minimum of two electrodes and is measured by counting the number of heartbeats in a given time period, such as one minute, or by assessing the time between successive heartbeats. Psychological activity can prompt increases and decreases in heart rate, often in less than a second, making heart rate a sensitive measure of cognition. Measures of heart rate variability are concerned with consistency in the time interval between heartbeats. Changes in heart rate variability are associated with stress as well as psychiatric conditions. Figure \(1\) is an example of an electrocardiogram, which is used to measure heart rate and heart rate variability. These cardio- vascular measures allow researchers to monitor SNS and PNS reactivity to various stimuli or situations. For example, when an arachnophobe views pictures of spiders, does their heart rate increase more than that of a person not afraid of spiders?
Electromyography (EMG) measures electrical activity produced by skeletal muscles. Similar to EEG, EMG measures the voltage between two points. This technique can be used to determine when a participant first initiates muscle activity to engage in a motor response to a stimulus or the degree to which a participant begins to engage in an incorrect response (such as pressing the wrong button), even if it is never visibly executed. It has also been used in emotion research to identify activity in muscles that are used to produce smiles and frowns. Using EMG, it is possible to detect very small facial movements that are not observable from looking at the face. The temporal resolution of EMG is similar to that of EEG and MEG.
Valuable information can also be gleaned from eye blinks, eye movements, and pupil diameter. Eye blinks are most often assessed using EMG electrodes placed just below the eyelid, but electrical activity associated directly with eye blinks or eye movements can be measured with electrodes placed on the face near the eyes, because there is voltage across the entire eyeball. Another option for the measurement of eye movement is a camera used to record video of an eye. This video method is particularly valuable when determination of absolute direction of gaze (not just change in direction of gaze) is of interest, such as when the eyes scan a picture. With the help of a calibration period in which a participant looks at multiple, known targets, eye position is then extracted from each video frame during the main task and compared with data from the calibration phase, allowing researchers to identify the sequence, direction, and duration of gaze fixations. For example, when viewing pleasant or unpleasant images, people spend different amounts of time looking at the most arousing parts. This, in turn, can vary as a function of psychopathology. Additionally, the diameter of a participant’s pupil can be measured and recorded over time from the video record. As with heart rate, pupil diameter is controlled by competing inputs from the SNS and PNS. Pupil diameter is commonly used as an index of mental effort when performing a task.
8.04: When To Use What
As the reader, you may be wondering, how do I know what tool is right for a given question? Generally, there are no definitive answers. If you wanted to know the temperature in the morning, would you check your phone? Look outside to see how warm it looks? Ask your roommate what he or she is wearing today? Look to see what other people are wearing? There is not a single way to answer the question. The same is true for research questions. However, there are some guidelines that one can consider. For example, if you are interested in what brain structures are associated with cognitive control, you wouldn’t use peripheral nervous system measures. A technique such as fMRI or PET might be more appropriate. If you are interested in how cognitive control unfolds over time, EEG or MEG would be a good choice. If you are interested in studying the bodily response to fear in different groups of people, peripheral nervous system measures might be most appropriate. The key to deciding what method is most appropriate is properly defining the question that you are trying to answer. What aspects are most interesting? Do you care about identifying the most relevant brain structures? Temporal dynamics? Bodily responses? Then, it is important to think about the strengths and weaknesses of the different psycho- physiological measures and pick one, or several, whose attributes work best for the question at hand. In fact, it is common to record several at once.
8.05: Conclusion
The outline of psychophysiological methods above provides a glimpse into the exciting techniques that are available to researchers studying a broad range of topics from clinical to social to cognitive psychology. Some of the most interesting psychophysiological studies use several methods, such as in sleep assessments or multimodal neuroimaging. Psychophysiological methods have applications outside of mainstream psychology in areas where psychological phenomena are central, such as economics, health-related decision making, and brain–computer interfaces. Examples of applications for each method are provided above, but this list is by no means exhaustive. Furthermore, the field is continually evolving, with new methods and new applications being developed. The wide variety of methods and applications provide virtually limitless possibilities for researchers.
REFERENCE
Van Horn, J. D., Irimia, A., Torgerson, C. M., Chambers, M. C., Kikinis, R., & Toga, A. W. (2012). Mapping connectivity damage in the case of Phineas Gage. PloS One, 7(5), e37454. doi.org/10.1371/journal. pone.0037454 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/08%3A_Psychophysiological_Methods_in_Neuroscience/8.03%3A_Peripheral_Nervous_System.txt |
Consciousness describes our awareness of internal and external stimuli. Awareness of internal stimuli includes feeling pain, hunger, thirst, sleepiness, and being aware of our thoughts and emotions. Awareness of external stimuli includes seeing the light from the sun, feeling the warmth of a room, and hearing the voice of a friend. We experience different states of consciousness and different levels of awareness on a regular basis. We might even describe consciousness as a continuum that ranges from full awareness to a deep sleep. Sleep is a state marked by relatively low levels of physical activity and reduced sensory awareness that is distinct from periods of rest that occur during wakeful- ness. Wakefulness is characterized by high levels of sensory awareness, thought, and behavior. Between these extremes are states of consciousness related to daydreaming, intoxication as a result of alcohol or other drug use, meditative states, hypnotic states, and altered states of consciousness following sleep deprivation. We might also experience unconscious states of being via drug-induced anesthesia for medical purposes. Often, we are not completely aware of our surroundings, even when we are fully awake. For instance, have you ever daydreamed while driving home from work or school without really thinking about the drive itself? You were capable of engaging in the all of the complex tasks involved with operating a motor vehicle even though you were not aware of doing so. Many of these processes, like much of psychological behavior, are rooted in our biology.
9.02: Biological Rhythms
Biological rhythms are internal rhythms of biological activity. A woman’s menstrual cycle is an example of a biological rhythm—a recurring, cyclical pattern of bodily changes. One complete menstrual cycle takes about 28 days—a lunar month—but many biological cycles are much shorter. For example, body temperature fluctuates cyclically over a 24-hour period (Figure \(1\)). Alertness is associated with higher body temperatures, and sleepiness with lower body temperatures.
This pattern of temperature fluctuation, which repeats every day, is one example of a circadian rhythm. A circadian rhythm is a biological rhythm that takes place over a period of about 24 hours. Our sleep-wake cycle, which is linked to our environment’s natural light-dark cycle, is perhaps the most obvious example of a circadian rhythm, but we also have daily fluctuations in heart rate, blood pressure, blood sugar, and body temperature. Some circadian rhythms play a role in changes in our state of consciousness.
If we have biological rhythms, then is there some sort of biological clock? In the brain, the hypothalamus, which lies above the pituitary gland, is a main center of homeostasis. Homeostasis is the tendency to maintain a balance, or optimal level, within a biological system.
The brain’s clock mechanism is located in an area of the hypothalamus known as the suprachiasmatic nucleus (SCN). The axons of light-sensitive neurons in the retina pro- vide information to the SCN based on the amount of light present, allowing this internal clock to be synchronized with the outside world (Klein et al., 1991; Welsh et al., 2010) (Figure \(2\)).
9.03: Problems With Circadian Rhythms
Generally, and for most people, our circadian cycles are aligned with the outside world. For example, most people sleep during the night and are awake during the day. One important regulator of sleep-wake cycles is the hormone melatonin. The pineal gland, an endocrine structure located inside the brain that releases melatonin, is thought to be involved in the regulation of various biological rhythms and of the immune system during sleep (Hardeland et al., 2006). Melatonin release is stimulated by darkness and inhibited by light.
There are individual differences with regards to our sleep- wake cycle. For instance, some people would say they are morning people, while others would consider themselves to be night owls. These individual differences in circadian patterns of activity are known as a person’s chronotype, and research demonstrates that morning larks and night owls differ with regard to sleep regulation (Taillard et al., 2003). Sleep regulation refers to the brain’s control of switching between sleep and wakefulness as well as coordinating this cycle with the outside world.
LINK TO LEARNING
Watch this brief video describing circadian rhythms and how they affect sleep.
Disruptions of Normal Sleep
Whether lark, owl, or somewhere in between, there are situations in which a person’s circadian clock gets out of synchrony with the external environment. One way that this happens involves traveling across multiple time zones. When we do this, we often experience jet lag. Jet lag is a collection of symptoms that results from the mismatch between our internal circadian cycles and our environment. These symptoms include fatigue, sluggishness, irritability, and insomnia (i.e., a consistent difficulty in falling or staying asleep for at least three nights a week over a month’s time) (Roth, 2007).
Individuals who do rotating shift work are also likely to experience disruptions in circadian cycles. Rotating shift work refers to a work schedule that changes from early to late on a daily or weekly basis. For example, a person may work from 7:00 a.m. to 3:00 p.m. on Monday, 3:00 a.m. to 11:00 a.m. on Tuesday, and 11:00 a.m. to 7:00 p.m. on Wednesday. In such instances, the individual’s schedule changes so frequently that it becomes difficult for a normal circadian rhythm to be maintained. This often results in sleeping problems, and it can lead to signs of depression and anxiety. These kinds of schedules are common for individuals working in health care professions and service industries, and they are associated with persistent feelings of exhaustion and agitation that can make someone more prone to making mistakes on the job (Gold et al., 1992; Presser, 1995).
Rotating shift work has pervasive effects on the lives and experiences of individuals engaged in that kind of work, which is clearly illustrated in stories reported in a qualitative study that researched the experiences of middle-aged nurses who worked rotating shifts (West et al., 2009). Several of the nurses interviewed commented that their work schedules affected their relationships with their family. One of the nurses said,
If you’ve had a partner who does work regular job 9 to 5 office hours . . . the ability to spend time, good time with them when you’re not feeling absolutely exhausted . . . that would be one of the problems that I’ve encountered. (West et al., 2009, p. 114)
While disruptions in circadian rhythms can have negative consequences, there are things we can do to help us realign our biological clocks with the external environment. Some of these approaches, such as using a bright light as shown in Figure \(1\), have been shown to alleviate some of the problems experienced by individuals suffering from jet lag or from the consequences of rotating shift work. Because the biological clock is driven by light, exposure to bright light during working shifts and dark exposure when not working can help combat insomnia and symptoms of anxiety and depression (Huang et al., 2013).
LINK TO LEARNING
Watch this video to learn tips on how to overcome jet lag.
Insufficient Sleep
When people have difficulty getting sleep due to their work or the demands of day-to-day life, they accumulate a sleep debt. A person with a sleep debt does not get sufficient sleep on a chronic basis. The consequences of sleep debt include decreased levels of alertness and mental efficiency. Interestingly, since the advent of electric light, the amount of sleep that people get has declined. While we certainly welcome the convenience of having the darkness lit up, we also suffer the consequences of reduced amounts of sleep because we are more active during the nighttime hours than our ancestors were. As a result, many of us sleep less than 7–8 hours a night and accrue a sleep debt. While there is tremendous variation in any given individual’s sleep needs, the National Sleep Foundation (n.d.) cites research to estimate that newborns require the most sleep (between 12 and 18 hours a night) and that this amount declines to just 7–9 hours by the time we are adults.
If you lie down to take a nap and fall asleep very easily, chances are you may have sleep debt. Given that college students are notorious for suffering from significant sleep debt (Hicks et al., 1992; Hicks et al., 2001; Miller et al., 2010), chances are you and your classmates deal with sleep debt– related issues on a regular basis. In 2015, the National Sleep Foundation updated their sleep duration hours, to better accommodate individual differences. Figure \(2\) shows the new recommendations.
Sleep debt and sleep deprivation have significant negative psychological and physiological consequences Figure \(3\). As mentioned earlier, lack of sleep can result in decreased mental alertness and cognitive function. In addition, sleep deprivation often results in depression-like symptoms. These effects can occur as a function of accumulated sleep debt or in response to more acute periods of sleep deprivation. It may surprise you to know that sleep deprivation is associated with obesity, increased blood pressure, increased levels of stress hormones, and reduced immune functioning (Banks & Dinges, 2007). A sleep deprived individual generally will fall asleep more quickly than if she were not sleep deprived. Some sleep-deprived individuals have difficulty staying awake when they stop moving (e.g., sitting and watching television or driving a car). That is why individuals suffering from sleep deprivation can put themselves and others at risk when they get behind the wheel of a car or work with dangerous machinery. Some research suggests that sleep deprivation affects cognitive and motor function as much as, if not more than, alcohol intoxication (Williamson & Feyer, 2000).
LINK TO LEARNING
To assess your own sleeping habits, read this article about sleep needs.
The amount of sleep we get varies across the lifespan. When we are very young, we spend up to 16 hours a day sleeping. As we grow older, we sleep less. In fact, a meta-analysis, which is a study that combines the results of many related studies, conducted within the last decade indicates that by the time we are 65 years old, we average fewer than 7 hours of sleep per day (Ohayon et al., 2004). As the amount of time we sleep varies over our lifespan, presumably the sleep debt would adjust accordingly.
REFERENCES
Banks, S., & Dinges, D. F. (2007). Behavioral and physiological consequences of sleep restriction. Journal of Clinical Sleep Medicine, 3, 519–528. https://doi.org/10.5664/jcsm.26918
Gold, D. R., Rogacz, S. R., Bock, N., Tosteson, T. D., Baum, T. M., Speizer, F. M., & Czeisler, C. A. (1992). Rotating shift work, sleep, and accidents related to sleepiness in hospital nurses. American Journal of Public Health, 82, 1011–1014. https://doi.org/10.2105/AJPH.82.7.1011
Hardeland, R., Pandi-Perumal, S. R., & Cardinali, D. P. (2006). Melatonin. International Journal of Biochemistry & Cell Biology, 38, 313–316. https://doi.org/10.1016/j.biocel.2005.08.020
Hicks, R. A., Fernandez, C., & Pelligrini, R. J. (2001). The changing sleep habits of university students: An update. Perceptual and Motor Skills, 93, 648. https://doi.org/10.2466/pms.2001.93.3.648
Hicks, R. A., Johnson, C., & Pelligrini, R. J. (1992). Changes in the self-reported consistency of normal habitual sleep duration of college students (1978 and 1992). Perceptual and Motor Skills, 75, 1168–1170. https://doi.org/10.2466/pms.1992.75.3f.1168
Huang, L. B., Tsai, M. C., Chen, C. Y., & Hsu, S. C. (2013). The effectiveness of light/dark exposure to treat insomnia in female nurses undertaking shift work during the evening/night shift. Journal of Clinical Sleep Medicine, 9, 641–646. https://doi.org/10.5664/jcsm.2824
Klein, D. C., Moore, R. Y., & Reppert, S. M. (Eds.). (1991). Suprachiasmatic nucleus: The mind’s clock. Oxford University Press.
Miller, N. L., Shattuck, L. G., & Matsangas, P. (2010). Longitudinal study of sleep patterns of United States Military Academy cadets. Sleep, 33, 1623–1631. https://doi.org/10.1093/sleep/33.12.1623
National Sleep Foundation. (n.d.). How much sleep do we really need? www.sleepfoundation.org/arti...much-sleep-do- we-really-need
Ohayon, M. M., Carskadon, M. A., Guilleminault, C., & Vitiello, M. V. (2004). Meta-analysis of quantitative sleep parameters from childhood to old age in healthy individuals: Developing normative sleep values across the human lifespan. Sleep, 27, 1255–1273. https://doi.org/10.1093/sleep/27.7.1255
Presser, H. B. (1995). Job, family, and gender: Determinants of nonstandard work schedules among employed Americans in 1991. Demography, 32, 577–598. https://doi.org/10.2307/2061676
Roth, T. (2007). Insomnia: Definition, prevalence, etiology, and consequences. Journal of Clinical Sleep Medicine, 3(5 Suppl.), S7–S10. https://doi.org/10.5664/jcsm.26929
Taillard, J., Philip, P., Coste, O., Sagaspe, P., & Bioulac, B. (2003). The circadian and homeostatic modulation of sleep pressure during wakefulness differs between morning and evening chronotypes.
Journal of Sleep Research, 12, 275–282. doi.org/10.1046/j.0962-1105.2003.00369.x
Waterhouse, J., Fukuda, Y., & Morita, T. (2012). Daily rhythms of the sleep-wake cycle. Journal of Physiological Anthropology, 31(1), 5. https://doi.org/10.1186/1880-6805-31-5
Welsh, D. K. Takahashi, J. S., & Kay, S. A. (2010). Suprachiasmatic nucleus: Cell autonomy and network properties. Annual Review of Physiology, 72, 551–577. doi.org/10.1146/annurev-physiol- 021909-135919
West, S., Boughton, M., & Byrnes, M. (2009). Juggling multiple temporalities: The shift work story of mid-life nurses. Journal of Nursing Management, 17, 110–119. doi.org/10.1111/ j.1365-2834.2008.00920.x
Williamson, A. M., & Feyer, A. M. (2000). Occupational and Environmental Medicine, 57, 649–655. doi.org/10.1136/ oem.57.10.649 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/09%3A_What_Is_Consciousness/9.01%3A_Introduction.txt |
We spend approximately one-third of our lives sleeping. Given that the average life expectancy for U.S. citizens falls between 73 and 79 years old (Singh & Siahpush, 2006), we can expect to spend approximately 25 years of our lives sleeping. Some animals never sleep (e.g., several fish and amphibian species); other animals can go extended periods of time without sleep and without apparent negative consequences (e.g., dolphins); yet some animals (e.g., rats) die after two weeks of sleep deprivation (Siegel, 2008). Why do we devote so much time to sleeping? Is it absolutely essential that we sleep? This section will consider these questions and explore various explanations for why we sleep.
10.02: What Is Sleep
You have read that sleep is distinguished by low levels of physical activity and reduced sensory awareness. As discussed by Siegel (2008), a definition of sleep must also include mention of the interplay of the circadian and homeostatic mechanisms that regulate sleep. Homeostatic regulation of sleep is evidenced by sleep rebound following sleep deprivation. Sleep rebound refers to the fact that a sleep-deprived individual will tend to take a shorter time to fall asleep during subsequent opportunities for sleep. Sleep is characterized by certain patterns of activity of the brain that can be visualized using electroencephalography (EEG), and different phases of sleep can be differentiated using EEG as well (Figure \(1\)).
Sleep-wake cycles seem to be controlled by multiple brain areas acting in conjunction with one another. Some of these areas include the thalamus, the hypothalamus, and the pons. As already mentioned, the hypothalamus contains the SCN— the biological clock of the body—in addition to other nuclei that, in conjunction with the thalamus, regulate slow-wave sleep. The pons is important for regulating rapid eye movement (REM) sleep (National Institutes of Health, 2007).
Sleep is also associated with the secretion and regulation of a number of hormones from several endocrine glands, including melatonin, follicle stimulating hormone (FSH), luteinizing hormone (LH), and growth hormone (National Institutes of Health, 2007). You have read that the pineal gland releases melatonin during sleep (Figure \(2\)). Melatonin is thought to be involved in the regulation of various biological rhythms and the immune system (Hardeland et al., 2006). During sleep, the pituitary gland secretes both FSH and LH which are important in regulating the reproductive system (Christensen et al., 2012; Sofikitis et al., 2008). The pituitary gland also secretes growth hormone during sleep, which plays a role in physical growth and maturation as well as other metabolic processes (Bartke et al., 2013).
10.03: Why Do We Sleep
Given the central role that sleep plays in our lives and the number of adverse consequences that have been associated with sleep deprivation, one would think that we would have a clear understanding of why it is that we sleep. Unfortunately, this is not the case; however, several hypotheses have been proposed to explain the function of sleep.
Adaptive Function of Sleep
One popular hypothesis of sleep incorporates the perspective of evolutionary psychology. Evolutionary psychology is a discipline that studies how universal patterns of behavior and cognitive processes have evolved over time as a result of natural selection. Variations and adaptations in cognition and behavior make individuals more or less successful in reproducing and passing their genes to their offspring. One hypothesis from this perspective might argue that sleep is essential to restore resources that are expended during the day. Just as bears hibernate in the winter when resources are scarce, perhaps people sleep at night to reduce their energy expenditures. While this is an intuitive explanation of sleep, there is little research that supports this explanation. In fact, it has been suggested that there is no reason to think that energetic demands could not be addressed with periods of rest and inactivity (Frank, 2006; Rial et al., 2007), and some research has actually found a negative correlation between energetic demands and the amount of time spent sleeping (Capellini et al., 2008).
Another evolutionary hypothesis of sleep holds that our sleep patterns evolved as an adaptive response to predatory risks, which increase in darkness. Thus we sleep in safe areas to reduce the chance of harm. Again, this is an intuitive and appealing explanation for why we sleep. Perhaps our ancestors spent extended periods of time asleep to reduce attention to themselves from potential predators. Comparative research indicates, however, that the relationship that exists between predatory risk and sleep is very complex and equivocal. Some research suggests that species that face higher predatory risks sleep fewer hours than other species (Capellini et al., 2008), while other researchers suggest there is no relation- ship between the amount of time a given species spends in deep sleep and its predation risk (Lesku et al., 2006).
It is quite possible that sleep serves no single universally adaptive function, and different species have evolved different patterns of sleep in response to their unique evolutionary pressures. While we have discussed the negative outcomes associated with sleep deprivation, it should be pointed out that there are many benefits that are associated with adequate amounts of sleep. A few such benefits listed by the National Sleep Foundation (n.d.) include maintaining healthy weight, lowering stress levels, improving mood, and increasing motor coordination, as well as a number of benefits related to cognition and memory formation.
Cognitive Function of Sleep
Another theory regarding why we sleep involves sleep’s importance for cognitive function and memory formation (Rattenborg et al., 2007). Indeed, we know sleep deprivation results in disruptions in cognition and memory deficits (Brown, 2012), leading to impairments in our abilities to maintain attention, make decisions, and recall long-term memories. Moreover, these impairments become more severe as the amount of sleep deprivation increases (Alhola & Polo-Kantola, 2007).
Furthermore, slow-wave sleep after learning a new task can improve resultant performance on that task (Huber et al., 2004) and seems essential for effective memory formation (Stickgold, 2005). Understanding the impact of sleep on cognitive function should help you understand that cramming all night for a test may be not effective and can even prove counterproductive.
LINK TO LEARNING
Watch this brief video describing sleep deprivation in college students.
Here’s another brief video describing sleep tips for college students.
Sleep has also been associated with other cognitive benefits. Research indicates that included among these possible benefits are increased capacities for creative thinking (Cai et al., 2009; Wagner et al., 2004), language learning (Fenn et al., 2003; Gómez et al., 2006), and inferential judgments (Ellenbogen et al., 2007). It is possible that even the processing of emotional information is influenced by certain aspects of sleep (Walker, 2009).
LINK TO LEARNING
Watch this brief video describing the relationship between sleep and memory.
REFERENCES
Alhola, P., & Polo-Kantola, P. (2007). Sleep deprivation: Impact on cognitive performance. Neuropsychiatric Disease and Treatment, 3, 553–557.
Bartke, A., Sun, L. Y., & Longo, V. (2013). Somatotropic signaling: Trade-offs between growth, reproductive development, and longevity. Physiological Reviews, 93, 571–598. doi.org/10.1152/ physrev.00006.2012
Brown, L. K. (2012). Can sleep deprivation studies explain why human adults sleep? Current Opinion in Pulmonary Medicine, 18, 541–545. doi.org/10.1097/MCP.0b013e3283596740
Cai, D. J., Mednick, S. A., Harrison, E. M., Kanady, J. C., & Mednick, S. C. (2009). REM, not incubation, improves creativity by priming associative networks. Proceedings of the National Academy of Sciences, 106, 10130–10134. https://doi.org/10.1073/pnas.0900271106
Capellini, I., Barton, R. A., McNamara, P., Preston, B. T., & Nunn, C. L. (2008). Phylogenetic analysis of the ecology and evolution of mammalian sleep. Evolution, 62, 1764–1776. doi.org/10.1111/ j.1558-5646.2008.00392.x
Christensen, A., Bentley, G. E., Cabrera, R., Ortega, H. H., Perfito, N., Wu, T. J., & Micevych, P. (2012). Hormonal regulation of female reproduction. Hormone and Metabolic Research, 44, 587–591. https://doi.org/10.1055/s-0032-1306301
Ellenbogen, J. M., Hu, P. T., Payne, J. D., Titone, D., & Walker, M. P. (2007). Human relational memory requires time and sleep.
Proceedings of the National Academy of Sciences, 104, 7723–7728. https://doi.org/10.1073/pnas.0700094104
Fenn, K. M., Nusbaum, H. C., & Margoliash, D. (2003). Consolidation during sleep of perceptual learning of spoken language. Nature, 425, 614–616. https://doi.org/10.1038/nature01951
Frank, M. G. (2006). The mystery of sleep function: Current perspectives and future directions. Reviews in the Neurosciences, 17, 375–392. https://doi.org/10.1515/revneuro.2006.17.4.375
Gómez, R. L., Bootzin, R. R., & Nadel, L. (2006). Naps promote abstraction in language-learning infants. Psychological Science, 17, 670–674. https://doi.org/10.1111/j.1467-9280.2006.01764.x
Hardeland, R., Pandi-Perumal, S. R., & Cardinali, D. P. (2006). Melatonin. International Journal of Biochemistry & Cell Biology, 38, 313–316. https://doi.org/10.1016/j.biocel.2005.08.020
Huber, R., Ghilardi, M. F., Massimini, M., & Tononi, G. (2004). Local sleep and learning. Nature, 430(6995), 78–81. https://doi.org/10.1038/ nature02663
Lesku, J. A., Roth, T. C., II, Amlaner, C. J., & Lima, S. L. (2006). A phylogenetic analysis of sleep architecture in mammals: The integration of anatomy, physiology, and ecology. The American Naturalist, 168, 441–453. https://doi.org/10.1086/506973
National Institutes of Health. (2007). Information about sleep. In NIH curriculum supplement series [Internet]. www.ncbi.nlm.nih. gov/books/NBK20359/
National Sleep Foundation. (n.d.). How much sleep do we really need? www.sleepfoundation.org/arti...much-sleep-do- we-really-need
Rattenborg, N. C., Lesku, J. A., Martinez-Gonzalez, D., & Lima, S. L. (2007). The non-trivial functions of sleep. Sleep Medicine Reviews, 11, 405–409. https://doi.org/10.1016/j.smrv.2007.04.003
Rial, R. V., Nicolau, M. C., Gamundí, A., Akaârir, M., Aparicio, S., Garau, C., Tejada, S., Roca, C., Gené, L., Moranta, D., & Esteban, S. (2007). The trivial function of sleep. Sleep Medicine Reviews, 11, 311–325. https:// doi.org/10.1016/j.smrv.2007.03.001
Siegel, J. M. (2008). Do all animals sleep? Trends in Neuroscience, 31, 208–213. https://doi.org/10.1016/j.tins.2008.02.001
Singh, G. K., & Siahpush, M. (2006). Widening socioeconomic inequalities in US life expectancy, 1980–2000. International Journal of Epidemiology, 35, 969–979. https://doi.org/10.1093/ije/dyl083
Sofikitis, N., Giotitsas, N., Tsounapi, P., Baltogiannis, D., Giannakis, D., & Pardalidis, N. (2008). Hormonal regulation of spermatogenesis and spermiogenesis. Journal of Steroid Biochemistry and Molecular Biology, 109, 323–330. https://doi.org/10.1016/j.jsbmb.2008.03.004
Stickgold, R. (2005). Sleep-dependent memory consolidation. Nature, 437, 1272–1278. https://doi.org/10.1038/nature04286
Wagner, U., Gais, S., Haider, H., Verleger, R., & Born, J. (2004). Sleep inspires insight. Nature, 427(6972), 352–355. dx.doi. org/10.1038/nature02223
Walker, M. P. (2009). The role of sleep in cognition and emotion. Annals of the New York Academy of Sciences, 1156, 168–197. doi. org/10.1111/j.1749-6632.2009.04416.x | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/10%3A_Sleep_and_Why_We_Sleep/10.01%3A_Introduction.txt |
Sleep is not a uniform state of being. Instead, sleep is com- posed of several different stages that can be differentiated from one another by the patterns of brain wave activity that occur during each stage. These changes in brain wave activity can be visualized using EEG and are distinguished from one another by both the frequency and amplitude of brain waves (Figure \(1\)). Sleep can be divided into two different general phases: REM sleep and non-REM sleep. Rapid eye movement (REM) sleep is characterized by darting movements of the eyes under closed eyelids. Brain waves during REM sleep appear very similar to brain waves during wakefulness. In contrast, non-REM (NREM) sleep is subdivided into four stages distinguished from each other and from wakefulness by characteristic patterns of brain waves. The first four stages of sleep are NREM sleep, while the fifth and final stage of sleep is REM sleep. In this section, we will discuss each of these stages of sleep and their associated patterns of brain wave activity.
11.02: Nrem Stages Of Sleep
The first stage of NREM sleep is known as stage 1 sleep. Stage 1 sleep is a transitional phase that occurs between wakeful- ness and sleep, the period during which we drift off to sleep. During this time, there is a slowdown in both the rates of respiration and heartbeat. In addition, stage 1 sleep involves a marked decrease in both overall muscle tension and core body temperature.
In terms of brain wave activity, stage 1 sleep is associated with both alpha and theta waves. The early portion of stage 1 sleep produces alpha waves, which are relatively low frequency (8–13 Hz), high amplitude patterns of electrical activity (waves) that become synchronized (Figure \(1\)). This pattern of brain wave activity resembles that of someone who is very relaxed, yet awake. As an individual continues through stage 1 sleep, there is an increase in theta wave activity. Theta waves are even lower frequency (4–7 Hz), higher amplitude brain waves than alpha waves. It is relatively easy to wake someone from stage 1 sleep; in fact, people often report that they have not been asleep if they are awoken during stage 1 sleep.
As we move into stage 2 sleep, the body goes into a state of deep relaxation. Theta waves still dominate the activity of the brain, but they are interrupted by brief bursts of activity known as sleep spindles (Figure \(2\)). A sleep spindle is a rapid burst of higher frequency brain waves that may be important for learning and memory (Fogel & Smith, 2011; Poe et al., 2010). In addition, the appearance of K-complexes is often associated with stage 2 sleep. A K-complex is a very high amplitude pattern of brain activity that may in some cases occur in response to environmental stimuli. Thus, K-complexes might serve as a bridge to higher levels of arousal in response to what is going on in our environments (Halász, 1993; Steriade & Amzica, 1998).
Stage 3 and stage 4 of sleep are often referred to as deep sleep or slow-wave sleep because these stages are characterized by low frequency (up to 4 Hz), high amplitude delta waves (Figure \(3\)). During this time, an individual’s heart rate and respiration slow dramatically. It is much more difficult to awaken someone from sleep during stage 3 and stage 4 than during earlier stages. Interestingly, individuals who have increased levels of alpha brain wave activity (more often associated with wakefulness and transition into stage 1 sleep) during stage 3 and stage 4 often report that they do not feel refreshed upon waking, regardless of how long they slept (Stone et al., 2008). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/11%3A_Stages_of_Sleep/11.01%3A_Introduction.txt |
As mentioned earlier, REM sleep is marked by rapid movements of the eyes. The brain waves associated with this stage of sleep are very similar to those observed when a person is awake, as shown in Figure \(1\), and this is the period of sleep in which dreaming occurs. It is also associated with paralysis of muscle systems in the body with the exception of those that make circulation and respiration possible. Therefore, no movement of voluntary muscles occurs during REM sleep in a normal individual; REM sleep is often referred to as paradoxical sleep because of this combination of high brain activity and lack of muscle tone. Like NREM sleep, REM has been implicated in various aspects of learning and memory (Wagner et al., 2001), although there is disagreement within the scientific community about how important both NREM and REM sleep are for normal learning and memory (Siegel, 2001).
If people are deprived of REM sleep and then allowed to sleep without disturbance, they will spend more time in REM sleep in what would appear to be an effort to recoup the lost time in REM. This is known as the REM rebound, and it suggests that REM sleep is also homeostatically regulated. Aside from the role that REM sleep may play in processes related to learning and memory, REM sleep may also be involved in emotional processing and regulation. In such instances, REM rebound may actually represent an adaptive response to stress in nondepressed individuals by suppressing the emotional salience of aversive events that occurred in wakefulness (Suchecki et al., 2012).
While sleep deprivation in general is associated with a number of negative consequences (Brown, 2012), the con- sequences of REM deprivation appear to be less profound (as discussed in Siegel, 2001). In fact, some have suggested that REM deprivation can actually be beneficial in some circumstances. For instance, REM sleep deprivation has been demonstrated to improve symptoms of people suffering from major depression, and many effective antidepressant medications suppress REM sleep (Riemann et al., 2001; Vogel, 1975).
It should be pointed out that some reviews of the literature challenge this finding, suggesting that sleep deprivation that is not limited to REM sleep is just as effective or more effective at alleviating depressive symptoms among some patients suffering from depression. In either case, why sleep deprivation improves the mood of some patients is not entirely understood (Giedke & Schwärzler, 2002). Recently, however, some have suggested that sleep deprivation might change emotional processing so that various stimuli are more likely to be perceived as positive in nature (Gujar et al., 2011).
The hypnogram below (Figure \(2\)) shows a person’s passage through the stages of sleep.
LINK TO LEARNING
View this video that describes the various stages of sleep.
Dreams
The meaning of dreams varies across different cultures and periods of time. By the late nineteenth century, German psychiatrist Sigmund Freud had become convinced that dreams represented an opportunity to gain access to the unconscious. By analyzing dreams, Freud thought people could increase self-awareness and gain valuable insight to help them deal with the problems they faced in their lives. Freud made distinctions between the manifest content and the latent content of dreams. Manifest content is the actual content, or storyline, of a dream. Latent content, on the other hand, refers to the hidden meaning of a dream. For instance, if a woman dreams about being chased by a snake, Freud might have argued that this represents the woman’s fear of sexual intimacy, with the snake serving as a symbol of a man’s penis.
Freud was not the only theorist to focus on the content of dreams. The twentieth century Swiss psychiatrist Carl Jung believed that dreams allowed us to tap into the collective unconscious. The collective unconscious, as described by Jung, is a theoretical repository of information he believed to be shared by everyone. According to Jung, certain symbols in dreams reflected universal archetypes with meanings that are similar for all people regardless of culture or location.
The sleep and dreaming researcher Rosalind Cartwright, however, believes that dreams simply reflect life events that are important to the dreamer. Unlike Freud and Jung, Cart- wright’s ideas about dreaming have found empirical support. For example, she and her colleagues published a study in which women going through divorce were asked several times over a five-month period to report the degree to which their former spouses were on their minds. These same women were awakened during REM sleep in order to provide a detailed account of their dream content. There was a significant positive correlation between the degree to which women thought about their former spouses during waking hours and the number of times their former spouses appeared as characters in their dreams (Cartwright et al., 2006). Recent research (Horikawa et al., 2013) has uncovered new techniques by which researchers may effectively detect and classify the visual images that occur during dreaming by using fMRI for neural measurement of brain activity patterns, opening the way for additional research in this area.
Recently, neuroscientists have also become interested in understanding why we dream. For example, John Hob- son (2009) suggests that dreaming may represent a state of protoconsciousness. In other words, dreaming involves constructing a virtual reality in our heads that we might use to help us during wakefulness. Among a variety of neurobiological evidence, Hobson cites research on lucid dreams as an opportunity to better understand dreaming in general. Lucid dreams are dreams in which certain aspects of wakefulness are maintained during a dream state. In a lucid dream, a per- son becomes aware of the fact that they are dreaming, and as such, they can control the dream’s content (LaBerge, 1990).
REFERENCES
Brown, L. K. (2012). Can sleep deprivation studies explain why human adults sleep? Current Opinion in Pulmonary Medicine, 18, 541–545. doi.org/10.1097/MCP.0b013e3283596740
Cartwright, R., Agargun, M. Y., Kirkby, J., & Friedman, J. K. (2006). Relation of dreams to waking concerns. Psychiatry Research, 141, 261–270. https://doi.org/10.1016/j.psychres.2005.05.013
Fogel, S. M., & Smith, C. T. (2011). The function of the sleep spindle: A physiological index of intelligence and a mechanism for sleep- dependent memory consolidation. Neuroscience & Biobehavioral Reviews, 35, 1154–1165. doi.org/10.1016/j.neubiorev. 2010.12.003
Giedke, H., & Schwärzler, F. (2002). Therapeutic use of sleep deprivation in depression. Sleep Medicine Reviews, 6, 361–377. https://doi.org/ 10.1053/smrv.2002.0235
Gujar, N., Yoo, S., Hu, P., & Walker, M. P. (2011). Sleep deprivation amplifies reactivity of brain reward networks, biasing the appraisal of positive emotional experiences. Journal of Neuroscience, 31, 4466–4474. https://doi.org/10.1523/JNEUROSCI.3220-10.2011
Halász, P. (1993). Arousals without awakening—Dynamic aspect of sleep. Physiology & Behavior, 54, 795–802. doi.org/10.1016/ 0031-9384(93)90094-V
Hobson, J. A. (2009). REM sleep and dreaming: Towards a theory of protoconsciousness. Nature Reviews Neuroscience, 10, 803–814. https://doi.org/10.1038/nrn2716
Horikawa, T., Tamaki, M., Miyawaki, Y., & Kamitani, Y. (2013). Neural decoding of visual imagery during sleep. Science, 340(6132), 639–642. https://doi.org/10.1126/science.1234330
LaBerge, S. (1990). Lucid dreaming: Psychophysiological studies of consciousness during REM sleep. In R. R. Bootzen, J. F. Kihlstrom, & D. L. Schacter (Eds.), Sleep and cognition (pp. 109–126). American Psychological Association.
Poe, G. R., Walsh, C. M., & Bjorness, T. E. (2010). Cognitive neuroscience of sleep. Progress in Brain Research, 185, 1–19. https://doi.org/ 10.1016/B978-0-444-53702-7.00001-4
Riemann, D., Berger, M., & Volderholzer, U. (2001). Sleep and depression—Results from psychobiological studies: An overview. Biological Psychology, 57, 67–103. doi.org/10.1016/ S0301-0511(01)00090-4
Siegel, J. M. (2001). The REM sleep-memory consolidation hypothesis. Science, 294, 1058–1063. https://doi.org/10.1126/science.1063049
Steriade, M., & Amzica, F. (1998). Slow sleep oscillation, rhythmic K-complexes, and their paroxysmal developments. Journal of Sleep Research, 7(1 Suppl.), 30–35. doi.org/10.1046/ j.1365-2869.7.s1.4.x
Stone, K. C., Taylor, D. J., McCrae, C. S., Kalsekar, A., & Lichstein, K. L. (2008). Nonrestorative sleep. Sleep Medicine Reviews, 12, 275–288. https://doi.org/10.1016/j.smrv.2007.12.002
Suchecki, D., Tiba, P. A., & Machado, R. B. (2012). REM sleep rebound as an adaptive response to stressful situations. Frontiers in Neuroscience, 3, 1–12. https://doi.org/10.3389/fneur.2012.00041
Vogel, G. W. (1975). A review of REM sleep deprivation. Archives of General Psychiatry, 32, 749–761. doi.org/10.1001/ archpsyc.1975.01760240077006
Wagner, U., Gais, S., & Born, J. (2001). Emotional memory formation is enhanced across sleep intervals with high amounts of rapid eye movement sleep. Learning & Memory, 8, 112–119. doi. org/10.1101/lm.36801 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/11%3A_Stages_of_Sleep/11.03%3A_REM_Sleep.txt |
Many people experience disturbances in their sleep at some point in their lives. Depending on the population and sleep disorder being studied, between 30% and 50% of the population suffers from a sleep disorder at some point in their lives (Bixler et al., 1979; Hossain & Shapiro, 2002; Ohayon, 1997, 2002; Ohayon & Roth, 2002). This section will describe several sleep disorders as well as some of their treatment options.
12.02: Insomnia
Insomnia, a consistent difficulty in falling or staying asleep, is the most common of the sleep disorders. Individuals with insomnia often experience long delays between the times that they go to bed and actually fall asleep. In addition, these individuals may wake up several times during the night only to find that they have difficulty getting back to sleep. As mentioned earlier, one of the criteria for insomnia involves experiencing these symptoms for at least three nights a week for at least one month’s time (Roth, 2007).
It is not uncommon for people suffering from insomnia to experience increased levels of anxiety about their inability to fall asleep. This becomes a self-perpetuating cycle because increased anxiety leads to increased arousal, and higher levels of arousal make the prospect of falling asleep even more unlikely. Chronic insomnia is almost always associated with feeling over- tired and may be associated with symptoms of depression.
There may be many factors that contribute to insomnia, including age, drug use, exercise, mental status, and bedtime routines. Not surprisingly, insomnia treatment may take one of several different approaches. People who suffer from insomnia might limit their use of stimulant drugs (such as caffeine) or increase their amount of physical exercise during the day. Some people might turn to over-the-counter (OTC) or prescribed sleep medications to help them sleep, but this should be done sparingly because many sleep medications result in dependence and alter the nature of the sleep cycle, and they can increase insomnia over time. Those who continue to have insomnia, particularly if it affects their quality of life, should seek professional treatment.
Some forms of psychotherapy, such as cognitive-behavioral therapy, can help sufferers of insomnia. Cognitive-behavioral therapy is a type of psychotherapy that focuses on cognitive processes and problem behaviors. The treatment of insomnia likely would include stress management techniques and changes in problematic behaviors that could contribute to insomnia (e.g., spending more waking time in bed). Cognitive-behavioral therapy has been demonstrated to be quite effective in treating insomnia (Savard et al., 2005; Williams et al., 2013).
12.03: Parasomnias
A parasomnia is one of a group of sleep disorders in which unwanted, disruptive motor activity and/or experiences during sleep play a role. Parasomnias can occur in either REM or NREM phases of sleep. Sleepwalking, restless leg syndrome, and night terrors are all examples of parasomnias (Mahowald & Schenck, 2000).
Sleepwalking
In sleepwalking, or somnambulism, the sleeper engages in relatively complex behaviors ranging from wandering about to driving an automobile. During periods of sleepwalking, sleepers often have their eyes open, but they are not responsive to attempts to communicate with them. Sleepwalking most often occurs during slow-wave sleep, but it can occur at any time during a sleep period in some affected individuals (Mahowald & Schenck, 2000).
DIG DEEPER
A Sleepwalking Defense?
On January 16, 1997, Scott Falater sat down to dinner with his wife and children and told them about difficulties he was experiencing on a project at work. After dinner, he prepared some materials to use in leading a church youth group the following morning, and then he attempted to repair the family’s swimming pool pump before retiring to bed. The following morning, he awoke to barking dogs and unfamiliar voices from downstairs. As he went to investigate what was going on, he was met by a group of police officers who arrested him for the murder of his wife (Cartwright, 2004; CNN, 1999).
yarmila Falater’s body was found in the family’s pool with 44 stab wounds. A neighbor called the police after witnessing Falater standing over his wife’s body before dragging her into the pool. Upon a search of the premises, police found blood-stained clothes and a bloody knife in the trunk of Falater’s car, and he had blood stains on his neck.
Remarkably, Falater insisted that he had no recollection of hurting his wife in any way. His children and his wife’s parents all agreed that Falater had an excellent relationship with his wife and they couldn’t think of a reason that would provide any sort of motive to murder her (Cartwright, 2004).
Scott Falater had a history of regular episodes of sleepwalking as a child, and he had even behaved violently toward his sister once when she tried to prevent him from leaving their home in his pajamas during a sleepwalking episode. He suffered from no apparent anatomical brain anomalies or psychological disorders. It appeared that Scott Falater had killed his wife in his sleep, or at least, that is the defense he used when he was tried for his wife’s murder (Cartwright, 2004; CNN, 1999). In Falater’s case, a jury found him guilty of first degree murder in June of 1999 (CNN, 1999); however, there are other murder cases where the sleepwalking defense has been used successfully. As scary as it sounds, many sleep researchers believe that homicidal sleepwalking is possible in individuals suffering from the types of sleep disorders described in this module (Broughton et al., 1994; Cartwright, 2004; Mahowald et al., 2005; Pressman, 2007). ■
REM Sleep Behavior Disorder (RBD)
REM sleep behavior disorder (RBD) occurs when the muscle paralysis associated with the REM sleep phase does not occur. Individuals who suffer from RBD have high levels of physical activity during REM sleep, especially during disturbing dreams. These behaviors vary widely, but they can include kicking, punching, scratching, yelling, and behaving like an animal that has been frightened or attacked. People who suffer from this disorder can injure themselves or their sleeping partners when engaging in these behaviors. Furthermore, these types of behaviors ultimately disrupt sleep, although affected individuals have no memories that these behaviors have occurred (Arnulf, 2012).
This disorder is associated with a number of neurodegenerative diseases such as Parkinson’s disease. In fact, this relationship is so robust that some view the presence of RBD as a potential aid in the diagnosis and treatment of a number of neurodegenerative diseases (Ferini-Strambi, 2011). Clonazepam, an anti-anxiety medication with sedative properties, is most often used to treat RBD. It is administered alone or in conjunction with doses of melatonin (the hormone secreted by the pineal gland). As part of treatment, the sleeping environment is often modified to make it a safer place for those suffering from RBD (Zanigni et al., 2011).
Other Parasomnias
A person with restless leg syndrome has uncomfortable sensations in the legs during periods of inactivity or when trying to fall asleep. This discomfort is relieved by deliberately moving the legs, which, not surprisingly, contributes to difficulty in falling or staying asleep. Restless leg syndrome is quite common and has been associated with a number of other medical diagnoses, such as chronic kidney disease and diabetes (Mahowald & Schenck, 2000). There are a variety of drugs that treat restless leg syndrome: benzodiazepines, opiates, and anticonvulsants (Restless Legs Syndrome Foundation, n.d.).
Night terrors result in a sense of panic in the sufferer and are often accompanied by screams and attempts to escape from the immediate environment (Mahowald & Schenck, 2000). Although individuals suffering from night terrors appear to be awake, they generally have no memories of the events that occurred, and attempts to console them are ineffective. Typically, individuals suffering from night terrors will fall back asleep again within a short time. Night terrors apparently occur during the NREM phase of sleep (Provini et al., 2011). Generally, treatment for night terrors is unnecessary unless there is some underlying medical or psychological condition that is contributing to the night terrors (Mayo Clinic, n.d.). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/12%3A_Sleep_Problems_and_Disorders/12.01%3A_Introduction.txt |
Sleep apnea is defined by episodes during which a sleeper’s breathing stops. These episodes can last 10–20 seconds or longer and often are associated with brief periods of arousal. While individuals suffering from sleep apnea may not be aware of these repeated disruptions in sleep, they do experience increased levels of fatigue. Many individuals diagnosed with sleep apnea first seek treatment because their sleeping partners indicate that they snore loudly and/or stop breathing for extended periods of time while sleeping (Henry & Rosenthal, 2013). Sleep apnea is much more common in overweight people and is often associated with loud snoring. Surprisingly, sleep apnea may exacerbate cardiovascular disease (Sánchezde-la-Torre et al., 2012). While sleep apnea is less common in thin people, anyone, regardless of their weight, who snores loudly or gasps for air while sleeping should be checked for sleep apnea.
While people are often unaware of their sleep apnea, they are keenly aware of some of the adverse consequences of insufficient sleep. Consider a patient who believed that as a result of his sleep apnea he “had three car accidents in six weeks. They were ALL my fault. Two of them I didn’t even know I was involved in until afterwards” (Henry & Rosenthal, 2013, p. 52). It is not uncommon for people suffering from undiagnosed or untreated sleep apnea to fear that their careers will be affected by the lack of sleep, illustrated by this statement from another patient: “I’m in a job where there’s a premium on being mentally alert. I was really sleepy and having trouble concentrating. It was getting to the point where it was kind of scary” (Henry & Rosenthal, 2013, p. 52).
There are two types of sleep apnea: obstructive sleep apnea and central sleep apnea. Obstructive sleep apnea occurs when an individual’s airway becomes blocked during sleep, and air is prevented from entering the lungs. In central sleep apnea, disruption in signals sent from the brain that regulate breathing cause periods of interrupted breathing (White, 2005).
One of the most common treatments for sleep apnea involves the use of a special device during sleep. A continuous positive airway pressure (CPAP) device includes a mask that fits over the sleeper’s nose and mouth, which is connected to a pump that pumps air into the person’s airways, forcing them to remain open, as shown in Figure \(1\). Some newer CPAP masks are smaller and cover only the nose. This treatment option has proven to be effective for people suffering from mild to severe cases of sleep apnea (McDaid et al., 2009). However, alternative treatment options are being explored because consistent compliance by users of CPAP devices is a problem. Recently, a new EPAP (expiratory positive air pressure) device has shown promise in double-blind trials as one such alternative (Berry et al., 2011).
12.05: SIDS
In sudden infant death syndrome (SIDS) an infant stops breathing during sleep and dies. Infants younger than 12 months appear to be at the highest risk for SIDS, and boys have a greater risk than girls. A number of risk factors have been associated with SIDS, including premature birth, smoking within the home, and hyperthermia. There may also be differences in both brain structure and function in infants that die from SIDS (Berkowitz, 2012; Mage & Donner, 2006; Thach, 2005).
The substantial amount of research on SIDS has led to a number of recommendations to parents to protect their children (Figure \(1\)). For one, research suggests that infants should be placed on their backs when put down to sleep, and their cribs should not contain any items that pose suffocation threats, such as blankets, pillows, or padded crib bumpers (cushions that cover the bars of a crib). Infants should not have caps placed on their heads when put down to sleep in order to prevent overheating, and people in the child’s house- hold should abstain from smoking in the home. Recommendations like these have helped to decrease the number of infant deaths from SIDS in recent years (Mitchell, 2009; Task Force on Sudden Infant Death Syndrome, 2011). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/12%3A_Sleep_Problems_and_Disorders/12.04%3A_Sleep_Apnea.txt |
Unlike the other sleep disorders described in this section, a person with narcolepsy cannot resist falling asleep at inopportune times. These sleep episodes are often associated with cataplexy, which is a lack of muscle tone or muscle weakness, and in some cases involves complete paralysis of the voluntary muscles. This is similar to the kind of paralysis experienced by healthy individuals during REM sleep (Burgess & Scammell, 2012; Hishikawa & Shimizu, 1995; Luppi et al., 2011). Narcoleptic episodes take on other features of REM sleep. For example, around one third of individuals diagnosed with narcolepsy experience vivid, dream-like hallucinations during narcoleptic attacks (Chokroverty, 2010).
Surprisingly, narcoleptic episodes are often triggered by states of heightened arousal or stress. The typical episode can last from a minute or two to half an hour. Once awakened from a narcoleptic attack, people report that they feel refreshed (Chokroverty, 2010). Obviously, regular narcoleptic episodes could interfere with the ability to perform one’s job or complete schoolwork, and in some situations, narcolepsy can result in significant harm and injury (e.g., driving a car or operating machinery or other potentially dangerous equipment).
Generally, narcolepsy is treated using psychomotor stimulant drugs, such as amphetamines (Mignot, 2012). These drugs promote increased levels of neural activity. Narcolepsy is associated with reduced levels of the signaling molecule hypocretin in some areas of the brain (De la Herrán-Arita & Druck- er-Colín, 2012; Han, 2012), and the traditional stimulant drugs do not have direct effects on this system. Therefore, it is quite likely that new medications that are developed to treat narcolepsy will be designed to target the hypocretin system.
There is a tremendous amount of variability among sufferers, both in terms of how symptoms of narcolepsy manifest and the effectiveness of currently available treatment options. This is illustrated by McCarty’s (2010) case study of a 50-year- old woman who sought help for the excessive sleepiness during normal waking hours that she had experienced for several years. She indicated that she had fallen asleep at inappropriate or dangerous times, including while eating, while socializing with friends, and while driving her car. During periods of emotional arousal, the woman complained that she felt some weakness in the right side of her body. Although she did not experience any dream-like hallucinations, she was diagnosed with narcolepsy as a result of sleep testing. In her case, the fact that her cataplexy was confined to the right side of her body was quite unusual. Early attempts to treat her condition with a stimulant drug alone were unsuccessful. How- ever, when a stimulant drug was used in conjunction with a popular antidepressant, her condition improved dramatically.
REFERENCES
Arnulf, I. (2012). REM sleep behavior disorder: Motor manifestations and pathophysiology. Movement Disorders, 27, 677–689. doi. org/10.1002/mds.24957
Berkowitz, C. D. (2012). Sudden infant death syndrome, sudden unexpected infant death, and apparent life-threatening events. Advances in Pediatrics, 59, 183–208. https://doi.org/10.1016/j.yapd.2012.04.011
Berry, R. B., Kryger, M. H., & Massie, C. A. (2011). A novel nasal expiratory positive airway pressure (EPAP) device for the treatment of obstructive sleep apnea: A randomized controlled trial. Sleep, 34, 479–485. https://doi.org/10.1093/sleep/34.4.479
Bixler, E. O., Kales, A., Soldatos, C. R., Kales, J. D., & Healey, S. (1979). Prevalence of sleep disorders in the Los Angeles metropolitan area. American Journal of Psychiatry, 136, 1257–1262.
Broughton, R., Billings, R., Cartwright, R., Doucette, D., Edmeads, J., Edwardh, M., Ervin, F., Orchard, B., Hill, R., & Turrell, G. (1994). Homicidal somnambulism: A case report. Sleep, 17, 253–264. https:// doi.org/10.1093/sleep/17.3.253
Burgess, C. R., & Scammell, T. E. (2012). Narcolepsy: Neural mechanisms of sleepiness and cataplexy. Journal of Neuroscience, 32, 12305– 12311. https://doi.org/10.1523/JNEUROSCI.2630-12.2012
Cartwright, R. (2004). Sleepwalking violence: A sleep disorder, a legal dilemma, and a psychological challenge. American Journal of Psychiatry, 161, 1149–1158. https://doi.org/10.1176/appi.ajp.161.7.1149
Chokroverty, S. (2010). Overview of sleep & sleep disorders. Indian Journal of Medical Research, 131, 126–140.
CNN. (1999, June 25). ‘Sleepwalker’ convicted of murder. www. cnn.com/US/9906/25/sleepwalker.01/
De la Herrán-Arita, A. K., & Drucker-Colín, R. (2012). Models for narcolepsy with cataplexy drug discovery. Expert Opinion on Drug Discovery, 7, 155–164. https://doi.org/10.1517/17460441.2012.651454
Ferini-Strambi, L. (2011). Does idiopathic REM sleep behavior disorder (iRBD) really exist? What are the potential markers of neurodegeneration in iRBD? Sleep Medicine, 12(2 Suppl.), S43–S49. https://doi.org/10.1016/j.sleep.2011.10.010
Guilleminault, C., Kirisoglu, C., Bao, G., Arias, V., Chan, A., & Li, K. K. (2005). Adult chronic sleepwalking and its treatment based on polysomnography. Brain, 128, 1062–1069. doi.org/10.1093/ brain/awh481
Han, F. (2012). Sleepiness that cannot be overcome: Narcolepsy and cataplexy. Respirology, 17, 1157–1165. doi. org/10.1111/j.1440-1843.2012.02178.x
Henry, D., & Rosenthal, L. (2013). “Listening for his breath:” The significance of gender and partner reporting on the diagnosis, management, and treatment of obstructive sleep apnea. Social Science & Medicine, 79, 48–56. https://doi.org/10.1016/j.socscimed.2012.05.021
Hishikawa, Y., & Shimizu, T. (1995). Physiology of REM sleep, cataplexy, and sleep paralysis. Advances in Neurology, 67, 245–271.
Hossain, J. L., & Shapiro, C. M. (2002). The prevalence, cost implications, and management of sleep disorders: An overview. Sleep and Breathing, 6, 85–102. https://doi.org/10.1055/s-2002-32322
Luppi, P.-H., Clément, O., Sapin, E., Gervasoni, D., Peyron, C., Léger, L., Salvert, D., & Fort, P. (2011). The neuronal network responsible for paradoxical sleep and its dysfunctions causing narcolepsy and rapid eye movement (REM) behavior disorder. Sleep Medicine Reviews, 15, 153–163. https://doi.org/10.1016/j.smrv.2010.08.002
Mage, D. T., & Donner, M. (2006). Female resistance to hypoxia: Does it explain the sex difference in mortality rates? Journal of Women’s Health, 15, 786–794. https://doi.org/10.1089/jwh.2006.15.786
Mahowald, M. W., & Schenck, C. H. (2000). Diagnosis and management of parasomnias. Clinical Cornerstone, 2, 48–54. doi. org/10.1016/S1098-3597(00)90040-1
Mahowald, M. W., Schenck, C. H., & Cramer Bornemann, M. A. (2005). Sleep-related violence. Current Neurology and Neuroscience Reports, 5, 153–158. https://doi.org/10.1007/s11910-005-0014-3
Mayo Clinic. (n.d.). Sleep terrors (night terrors): Diagnosis and treatment. www.mayoclinic.org/diseases-c...errors/basics/ treatment/con-20032552
McCarty, D. E. (2010). A case of narcolepsy with strictly unilateral cataplexy. Journal of Clinical Sleep Medicine, 15, 75–76. doi. org/10.5664/jcsm.27714
McDaid, C., Durée, K. H., Griffin, S. C., Weatherly, H. L. A., Stradling, J. R., Davies, R. J. O., Sculpher, M. J., & Westwood, M. E. (2009). A systematic review of continuous positive airway pressure for obstructive sleep apnoea-hypopnoea syndrome. Sleep Medicine Reviews, 13, 427–436. https://doi.org/10.1016/j.smrv.2009.02.004
Mignot, E. J. M. (2012). A practical guide to the therapy of narcolepsy and hypersomnia syndromes. Neurotherapeutics, 9, 739–752. doi. org/10.1007/s13311-012-0150-9
Mitchell, E. A. (2009). SIDS: Past, present and future. Acta Paediatrica, 98, 1712–1719. doi.org/10.1111/j.1651-2227.2009.01503.x
Ohayon, M. M. (1997). Prevalence of DSM-IV diagnostic criteria of insomnia: Distinguishing insomnia related to mental disorders from sleep disorders. Journal of Psychiatric Research, 31, 333–346. https://doi.org/10.1016/S0022-3956(97)00002-2
Ohayon, M. M. (2002). Epidemiology of insomnia: What we know and what we still need to learn. Sleep Medicine Reviews, 6, 97–111. https://doi.org/10.1053/smrv.2002.0186
Ohayon, M. M., & Roth, T. (2002). Prevalence of restless legs syndrome and periodic limb movement disorder in the general population. Journal of Psychosomatic Research, 53, 547–554. doi. org/10.1016/S0022-3999(02)00443-9
Pressman, M. R. (2007). Disorders of arousal from sleep and violent behavior: The role of physical contact and proximity. Sleep, 30, 1039–1047. https://doi.org/10.1093/sleep/30.8.1039
Provini, F., Tinuper, P., Bisulli, F., & Lugaresi, E. (2011). Arousal disorders. Sleep Medicine, 12(2 Suppl.), S22–S26. doi.org/10.1016/j. sleep.2011.10.007
Restless Legs Syndrome Foundation. (n.d.). Restless legs syndrome: Causes, diagnosis, and treatment for the patient living with restless legs syndrome (RSL) [Brochure]. www.rls.org/ understanding-rls
Roth, T. (2007). Insomnia: Definition, prevalence, etiology, and consequences. Journal of Clinical Sleep Medicine, 3(5 Suppl.), S7–S10. https://doi.org/10.5664/jcsm.26929
Sánchez-de-la-Torre, M., Campos-Rodriguez, F., & Barbé, F. (2012). Obstructive sleep apnoea and cardiovascular disease. The Lancet Respiratory Medicine, 1, 31–72. doi.org/10.1016/ S2213-2600(12)70051-6
Savard, J., Simard, S., Ivers, H., & Morin, C. M. (2005). Randomized study on the efficacy of cognitive-behavioral therapy for insomnia secondary to breast cancer, part I: Sleep and psychological effects. Journal of Clinical Oncology, 23, 6083–6096. doi.org/10.1200/ JCO.2005.09.548
Task Force on Sudden Infant Death Syndrome. (2011). SIDS and other sleep-related infant deaths: Expansion of recommendations for a safe infant sleeping environment. Pediatrics, 128, 1030–1039. https:// doi.org/10.1542/peds.2011-2285
Thach, B. T. (2005). The role of respiratory control disorders in SIDS. Respiratory Physiology & Neurobiology, 149, 343–353. doi. org/10.1016/j.resp.2005.06.011
White, D. P. (2005). Pathogenesis of obstructive and central sleep apnea. American Journal of Respiratory and Critical Care Medicine, 172, 1363–1370. https://doi.org/10.1164/rccm.200412-1631SO
Williams, J., Roth, A., Vatthauer, K., & McCrae, C. S. (2013). Cognitive behavioral treatment of insomnia. Chest, 143, 554–565. doi. org/10.1378/chest.12-0731
Zanigni, S., Calandra-Buonaura, G., Grimaldi, D., & Cortelli, P. (2011). REM behaviour disorder and neurodegenerative diseases. Sleep Medicine, 12(2 Suppl.), S54–S58. https://doi.org/10.1016/j.sleep.2011.10.012 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/12%3A_Sleep_Problems_and_Disorders/12.06%3A_Narcolepsy.txt |
Whereas other animals rely primarily on hearing, smell, or touch to understand the world around them, human beings rely in large part on vision. A large part of our cerebral cortex is devoted to seeing, and we have substantial visual skills. Seeing begins when light falls on the eyes, initiating the process of transduction. Once this visual information reaches the visual cortex, it is processed by a variety of neurons that detect colors, shapes, and motion and that create meaningful perceptions out of the incoming stimuli.
The air around us is filled with a sea of electromagnetic energy; pulses of energy waves that can carry information from place to place. As you can see in Figure \(1\), electromagnetic waves vary in their wavelength—the distance between one wave peak and the next wave peak—with the shortest gamma waves being only a fraction of a millimeter in length and the longest radio waves being hundreds of kilometers long. Humans are blind to almost all of this energy; our eyes detect only the range from about 400 to 700 billionths of a meter, the part of the electromagnetic spectrum known as the visible spectrum.
13.02: The Sensing Eye And The Perceiving Visual Cortex
As you can see in Figure \(1\), light enters the eye through the cornea, a clear covering that protects the eye and begins to focus the incoming light. The light then passes through the pupil, a small opening in the center of the eye. The pupil is surrounded by the iris, the colored part of the eye that controls the size of the pupil by constricting or dilating in response to light intensity. When we enter a dark movie theater on a sunny day, for instance, muscles in the iris open the pupil and allow more light to enter. Complete adaptation to the dark may take up to 20 minutes.
Behind the pupil is the lens, a structure that focuses the incoming light on the retina, the layer of tissue at the back of the eye that contains photoreceptor cells. As our eyes move from near objects to distant objects, a process known as visual accommodation occurs. Visual accommodation is the process of changing the curvature of the lens to keep the light entering the eye focused on the retina. Rays from the top of the image strike the bottom of the retina and vice versa, and rays from the left side of the image strike the right part of the retina and vice versa, causing the image on the retina to be upside down and backward. Furthermore, the image projected on the retina is flat, and yet our final perception of the image will be three dimensional.
Accommodation is not always perfect, and in some cases the light that is hitting the retina is a bit out of focus. As you can see in Figure \(2\), if the focus is in front of the retina, we say that the person is nearsighted, and when the focus is behind the retina we say that the person is farsighted. Eye- glasses and contact lenses correct this problem by adding another lens in front of the eye, and laser eye surgery corrects the problem by reshaping the eye’s own lens.
The retina contains layers of neurons specialized to respond to light (see Figure \(3\)). As light falls on the retina, it first activates receptor cells known as rods and cones. The activation of these cells then spreads to the bipolar cells and then to the ganglion cells, which gather together and converge, like the strands of a rope, forming the optic nerve. The optic nerve is a collection of millions of ganglion neurons that sends vast amounts of visual information, via the thalamus, to the brain. Because the retina and the optic nerve are active processors and analyzers of visual information, it is not inappropriate to think of these structures as an extension of the brain itself.
Rods are visual neurons that specialize in detecting black, white, and gray colors. There are about 120 million rods in each eye. The rods do not provide a lot of detail about the images we see, but because they are highly sensitive to shorter-waved (darker) and weak light, they help us see in dim light, for instance, at night. Because the rods are located primarily around the edges of the retina, they are particularly active in peripheral vision (when you need to see something at night, try looking away from what you want to see). Cones are visual neurons that are specialized in detecting fine detail and colors. The 5 million or so cones in each eye enable us to see in color, but they operate best in bright light. The cones are located primarily in and around the fovea, which is the central point of the retina.
To demonstrate the difference between rods and cones in attention to detail, choose a word in this text and focus on it. Do you notice that the words a few inches to the side seem more blurred? This is because the word you are focusing on strikes the detail-oriented cones, while the words surrounding it strike the
less-detail-oriented rods, which are located on the periphery.
As you can see in Figure \(5\), the sensory information received by the retina is relayed through the thalamus to corresponding areas in the visual cortex, which is located in the occipital lobe at the back of the brain. Although the principle of contralateral control might lead you to expect that the left eye would send information to the right brain hemisphere and vice versa, nature is smarter than that. In fact, the left and right eyes each send information to both the left and the right hemisphere, and the visual cortex processes each of the cues separately and in parallel. This is an adaptational advantage to an organism that loses sight in one eye, because even if only one eye is functional, both hemispheres will still receive input from it.
The visual cortex is made up of specialized neurons that turn the sensations they receive from the optic nerve into meaningful images. Because there are no photoreceptor cells at the place where the optic nerve leaves the retina, a hole or blind spot in our vision is created (see Figure \(6\)). When both of our eyes are open, we don’t experience a problem because our eyes are constantly moving, and one eye makes up for what the other eye misses. But the visual system is also designed to deal with this problem if only one eye is open—the visual cortex simply fills in the small hole in our vision with similar patterns from the surrounding areas, and we never notice the difference. The ability of the visual system to cope with the blind spot is another example of how sensation and perception work together to create meaningful experience.
Perception is created in part through the simultaneous action of thousands of feature detector neurons—specialized neurons, located in the visual cortex, that respond to the strength, angles, shapes, edges, and movements of a visual stimulus (Kelsey, 1997; Livingstone & Hubel, 1988). The feature detectors work in parallel, each performing a specialized function. When faced with a red square, for instance, the parallel line feature detectors, the horizontal line feature detectors, and the red color feature detectors all become activated. This activation is then passed on to other parts of the visual cortex where other neurons compare the information supplied by the feature detectors with images stored in memory. Suddenly, in a flash of recognition, the many neurons fire together, creating the single image of the red square that we experience (Rodriguez et al., 1999).
Some feature detectors are tuned to selectively respond to particularly important objects, for instance, faces, smiles, and other parts of the body (Downing et al., 2001; Haxby et al., 2001). When researchers disrupted face recognition areas of the cortex using the magnetic pulses of transcranial magnetic stimulation (TMS), people were temporarily unable to recognize faces, and yet they were still able to recognize houses (McKone et al., 2007; Pitcher et al., 2007). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/13%3A_Seeing/13.01%3A_Introduction.txt |
It has been estimated that the human visual system can detect and discriminate among 7 million color variations (Geldard, 1972), but these variations are all created by the combinations of the three primary colors: red, green, and blue. The shade of a color, known as hue, is conveyed by the wavelength of the light that enters the eye (we see shorter wavelengths as more blue and longer wavelengths as more red), and we detect brightness from the intensity or height of the wave (bigger or more intense waves are perceived as brighter) (Figure \(1\)).
In his important research on color vision, Hermann von Helmholtz (1821–1894) theorized that color is perceived because the cones in the retina come in three types. One type of cone reacts primarily to blue light (short wavelengths), another reacts primarily to green light (medium wavelengths), and a third reacts primarily to red light (long wavelengths). The visual cortex then detects and compares the strength of the signals from each of the three types of cones, creating the experience of color. According to this Young-Helmholtz trichromatic color theory, what color we see depends on the mix of the signals from the three types of cones. If the brain is receiving primarily red and blue signals, for instance, it will perceive purple; if it is receiving primarily red and green signals it will perceive yellow; and if it is receiving messages from all three types of cones it will perceive white.
The different functions of the three types of cones are apparent in people who experience color blindness—the inability to detect green and/or red colors. About 1 in 50 people, mostly men, lack functioning in the red- or green-sensitive cones, leaving them only able to experience either one or two colors (Figure \(2\)).
The trichromatic color theory cannot explain all of human vision, however. For one, although the color purple does appear to us as a mixing of red and blue, yellow does not appear to be a mix of red and green. And people with color blindness, who cannot see either green or red, nevertheless can still see yellow. An alternative approach to the Young-Helmholtz theory, known as the opponent-process color theory, proposes that we analyze sensory information not in terms of three colors but rather in three sets of “opponent colors”: red-green, yellow-blue, and white-black. Evidence for the opponent-process theory comes from the fact that some neurons in the retina and in the visual cortex are excited by one color (e.g., red) but inhibited by another color (e.g., green).
One example of opponent processing occurs in the experience of an afterimage. If you stare at the flag on the left side of Figure \(3\) for about 30 seconds (the longer you look, the better the effect), and then move your eyes to the blank area to the right of it, you will see the afterimage. When we stare at the green stripes, our green receptors habituate and begin to process less strongly, whereas the red receptors remain at full strength. When we switch our gaze, we see primarily the red part of the opponent process. Similar processes create blue after yellow and white after black.
The tricolor and the opponent-process mechanisms work together to produce color vision. When light rays enter the eye, the red, blue, and green cones on the retina respond in different degrees and send different strength signals of red, blue, and green through the optic nerve. The color signals are then processed both by the ganglion cells and by the neurons in the visual cortex (Gegenfurtner & Kiper, 2003).
13.04: Perceiving Form
One of the important processes required in vision is the perception of form. German psychologists in the 1930s and 1940s, including Max Wertheimer (1880–1943), Kurt Koffka (1886–1941), and Wolfgang Kohler (1887–1967), argued that we create forms out of their component sensations based on the idea of the gestalt, a meaningfully organized whole. The idea of the gestalt is that the “whole is more than the sum of its parts.” Some examples of how gestalt principles lead us to see more than what is actually there are summarized in Figure \(1\). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/13%3A_Seeing/13.03%3A_Perceiving_Color.txt |
Depth perception is the ability to perceive three-dimensional space and to accurately judge distance. Without depth perception, we would be unable to drive a car, thread a needle, or simply navigate our way around the supermarket (Howard & Rogers, 2001). Research has found that depth perception is in part based on innate capacities and in part learned through experience (Witherington, 2005).
Psychologists Eleanor Gibson and Richard Walk (1960) tested the ability to perceive depth in 6- to 14-month-old infants by placing them on a visual cliff, a mechanism that gives the perception of a dangerous drop-off, in which infants can be safely tested for their perception of depth (Figure \(1\)). The infants were placed on one side of the “cliff,” while their mothers called to them from the other side. Gibson and Walk found that most infants either crawled away from the cliff or remained on the board and cried because they wanted to go to their mothers, but the infants perceived a chasm that they instinctively could not cross. Further research has found that even very young children who cannot yet crawl are fearful of heights (Campos et al., 1970). On the other hand, studies have also found that infants improve their hand-eye coordination as they learn to better grasp objects and as they gain more experience in crawling, indicating that depth perception is also learned (Adolph, 2000).
Depth perception is the result of our use of depth cues, messages from our bodies and the external environment that supply us with information about space and distance. Binocular depth cues are depth cues that are created by retinal image disparity—that is, the space between our eyes—and thus require the coordination of both eyes. One outcome of retinal disparity is that the images projected on each eye are slightly different from each other. The visual cortex automatically merges the two images into one, enabling us to perceive depth. Three-dimensional movies make use of retinal disparity by using 3-D glasses that the viewer wears to create a different image on each eye. The perceptual system quickly, easily, and unconsciously turns the disparity into 3-D.
An important binocular depth cue is convergence, the inward turning of our eyes that is required to focus on objects that are less than about 50 feet away from us. The visual cortex uses the size of the convergence angle between the eyes to judge the object’s distance. You will be able to feel your eyes converging if you slowly bring a finger closer to your nose while continuing to focus on it. When you close one eye, you no longer feel the tension—convergence is a binocular depth cue that requires both eyes to work.
The visual system also uses accommodation to help deter- mine depth. As the lens changes its curvature to focus on distant or close objects, information relayed from the muscles attached to the lens helps us determine an object’s distance. Accommodation is only effective at short viewing distances, however, so while it comes in handy when threading a needle or tying shoe- laces, it is far less effective when driving or playing sports.
Although the best cues to depth occur when both eyes work together, we are able to see depth even with one eye closed. Monocular depth cues are depth cues that help us perceive depth using only one eye (Sekuler & Blake, 2006). Some of the most important are summarized in Table \(1\).
Name
Description
Example
Image
Position
We tend to see objects higher up in our field of vision as farther away.
In the image, the fence posts at left appear farther away not only because they become smaller but also because they appear higher up in the picture.
Relative size
Assuming that the objects in a scene are the same size, smaller objects are perceived as farther away.
At right, the cars in the distance appear smaller than those nearer to us.
Linear perspective
Parallel lines appear to converge at a distance
We know that the tracks at right are parallel. When they appear closer together, we determine they are farther away.
Light and shadow
The eye receives more reflected light from objects that are closer to us. Normally, light comes from above, so darker images are in shadow.
We see the ovals at right as extended and indented according to their shadowing. If we invert the picture, the images will reverse.
Interposition
When one object overlaps another object, we view it as closer.
At right, because the orange star covers the blue bar, it is seen as closer than the yellow moon.
Aerial perspective
Objects that appear hazy, or that are covered with smog or dust, appear farther away.
The artist who painted the picture on the right used aerial perspective to make the distant hills more hazy and thus appear farther away. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/13%3A_Seeing/13.05%3A_Perceiving_Depth.txt |
Many animals, including human beings, have very sophisticated perceptual skills that allow them to coordinate their own motion with the motion of moving objects in order to create a collision with that object. Bats and birds use this mechanism to catch up with prey, dogs use it to catch a Frisbee, and humans use it to catch a moving football. The brain detects motion partly from the changing size of an image on the retina (objects that look bigger are usually closer to us) and in part from the relative brightness of objects.
We also experience motion when objects near each other change their appearance. The beta effect refers to the perception of motion that occurs when different images are presented next to each other in succession (see the Beta Effect and Phi Phenomenon note). The visual cortex fills in the missing part of the motion and we see the object moving. The beta effect is used in movies to create the experience of motion. A related effect is the phi phenomenon, in which we perceive a sensation of motion caused by the appearance and disappearance of objects that are near each other. The phi phenomenon looks like a moving zone or cloud of background color surrounding the flashing objects. The beta effect and the phi phenomenon are other examples of the importance of the gestalt—our tendency to “see more than the sum of the parts.”
BETA EFFECT AND PHI PHENOMENON
In the beta effect, our eyes detect motion from a series of still images, each with the object in a different place. This is the fundamental mechanism of motion pictures (movies).
In the phi phenomenon, the perception of motion is based on the momentary hiding of an image.
KEY TAKEAWAYS
• Vision is the process of detecting the electromagnetic energy that surrounds us. Only a small fraction of the electromagnetic spec- trum is visible to humans.
• The visual receptor cells on the retina detect shape, color, motion, and depth.
• Light enters the eye through the transparent cornea and passes through the pupil at the center of the iris. The lens adjusts to focus the light on the retina, where it appears upside down and backward. Receptor cells on the retina are excited or inhibited by the light and send information to the visual cortex through the optic nerve.
• • The retina has two types of photoreceptor cells: rods, which detect brightness and respond to black and white, and cones, which respond to red, green, and blue. Color blindness occurs when people lack function in the red- or green-sensitive cones.
• • Feature detector neurons in the visual cortex help us recognize objects, and some neurons respond selectively to faces and other body parts.
• The Young-Helmholtz trichromatic color theory proposes that color perception is the result of the signals sent by the three types of cones, whereas the opponent-process color theory proposes that we perceive color as three sets of opponent colors: red-green, yellow-blue, and white-black.
• The ability to perceive depth occurs through the result of binocular and monocular depth cues.
• Motion is perceived as a function of the size and brightness of objects. The beta effect and the phi phenomenon are examples of perceived motion.
EXERCISES AND CRITICAL THINKING
1. Consider some ways that the processes of visual perception help you engage in an everyday activity, such as driving a car or riding a bicycle.
2. Imagine for a moment what your life would be like if you couldn’t see. Do you think you would be able to compensate for your loss of sight by using other senses?
REFERENCES
Adolph, K. E. (2000). Specificity of learning: Why infants fall over a veritable cliff. Psychological Science, 11(4), 290–295. doi. org/10.1111/1467-9280.00258
Campos, J. J., Langer, A., & Krowitz, A. (1970). Cardiac responses on the visual cliff in prelocomotor human infants. Science, 170(3954), 196–197. https://doi.org/10.1126/science.170.3954.196
Downing, P. E., Jiang, Y., Shuman, M., & Kanwisher, N. (2001). A cortical area selective for visual processing of the human body. Science, 293(5539), 2470–2473. https://doi.org/10.1126/science.1063414
Gegenfurtner, K. R., & Kiper, D. C. (2003). Color vision. Annual Review of Neuroscience, 26, 181–206. doi.org/10.1146/annurev. neuro.26.041002.131116
Geldard, F. A. (1972). The human senses (2nd ed.). John Wiley & Sons. Gibson, E. J., & Walk, R. D. (1960). The “visual cliff.” Scientific American,
202(4), 64–71. doi.org/10.1038/scientificamerican0460-64 Haxby, J. V., Gobbini, M. I., Furey, M. L., Ishai, A., Schouten, J. L., &
Pietrini, P. (2001). Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science, 293(5539), 2425–2430. https://doi.org/10.1126/science.1063736
Howard, I. P., & Rogers, B. J. (2001). Seeing in depth: Basic mechanisms (Vol. 1). Porteous.
Kelsey, C.A. (1997). Detection of visual information. In W. R. Hendee & P. N. T. Wells (Eds.), The perception of visual information (2nd ed.). Springer Verlag.
Livingstone M. S. (2000). Is it warm? Is it real? Or just low spatial frequency? Science, 290, 1299. doi.org/10.1126/ science.290.5495.1299b
Livingstone, M., & Hubel, D. (1998). Segregation of form, color, movement, and depth: Anatomy, physiology, and perception. Science, 240, 740–749. https://doi.org/10.1126/science.3283936
McKone, E., Kanwisher, N., & Duchaine, B. C. (2007). Can generic expertise explain special processing for faces? Trends in Cognitive Sciences, 11, 8–15. https://doi.org/10.1016/j.tics.2006.11.002
Pitcher, D., Walsh, V., Yovel, G., & Duchaine, B. (2007). TMS evidence for the involvement of the right occipital face area in early face processing. Current Biology, 17, 1568–1573. doi.org/10.1016/j. cub.2007.07.063
Rodriguez, E., George, N., Lachaux, J.-P., Martinerie, J., Renault, B., & Varela, F. J. (1999). Perception’s shadow: Long-distance synchronization of human brain activity. Nature, 397(6718), 430–433. https://doi.org/10.1038/17120
Sekuler, R., & Blake, R. (2006). Perception (5th ed.). McGraw-Hill. Witherington, D. C. (2005). The development of prospective grasping
control between 5 and 7 months: A longitudinal study. Infancy, 7(2), 143–161. doi.org/10.1207/s15327078in0702_2 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/13%3A_Seeing/13.06%3A_Perceiving_Motion.txt |
The eyes, ears, nose, tongue, and skin sense the world around us and, in some cases, perform preliminary infor- mation processing on the incoming data. But by and large, we do not experience sensation—we experience the out- come of perception—the total package that the brain puts together from the pieces it receives through our senses and that the brain creates for us to experience. When we look out the window at a view of the countryside, or when we look at the face of a good friend, we don’t just see a jumble of colors and shapes—we see, instead, an image of a countryside or an image of a friend (Goodale & Milner, 2006).
14.02: How The Perceptual System Interprets The Environment
This meaning-making involves the automatic operation of a variety of essential perceptual processes. One of these is sensory interactionthe working together of different senses to create experience. Sensory interaction is involved when taste, smell, and texture combine to create the flavor we experience in food. It is also involved when we enjoy a movie because of the way the images and the music work together.
Although you might think that we understand speech only through our sense of hearing, it turns out that the visual aspect of speech is also important. One example of sensory interaction is shown in the McGurk effect—an error in perception that occurs when we misperceive sounds because the audio and visual parts of the speech are mismatched. You can witness the effect yourself by viewing the following video clip.
THE MCGURK EFFECT
Watch this brief video about an auditory illusion known as the McGurk effect.
Other examples of sensory interaction include the experience of nausea that can occur when the sensory information being received from the eyes and the body does not match information from the vestibular system (Flanagan et al., 2004) and synesthesia—an experience in which one sensation (e.g., hearing a sound) creates experiences in another (e.g., vision). Most people do not experience synesthesia, but those who do link their perceptions in unusual ways, for instance, by experiencing color when they taste a particular food or by hearing sounds when they see certain objects (Ramachandran et al., 2005).
Another important perceptual process is selective attention—the ability to focus on some sensory inputs while tuning out others. View the following video clip and count the number of times the people playing with the ball pass it to each other. You may find that, like many other people who view it for the first time, you miss something important because you selectively attend to only one aspect of the video (Simons & Chabris, 1999).
SELECTIVE ATTENTION
Watch this video and carefully count how many times the people in white pass the ball to each other.
Selective attention also allows us to focus on a single talker at a party while ignoring other conversations that are occurring around us (Broadbent, 1958; Cherry, 1953). With- out this automatic selective attention, we’d be unable to focus on the single conversation we want to hear. But selective attention is not complete; we also, at the same time, monitor what’s happening in the channels we are not focusing on. Perhaps you have had the experience of being at a party and talking to someone in one part of the room, when suddenly you hear your name being mentioned by someone in another part of the room. This cocktail party phenomenon shows us that although selective attention is limiting what we process, we are nevertheless, at the same time, doing a lot of unconscious monitoring of the world around us—you didn’t know you were attending to the background sounds of the party, but evidently you were.
A second fundamental process of perception is sensory adaptation—a decreased sensitivity to a stimulus after pro- longed and constant exposure. When you step into a swimming pool, the water initially feels cold, but after a while you stop noticing it. After prolonged exposure to the same stimulus, our sensitivity toward it diminishes and we no longer perceive it. The ability to adapt to the things that don’t change around us is essential to our survival, as it leaves our sensory receptors free to detect the important and informative changes in our environment and to respond accordingly. We ignore the sounds that our car makes every day, which leaves us free to pay attention to the sounds that are different from normal, and thus likely to need our attention. Our sensory receptors are alert to novelty and are fatigued after constant exposure to the same stimulus.
If sensory adaptation occurs with all senses, why doesn’t an image fade away after we stare at it for a period of time? The answer is that, although we are not aware of it, our eyes are constantly flitting from one angle to the next, making thou- sands of tiny movements (called saccades) every minute. This constant eye movement guarantees that the image we are viewing always falls on fresh receptor cells. What would happen if we could stop the movement of our eyes? Psychologists have devised a way of testing the sensory adaptation of the eye by attaching an instrument that ensures a constant image is maintained on the eye’s inner surface. Participants are fitted with a contact lens that has a miniature slide projector attached to it. Because the projector follows the exact movements of the eye, the same image is always projected, stimulating the same spot, on the retina. Within a few seconds, interesting things begin to happen. The image will begin to vanish, then reappear, only to disappear again, either in pieces or as a whole. Even the eye experiences sensory adaptation (Yarbus, 1967).
One of the major problems in perception is to ensure that we always perceive the same object in the same way, despite the fact that the sensations that it creates on our receptors changes dramatically. The ability to perceive a stimulus as constant despite changes in sensation is known as perceptual constancy. Consider an image of a door as it swings. When it is closed, we see it as rectangular, but when it is open, we see only its edge and it appears as a line. But we never perceive the door as changing shape as it swings—perceptual mechanisms take care of the problem for us by allowing us to see a constant shape.
The visual system also corrects for color constancy. Imagine that you are wearing blue jeans and a bright white t-shirt. When you are outdoors, both colors will be at their brightest, but you will still perceive the white t-shirt as bright and the blue jeans as darker. When you go indoors, the light shining on the clothes will be significantly dimmer, but you will still perceive the t-shirt as bright. This is because we put colors in context and see that, compared to its surroundings, the white t-shirt reflects the most light (McCann, 1992). In the same way, a green leaf on a cloudy day may reflect the same wavelength of light as a brown tree branch does on a sunny day. Nevertheless, we still perceive the leaf as green and the branch as brown.
14.03: Illusions
Although our perception is very accurate, it is not perfect. Illusions occur when the perceptual processes that normally help us correctly perceive the world around us are fooled by a particular situation so that we see something that does not exist or that is incorrect. Figure \(1\) presents two situations in which our normally accurate perceptions of visual constancy have been fooled.
Another well-known illusion is the Mueller-Lyer illusion (see Figure \(2\)). The line segment in the bottom arrow looks longer to us than the one on the top, even though they are both actually the same length. It is likely that the illusion is, in part, the result of the failure of monocular depth cues—the bottom line looks like an edge that is normally farther away from us, whereas the top one looks like an edge that is normally closer.
The moon illusion refers to the fact that the moon is perceived to be about 50% larger when it is near the horizon than when it is seen overhead, despite the fact that both moons are the same size and cast the same size retinal image. The monocular depth cues of position and aerial perspective create the illusion that things that are lower and more hazy are farther away. The skyline of the horizon (trees, clouds, outlines of buildings) also gives a cue that the moon is far away, com- pared to a moon at its zenith. If we look at a horizon moon through a tube of rolled up paper, taking away the surrounding horizon cues, the moon will immediately appear smaller.
The Ponzo illusion operates on the same principle. As you can see in Figure \(3\), the top yellow bar seems longer than the bottom one, but if you measure them you’ll see that they are exactly the same length. The monocular depth cue of linear perspective leads us to believe that, given two similar objects, the distant one can only cast the same size retinal image as the closer object if it is larger. The topmost bar there- fore appears longer.
Illusions demonstrate that our perception of the world around us may be influenced by our prior knowledge. But the fact that some illusions exist in some cases does not mean that the perceptual system is generally inaccurate—in fact, humans normally become so closely in touch with their environment that the physical body and the particular environment we sense and perceive becomes embodied—that is, built into and linked with—our cognition, such that the worlds around us become part of our brain (Calvo & Gomila, 2008). The close relationship between people and their environments means that, although illusions can be created in the lab and under some unique situations, they may be less common with active observers in the real world (Runeson, 1988). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/14%3A_Accuracy_and_Inaccuracy_in_Perception/14.01%3A_Introduction.txt |
Our emotions, mindset, expectations, and the contexts in which our sensations occur all have a profound influence on perception. People who are warned that they are about to taste something bad rate what they do taste more negatively than people who are told that the taste won’t be so bad (Nitschke et al., 2006), and people perceive a child and adult pair as looking more alike when they are told that they are parent and child (Bressan & Dal Martello, 2002). Similarly, participants who see images of the same baby rate it as stronger and bigger when they are told it is a boy as opposed to when they are told it is a girl (Stern & Karraker, 1989), and research participants who learn that a child is from a lower-class background perceive the child’s scores on an intelligence test as lower than people who see the same test taken by a child they are told is from an upper-class background (Darley & Gross, 1983). Plassmann et al. (2008) found that wines were rated more positively and caused greater brain activity in brain areas associated with pleasure when they were said to cost more than when they were said to cost less. And even experts can be fooled: Professional referees tended to assign more penalty cards to soccer teams for videotaped fouls when they were told that the team had a history of aggressive behavior than when they had no such expectation (Jones et al., 2002).
Psychology In Everyday Life
How Understanding Sensation and Perception Can Save Lives
Human factors is the field of psychology that uses psycho- logical knowledge, including the principles of sensation and perception, to improve the development of technology. Human factors has worked on a variety of projects, ranging from nuclear reactor control centers and airplane cockpits to cell phones and websites (Proctor & van Zandt, 2008). For instance, modern televisions and computer monitors were developed on the basis of the trichromatic color theory, using three color elements placed close enough together so that the colors are blended by the eye. knowledge of the visual system also helped engineers create new kinds of displays, such as those used on notebook computers and music players, and better understand how using cell phones while driving may contribute to automobile accidents (lee & Strayer, 2004).
Human factors also has made substantial contributions to airline safety. About two thirds of accidents on commercial airplane flights are caused by human error (Nickerson, 1998). During takeoff, travel, and landing, the pilot simultaneously communicates with ground control, maneuvers the plane, scans the horizon for other aircraft, and operates controls. The need for a usable interface that works easily and naturally with the pilot’s visual perception is essential.
Psychologist Conrad kraft (1978) hypothesized that as planes land, with no other distance cues visible, pilots may be subjected to a type of moon illusion, in which the city lights beyond the runway appear much larger on the retina than they really are, deceiving the pilot into landing too early. kraft’s findings caused airlines to institute new flight safety measures, where copilots must call out the altitude progressively during the descent, which has probably decreased the number of landing accidents.
Figure \(1\) presents the design of an airplane instrument panel before and after it was redesigned by human factors psychologists. On the left is the initial design in which the controls were crowded and cluttered, in no logical sequence, each control performing one task. The controls were more or less the same in color, and the gauges were not easy to read. The redesigned digital cockpit shows a marked improvement in usability. More of the controls are color-coded and multifunctional so that there is less clutter on the dashboard. Screens make use of lCD and 3-D graphics. Text sizes are changeable—increasing readability—and many of the functions have become automated, freeing up the pilot’s concentration for more important activities.
One important aspect of the redesign was based on the principles of sensory adaptation. Displays that are easy to see in darker conditions quickly become unreadable when the sun shines directly on them. It takes the pilot a relatively long time to adapt to the suddenly much brighter display. Furthermore, perceptual contrast is important. The display cannot be so bright at night that the pilot is unable to see targets in the sky or on the land. Human factors psychologists used these principles to determine the appropriate stimulus intensity needed on these displays so that pilots would be able to read them accurately and quickly under a wide range of conditions. The psychologists accomplished this by developing an automatic control mechanism that senses the ambient light visible through the front cockpit windows and that detects the light falling on the display surface, and then automatically adjusts the intensity of the display for the pilot (Silverstein & Merrifield, 1985; Silverstein et al., 1990). ■
Our perceptions are also influenced by our desires and motivations. When we are hungry, food-related words tend to grab our attention more than non-food-related words (Mogg et al., 1998); we perceive objects we can reach as bigger than those we cannot reach (Witt & Proffitt, 2005); and people who favor a political candidate’s policies view the can- didate’s skin color more positively than do those who oppose the candidate’s policies (Caruso et al., 2009). Even our culture influences perception. Chua et al. (2005) showed American and Asian graduate students different images, such as an air- plane, an animal, or a train, against complex backgrounds. They found that (consistent with their overall individualistic orientation) the American students tended to focus more on the foreground image, while Asian students (consistent with their interdependent orientation) paid more attention to the image’s context. Furthermore, Asian American students focused more or less on the context depending on whether their Asian or their American identity had been activated.
KEY TAKEAWAYS
• Sensory interaction occurs when different senses work together, for instance, when taste, smell, and touch together produce the flavor of food.
• Selective attention allows us to focus on some sensory experiences while tuning out others.
• Sensory adaptation occurs when we become less sensitive to some aspects of our environment, freeing us to focus on more important changes.
• Perceptual constancy allows us to perceive an object as the same, despite changes in sensation.
• Cognitive illusions are examples of how our expectations can influence our perceptions.
• Our emotions, motivations, desires, and even our culture can influence our perceptions.
EXERCISE
1. Consider some cases where your expectations about what you think you might be going to experience have influenced your perceptions of what you actually experienced.
REFERENCES
Bressan, P., & Dal Martello, M. F. (2002). Talis pater, talis filius: Perceived resemblance and the belief in genetic relatedness. Psychological Science, 13, 213–218. https://doi.org/10.1111/1467-9280.00440
Broadbent, D. E. (1958). Perception and communication. Pergamon. Calvo, P., & Gomila, T. (Eds.). (2008). Handbook of cognitive science: An
embodied approach. Elsevier.
Caruso, E. M., Mead, N. L., & Balcetis, E. (2009). Political partisanship
influences perception of biracial candidates’ skin tone. Proceedings of the National Academy of Sciences, 106(48), 20168–20173. https:// doi.org/10.1073/pnas.0905362106
Cherry, E. C. (1953). Some experiments on the recognition of speech, with one and with two ears. Journal of the Acoustical Society of America, 25, 975–979. https://doi.org/10.1121/1.1907229
Chua, H. F., Boland, J. E., & Nisbett, R. E. (2005). Cultural variation in eye movements during scene perception. Proceedings of the National Academy of Sciences, 102, 12629–12633. doi.org/10.1073/ pnas.0506162102
Darley, J. M., & Gross, P. H. (1983). A hypothesis-confirming bias in labeling effects. Journal of Personality and Social Psychology, 44, 20–33. https://doi.org/10.1037/0022-3514.44.1.20
Flanagan, M. B., May, J. G., & Dobie, T. G. (2004). The role of vection, eye movements and postural instability in the etiology of motion sickness. Journal of Vestibular Research: Equilibrium and Orientation, 14(4), 335–346.
Goodale, M., & Milner, D. (2006). One brain—Two visual systems. Psychologist, 19(11), 660–663. https://thepsychologist.bps.org.uk/ volume-19/edition-11/one-brain-two-visual-systems
Jones, M. V., Paull, G. C., & Erskine, J. (2002). The impact of a team’s aggressive reputation on the decisions of association football referees. Journal of Sports Sciences, 20, 991–1000. https://doi.org/ 10.1080/026404102321011751
Kraft, C. (1978). A psychophysical approach to air safety: Simulator studies of visual illusions in night approaches. In H. L. Pick, H. W. Leibowitz, J. E. Singer, A. Steinschneider, & H. W. Steenson (Eds.), Psychology: From research to practice. Plenum Press.
Lee, J., & Strayer, D. (2004). Preface to the special section on driver distraction. Human Factors, 46(4), 583. doi.org/10.1518/ hfes.46.4.583.56811
McCann, J. J. (1992). Rules for colour constancy. Ophthalmic and Physiologic Optics, 12(2), 175–177. doi.org/10.1111/ j.1475-1313.1992.tb00285.x
Mogg, K., Bradley, B. P., Hyare, H., & Lee, S. (1998). Selective attention to food-related stimuli in hunger: Are attentional biases specific to emotional and psychopathological states, or are they also found in normal drive states? Behaviour Research and Therapy, 36(2), 227–237. https://doi.org/10.1016/S0005-7967(97)00062-4
Nickerson, R. S. (1998). Applied experimental psychology. Applied Psychology, 47, 155–173. doi.org/10.1111/j.1464-0597.1998. tb00019.x
Nitschke, J. B., Dixon, G. E., Sarinopoulos, I., Short, S. J., Cohen, J. D., Smith, E. E., & Davidson, R. J. (2006). Altering expectancy dampens neural response to aversive taste in primary taste cortex. Nature Neuroscience, 9, 435–442. https://doi.org/10.1038/nn1645
Plassmann, H., O’Doherty, J., Shiv, B., & Rangel, A. (2008). Marketing actions can modulate neural representations of experienced pleasantness. Proceedings of the National Academy of Sciences, 105(3), 1050–1054. https://doi.org/10.1073/pnas.0706929105
Proctor, R. W., & Van Zandt, T. (2008). Human factors in simple and
complex systems (2nd ed.). CRC Press.
Ramachandran, V. S., Hubbard, E. M., Robertson, L. C., & Sagiv, N. (2005).
The emergence of the human mind: Some clues from synesthesia. In Synesthesia: Perspectives from cognitive neuroscience (pp. 147–190). Oxford University Press.
Runeson, S. (1988). The distorted room illusion, equivalent configurations, and the specificity of static optic arrays. Journal of Experimental Psychology: Human Perception and Performance, 14(2), 295–304. https://doi.org/10.1037/0096-1523.14.2.295
Silverstein, L. D., Krantz, J. H., Gomer, F. E., Yeh, Y., & Monty, R. W. (1990). The effects of spatial sampling and luminance quantization on the image quality of color matrix displays. Journal of the Optical Society of America A, 7, 1955–1968. https://doi.org/10.1364/JOSAA.7.001955
Silverstein, L. D., & Merrifield, R. M. (1985). The development and evaluation of color systems for airborne applications: Phase I: Fundamental visual, perceptual, and display systems considerations (Tech. Report DOT/ FAA/PM085019). Federal Aviation Administration. https://apps.dtic.mil/dtic/tr/fulltext/u2/a159208.pdf
Simons, D. J., & Chabris, C. F. (1999). Gorillas in our midst: Sustained inattentional blindness for dynamic events. Perception, 28(9), 1059–1074. https://doi.org/10.1068/p281059
Stern, M., & Karraker, K. H. (1989). Sex stereotyping of infants: A review of gender labeling studies. Sex Roles, 20(9–10), 501–522. doi. org/10.1007/BF00288198
Witt, J. K., & Proffitt, D. R. (2005). See the ball, hit the ball: Apparent ball size is correlated with batting average. Psychological Science, 16(12), 937–938. https://doi.org/10.1111/j.1467-9280.2005.01640.x
Yarbus, A. L. (1967). Eye movements and vision. Plenum Press. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/14%3A_Accuracy_and_Inaccuracy_in_Perception/14.04%3A_The_Important_Role_Of_Expectations_In_Perception.txt |
Our emotions, thoughts, and behaviors play an import- ant role in our health. Not only do they influence our day-to-day health practices, but they can also influence how our body functions. This module provides an overview of health psychology, which is a field devoted to understanding the connections between psychology and health. Discussed here are examples of topics a health psychologist might study, including stress, psychosocial factors related to health and disease, how to use psychology to improve health, and the role of psychology in medicine.
15.02: What Is Health Psychology
Today, we face more chronic disease than ever before because we are living longer lives while also frequently behaving in unhealthy ways. One example of a chronic disease is coronary heart disease (CHD): It is the number one cause of death worldwide (World Health Organization, 2017). CHD develops slowly over time and typically appears midlife, but related heart problems can persist for years after the original diagnosis or cardiovascular event. In managing illnesses that persist over time (other examples might include cancer, diabetes, and long-term disability), many psychological factors will determine the progression of the ailment. For example, do patients seek help when appropriate? Do they follow doctor recommendations? Do they develop negative psychological symptoms due to lasting illness (e.g., depression)? Also important is that psychological factors can play a significant role in who develops these diseases, the prognosis, and the nature of the symptoms related to the illness. Health psychology is a relatively new, interdisciplinary field of study that focuses on these very issues, or more specifically, the role of psychology in maintaining health, as well as preventing and treating illness.
Consideration of how psychological and social factors influence health is especially important today because many of the leading causes of illness in developed countries are often attributed to psychological and behavioral factors. In the case of CHD, discussed above, psychosocial factors, such as excessive stress, smoking, unhealthy eating habits, and some personality traits can also lead to increased risk of disease and worse health outcomes. That being said, many of these factors can be adjusted using psychological techniques. For example, clinical health psychologists can improve health practices like poor dietary choices and smoking, they can teach important stress reduction techniques, and they can help treat psycho- logical disorders tied to poor health. Health psychology considers how the choices we make, the behaviors we engage in, and even the emotions we feel can play an important role in our overall health (Cohen & Herbert, 1996; Taylor, 2012).
Health psychology relies on the Biopsychosocial Model of Health. This model posits that biology, psychology, and social factors are just as important in the development of disease as biological causes (e.g., germs, viruses), which is consistent with the World Health Organization (1946) definition of health. This model replaces the older Biomedical Model of Health, which primarily considers the physical, or pathogenic, factors contributing to illness. Thanks to advances in medical technology, there is a growing understanding of the physiology underlying the mind–body connection and, in particular, the role that different feelings can have on our body’s function. Health psychology researchers working in the fields of psychosomatic medicine and psychoneuroimmunology, for example, are interested in understanding how psychological factors can “get under the skin” and influence our physiology in order to better understand how factors like stress can make us sick. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/15%3A_The_Healthy_Life/15.01%3A_Introduction.txt |
You probably know exactly what it’s like to feel stress, but what you may not know is that it can objectively influence your health. Answers to questions like “How stressed do you feel?” or “How overwhelmed do you feel?” can predict your likelihood of developing both minor illnesses and serious problems like future heart attack (Cohen et al., 2007). To understand how health psychologists study these types of associations, we will describe one famous example of a stress and health study. Imagine that you are a research subject for a moment. After you check into a hotel room as part of the study, the researchers ask you to report your general levels of stress. Not too surprising; however, what happens next is that you receive droplets of cold virus into your nose! The researchers intentionally try to make you sick by exposing you to an infectious illness. After they expose you to the virus, the researchers will then evaluate you for several days by asking you questions about your symptoms, monitoring how much mucus you are producing by weighing your used tissues, and taking body fluid samples—all to see if you are objectively ill with a cold. Now, the interesting thing is that not everyone who has drops of cold virus put in their nose develops the illness. Studies like this one find that people who are less stressed and those who are more positive at the beginning of the study are at a decreased risk of developing a cold (Cohen et al., 1991, 2006) (see Figure \(1\) for an example).
Importantly, it is not just major life stressors (e.g., a family death, a natural disaster) that increase the likelihood of get- ting sick. Even small daily hassles like getting stuck in traffic or fighting with your girlfriend can raise your blood pressure, alter your stress hormones, and even suppress your immune system function (DeLongis et al., 1988; Twisk et al., 1999).
It is clear that stress plays a major role in our mental and physical health, but what exactly is it? The term stress was originally derived from the field of mechanics where it is used to describe materials under pressure. The word was first used in a psychological manner by researcher Hans Selye. He was examining the effect of an ovarian hormone that he thought caused sickness in a sample of rats. Surprisingly, he noticed that almost any injected hormone produced this same sick- ness. He smartly realized that it was not the hormone under investigation that was causing these problems, but instead, the aversive experience of being handled and injected by researchers that led to high physiological arousal and, eventually, to health problems like ulcers. Selye (1946) coined the term stressor to label a stimulus that had this effect on the body and developed a model of the stress response called the general adaptation syndrome. Since then, psychologists have studied stress in a myriad of ways, including stress as negative events (e.g., natural disasters or major life changes like dropping out of school), as chronically difficult situations (e.g., taking care of a loved one with Alzheimer’s), as short- term hassles, as a biological fight-or-flight response, and even as clinical illness like post-traumatic stress disorder (PTSD). It continues to be one of the most important and well-studied psychological correlates of illness, because excessive stress causes potentially damaging wear and tear on the body and can influence almost any imaginable disease process. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/15%3A_The_Healthy_Life/15.03%3A_Stress_And_Health.txt |
An important question that health psychologists ask is, What keeps us protected from disease and alive longer? When considering this issue of resilience (Rutter, 1985), five factors are often studied in terms of their ability to protect (or sometimes harm) health. They are:
1. Coping
2. Control and Self-Efficacy
3. Social Relationships
4. Dispositions and Emotions
5. Stress Management
Coping Strategies
How individuals cope with the stressors they face can have a significant impact on health. Coping is often classified into two categories: problem-focused coping or emotion-focused coping (Carver et al., 1989). Problem-focused coping is thought of as actively addressing the event that is causing stress in an effort to solve the issue at hand. For example, say you have an important exam coming up next week. A problem-focused strategy might be to spend additional time over the weekend studying to make sure you understand all of the material. Emotion-focused coping, on the other hand, regulates the emotions that come with stress. In the above examination example, this might mean watching a funny movie to take your mind off the anxiety you are feeling. In the short term, emotion-focused coping might reduce feelings of stress, but problem-focused coping seems to have the greatest impact on mental wellness (Billings & Moos, 1981; Herman-Stabl et al., 1995). That being said, when events are uncontrollable (e.g., the death of a loved one), emotion-focused coping directed at managing your feelings, at first, might be the better strategy. Therefore, it is always important to consider the match of the stressor to the coping strategy when evaluating its plausible benefits.
Control and Self-Efficacy
Another factor tied to better health outcomes and an improved ability to cope with stress is having the belief that you have control over a situation. For example, in one study where participants were forced to listen to unpleasant (stressful) noise, those who were led to believe that they had control over the noise performed much better on proofreading tasks afterwards (Glass & Singer, 1972). In other words, even though participants did not have actual control over the noise, the control belief aided them in completing the task. In similar studies, perceived control benefited immune system functioning (Sieber et al., 1992). Outside of the laboratory, studies have shown that older residents in assisted living facilities, which are notorious for low control, lived longer and showed better health outcomes when given control over something as simple as watering a plant or choosing when student volunteers came to visit (Rodin & Langer, 1977; Schulz & Hanusa, 1978). In addition, feeling in control of a threatening situation can actually change stress hormone levels (Dicker- son & Kemeny, 2004). Believing that you have control over your own behaviors can also have a positive influence on important outcomes like smoking cessation, contraception use, and weight management (Wallston & Wallston, 1978). When individuals do not believe they have control, they do not try to change.
Self-efficacy is closely related to control, in that people with high levels of this trait believe they can complete tasks and reach their goals. Just as feeling in control can reduce stress and improve health, higher self-efficacy can reduce stress and negative health behaviors and is associated with better health (O’Leary, 1985).
Social Relationships
Research has shown that the impact of social isolation on our risk for disease and death is similar in magnitude to the risk associated with smoking regularly (Holt-Lunstad et al., 2010; House et al., 1988). In fact, the importance of social relationships for our health is so significant that some scientists believe our body has developed a physiological system that encourages us to seek out our relationships, especially in times of stress (Taylor et al., 2000). Social integration is the concept used to describe the number of social roles that you have (Cohen & Wills, 1985), as well as the lack of isolation. For example, you might be a daughter, a basketball team member, a Humane Society volunteer, a coworker, and a student. Maintaining these different roles can improve your health via encouragement from those around you to maintain a healthy lifestyle. Those in your social network might also provide you with social support (e.g., when you are under stress). This support might include emotional help (e.g., a hug when you need it), tangible help (e.g., lending you money), or advice. By helping to improve health behaviors and reduce stress, social relationships can have a powerful, protective impact on health and, in some cases, might even help people with serious illnesses stay alive longer (Spiegel et al., 1989).
Dispositions and Emotions: What’s Risky and What’s Protective?
Negative dispositions and personality traits have been strongly tied to an array of health risks. One of the earliest negative trait-to-health connections was discovered in the 1950s by two cardiologists. They made the interesting discovery that there were common behavioral and psychological patterns among their heart patients that were not present in other patient samples. This pattern included being competitive, impatient, hostile, and time urgent. They labeled it Type A behavior. Importantly, it was found to be associated with double the risk of heart disease as compared with Type B behavior (Friedman & Rosenman, 1959). Since the 1950s, researchers have discovered that it is the hostility and competitiveness components of Type A that are especially harmful to heart health (Iribarren et al., 2000; Matthews et al., 1977; Miller et al., 1996). Hostile individuals are quick to get upset, and this angry arousal can damage the arteries of the heart. In addition, given their negative personality style, hostile people often lack a heath-protective supportive social network.
Positive traits and states, on the other hand, are often health protective. For example, characteristics like positive emotions (e.g., feeling happy or excited) have been tied to a wide range of benefits such as increased longevity, a reduced likelihood of developing some illnesses, and better outcomes once you are diagnosed with certain diseases (e.g., heart disease, HIV) (Pressman & Cohen, 2005). Across the world, even in the most poor and underdeveloped nations, positive emotions are consistently tied to better health (Pressman et al., 2013). Positive emotions can also serve as the “antidote” to stress, protecting us against some of its damaging effects (Fredrickson, 2001; Pressman & Cohen, 2005) (see FiGURE 15.2). Similarly, looking on the bright side can also improve health. Optimism has been shown to improve coping, reduce stress, and predict better disease outcomes like recovering from a heart attack more rapidly (Kubzansky et al., 2001; Nes & Segerstrom, 2006; Scheier & Carver, 1985; Segerstrom et al., 1998).
Stress Management
About 20% of Americans report having stress, with 18- to 33-year-olds reporting the highest levels (American Psychological Association, 2012). Given that the sources of our stress are often difficult to change (e.g., personal finances, current job), a number of interventions have been designed to help reduce the aversive responses to duress. For example, relaxation activities and forms of meditation are techniques that allow individuals to reduce their stress via breathing exercises, muscle relaxation, and mental imagery. Physiological arousal from stress can also be reduced via biofeedback, a technique in which the individual is shown bodily information that is not normally available to them (e.g., heart rate), and then taught strategies to alter this signal. This type of intervention has even shown promise in reducing heart disease and hypertension risk, as well as other serious conditions (e.g., Moravec, 2008; Patel et al., 1981). But reducing stress does not have to be complicated! For example, exercise is a great stress reduction activity (Salmon, 2001) that has a myriad of health benefits. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/15%3A_The_Healthy_Life/15.04%3A__Protecting_Our_Health.txt |
As a student, you probably strive to maintain good grades, to have an active social life, and to stay healthy (e.g., by getting enough sleep), but there is a popular joke about what it’s like to be in college: you can only pick two of these things (see Figure \(1\) for an example). The busy life of a college student doesn’t always allow you to maintain all three areas of your life, especially during test-taking periods. In one study, researchers found that students taking exams were more stressed and, thus, smoked more, drank more caffeine, had less physical activity, and had worse sleep habits (Oaten & Chang, 2005), all of which could have detrimental effects on their health. Positive health practices are especially important in times of stress when your immune system is compromised due to high stress and the elevated frequency of exposure to the illnesses of your fellow students in lecture halls, cafeterias, and dorms.
Psychologists study both health behaviors and health habits. The former are behaviors that can improve or harm your health. Some examples include regular exercise, flossing, and wearing sunscreen, versus negative behaviors like drunk driving, pulling all-nighters, or smoking. These behaviors become habits when they are firmly established and performed automatically. For example, do you have to think about putting your seatbelt on or do you do it automatically? Habits are often developed early in life thanks to parental encouragement or the influence of our peer group.
While these behaviors sound minor, studies have shown that those who engaged in more of these protective habits (e.g., getting 7–8 hours of sleep regularly, not smoking or drinking excessively, exercising) had fewer illnesses, felt better, and were less likely to die over a 9- to 12-year follow-up period (Belloc & Breslow, 1972; Breslow & Enstrom, 1980). For college students, health behaviors can even influence academic performance. For example, poor sleep quality and quantity are related to weaker learning capacity and academic performance (Curcio et al., 2006). Due to the effects that health behaviors can have, much effort is put forward by psychologists to understand how to change unhealthy behaviors, and to understand why individuals fail to act in healthy ways. Health promotion involves enabling individuals to improve health by focusing on behaviors that pose a risk for future ill- ness, as well as spreading knowledge on existing risk factors. These might be genetic risks you are born with, or something you developed over time like obesity, which puts you at risk of Type 2 diabetes and heart disease, among other illnesses.
15.06: Psychology And Medicine
There are many psychological factors that influence medical treatment outcomes. For example, older individuals (Meara et al., 2004), women (Briscoe, 1987), and those from higher socioeconomic backgrounds (Adamson et al., 2008). are all more likely to seek medical care. On the other hand, some individuals who need care might avoid it due to financial obstacles or preconceived notions about medical practitioners or the illness. Thanks to the growing amount of medical information online, many people now use the Internet for health information, and 38% report that this influences their decision to see a doctor (Fox & Jones, 2009). Unfortunately, this is not always a good thing because individuals tend to do a poor job assessing the credibility of health information. For example, college-student participants reading online articles about HIV and syphilis rated a physician’s article and a college student’s article as equally credible if the participants said they were familiar with the health topic (Eastin, 2001). Credibility of health information often means how accurate or trustworthy the information is, and it can be influenced by irrelevant factors, such as the website’s design, logos, or the organization’s contact information (Freeman & Spyridakis, 2004). Similarly, many people post health questions on unmoderated online forums where anyone can respond, which allows for the possibility of inaccurate information being provided for serious medical conditions by unqualified individuals.
After individuals decide to seek care, there is also variability in the information they give their medical provider. Poor communication (e.g., due to embarrassment or feeling rushed) can influence the accuracy of the diagnosis and the effectiveness of the prescribed treatment. Similarly, there is variation following a visit to the doctor. While most individuals are tasked with a health recommendation (e.g., buying and using a medication appropriately, losing weight, going to another expert), not everyone adheres to medical recommendations (Dunbar-Jacob & Mortimer-Stephens, 2010). For example, many individuals take medications inappropriately (e.g., stopping early, not filling prescriptions) or fail to change their behaviors (e.g., quitting smoking). Unfortunately, get- ting patients to follow medical orders is not as easy as one would think. For example, in one study, over one third of diabetic patients failed to get proper medical care that would prevent or slow down diabetes-related blindness (Schoenfeld et al., 2001)! Fortunately, as mobile technology improves, physicians now have the ability to monitor adherence and work to improve it (e.g., with pill bottles that monitor if they are opened at the right time). Even text messages are useful for improving treatment adherence and outcomes in depression, smoking cessation, and weight loss (Cole-Lewis & Kershaw, 2010).
15.07: Being A Health Psychologist
Training as a clinical health psychologist provides a variety of possible career options. Clinical health psychologists often work on teams of physicians, social workers, allied health professionals, and religious leaders. These teams may be formed in locations like rehabilitation centers, hospitals, primary care offices, emergency care centers, or chronic illness clinics. Work in each of these settings will pose unique challenges in patient care, but the primary responsibility will be the same. Clinical health psychologists will evaluate physical, personal, and environmental factors contributing to illness and preventing improved health. In doing so, they will then help create a treatment strategy that takes into account all dimensions of a person’s life and health, which maximizes its potential for success. Those who specialize in health psychology can also conduct
research to discover new health predictors and risk factors, or develop interventions to prevent and treat illness. Researchers studying health psychology work in numerous locations, such as universities, public health departments, hospitals, and private organizations. In the related field of behavioral medicine, careers focus on the application of this type of research. Occupations in this area might include jobs in occupational therapy, rehabilitation, or preventive medicine. Training as a health psychologist provides a wide skill set applicable in a number of different professional settings and career paths. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/15%3A_The_Healthy_Life/15.05%3A_The_Importance_Of_Good_Health_Practices.txt |
Much of the past medical research literature provides an incomplete picture of human health. “Health care” is often “illness care.” That is, it focuses on the management of symptoms and illnesses as they arise. As a result, in many developed countries, we are faced with several health epidemics that are difficult and costly to treat. These include obesity, diabetes, and cardiovascular disease, to name a few. The National Institutes of Health have called for researchers to use the knowledge we have about risk factors to design effective interventions to reduce the prevalence of preventable illness. Additionally, there are a growing number of individuals across developed countries with multiple chronic illnesses and/or lasting disabilities, especially with older age. Addressing their needs and maintaining their quality of life will require skilled individuals who understand how to properly treat these populations. Health psychologists will be on the forefront of work in these areas.
With this focus on prevention, it is important that health psychologists move beyond studying risk (e.g., depression, stress, hostility, low socioeconomic status) in isolation, and move toward studying factors that confer resilience and protection from disease. There is, fortunately, a growing interest in studying the positive factors that protect our health (e.g., Diener & Chan, 2011; Pressman & Cohen, 2005; Richman et al., 2005) with evidence strongly indicating that people with higher positivity live longer, suffer fewer illnesses, and generally feel better. Seligman (2008) has even proposed a field of “Positive Health” to specifically study those who exhibit “above average” health—something we do not think about enough. By shifting some of the research focus to identifying and understanding these health-promoting factors, we may capitalize on this information to improve public health.
Innovative interventions to improve health are already in use and continue to be studied. With recent advances in technology, we are starting to see great strides made to improve health with the aid of computational tools. For example, there are hundreds of simple applications (apps) that use email and text messages to send reminders to take medication, as well as mobile apps that allow us to monitor our exercise levels and food intake (in the growing mobile-health, or m-health, field). These m-health applications can be used to raise health awareness, support treatment and compliance, and remotely collect data on a variety of outcomes. Also exciting are devices that allow us to monitor physiology in real time; for example, to better understand the stressful situations that raise blood pressure or heart rate. With advances like these, health psychologists will be able to serve the population better, learn more about health and health behavior, and develop excel- lent health-improving strategies that could be specifically targeted to certain populations or individuals. These leaps in equipment development, partnered with growing health psychology knowledge and exciting advances in neuroscience and genetic research, will lead health researchers and practitioners into an exciting new time where, hopefully, we will understand more and more about how to keep people healthy.
REFERENCES
Adamson, J., Ben-Shlomo, Y., Chaturvedi, N., & Donovan, J. (2008). Ethnicity, socio-economic position and gender—Do they affect reported health-care seeking behaviour? Social Science & Medicine, 57, 895–904. https://doi.org/10.1016/S0277-9536(02)00458-6
American Psychological Association. (2012). Stress in America 2012 [Press release]. www.apa.org/news/press/releases/stress/2012/ generations.aspx
Belloc, N. B., & Breslow, L. (1972). Relationship of physical health status and health practices. Preventive Medicine, 1, 409–421. doi. org/10.1016/0091-7435(72)90014-X
Billings, A. G., & Moos, R. H. (1981). The role of coping responses and social resources in attenuating the stress of life events. Journal of Behavioral Medicine, 4, 139–157. https://doi.org/10.1007/BF00844267
Breslow, L., & Enstrom, J. E. (1980). Persistence of health habits and their relationship to mortality. Preventive Medicine, 9, 469–483. https:// doi.org/10.1016/0091-7435(80)90042-0
Briscoe, M. E. (1987). Why do people go to the doctor? Sex differences in the correlates of GP consultation. Social Science & Medicine, 25, 507–513. https://doi.org/10.1016/0277-9536(87)90174-2
Carver, C. S., Scheier, M. F., & Weintraub, J. K. (1989). Assessing coping strategies: A theoretically based approach. Journal of Personality and Social Psychology, 56, 267–283. doi.org/10.1037/ 0022-3514.56.2.267
Cohen, S., & Herbert, T. B. (1996). Health psychology: Psychological factors and physical disease from the perspective of human psychoneuroimmunology. Annual Review of Psychology, 47, 113–142. https://doi.org/10.1146/annurev.psych.47.1.113
Cohen, S., & Wills, T. A. (1985). Stress, social support, and the buffering hypothesis. Psychological Bulletin, 98, 310–357. https://doi.org/ 10.1037/0033-2909.98.2.310
Cohen, S., Alper, C. M., Doyle, W. J., Treanor, J. J., & Turner, R. B. (2006). Positive emotional style predicts resistance to illness after experimental exposure to rhinovirus or influenza A virus. Psychosomatic Medicine, 68, 809–815. https://doi.org/10.1097/01. psy.0000245867.92364.3c
Cohen, S., Janicki-Deverts, D., & Miller, G. E. (2007). Psychological stress and disease. Journal of the American Medical Association, 298, 1685–1687. https://doi.org/10.1001/jama.298.14.1685
Cohen, S., Tyrrell, D. A., & Smith, A. P. (1991). Psychological stress and susceptibility to the common cold. New England Journal of Medicine, 325, 606–612. https://doi.org/10.1056/NEJM199108293250903
Cole-Lewis, H., & Kershaw, T. (2010). Text messaging as a tool for behavior change in disease prevention and management. Epidemiologic Reviews, 32, 56–69. doi.org/10.1093/epirev/ mxq004
Curcio, G., Ferrara, M., & De Gennaro, L. (2006). Sleep loss, learning capacity and academic performance. Sleep Medicine Reviews, 10, 323–337. https://doi.org/10.1016/j.smrv.2005.11.001
DeLongis, A., Folkman, S., & Lazarus, R. S. (1988). The impact of daily stress on health and mood: Psychological and social resources as mediators. Journal of Personality and Social Psychology, 54, 486–495. https://doi.org/10.1037/0022-3514.54.3.486
Dickerson, S. S., & Kemeny, M. E. (2004). Acute stressors and cortisol responses: A theoretical integration and synthesis of laboratory research. Psychological Bulletin, 130, 355–391. https://doi.org/ 10.1037/0033-2909.130.3.355
Diener, E., & Chan, M. Y. (2011). Happy people live longer: Subjective well-being contributes to health and longevity. Applied Psychology: Health and Well-Being, 3(1), 1–43. doi.org/10.1111/ j.1758-0854.2010.01045.x
Dunbar-Jacob, J., & Mortimer-Stephens, M. (2001). Treatment adherence in chronic disease. Journal of Clinical Epidemiology, 54(12), S57–S60. https://doi.org/10.1016/S0895-4356(01)00457-7
Eastin, M. S. (2001). Credibility assessments of online health information: The effects of source expertise and knowledge of content. Journal of Computer Mediated Communication, 6. doi.org/10.1111/ j.1083-6101.2001.tb00126.x
Fox, S. & Jones, S. (2009). The social life of health information. Pew Internet and American Life Project, California HealthCare Foundation. https://www.pewresearch.org/internet/2009/06/11/ the-social-life-of-health-information
Fredrickson, B. L. (2001). The role of positive emotions in positive psychology: The broaden-and-build theory of positive emotions. American Psychologist, 56, 218–226. doi.org/10.1037/ 0003-066X.56.3.218
Freeman, K. S., & Spyridakis, J. H. (2004). An examination of factors that affect the credibility of online health information. Technical Communication, 51, 239–263.
Friedman, M., & Rosenman, R. (1959). Association of specific overt behavior pattern with blood and cardiovascular findings. Journal of the American Medical Association, 169, 1286–1296. doi. org/10.1001/jama.1959.03000290012005
Glass, D. C., & Singer, J. E. (1972). Behavioral aftereffects of unpredictable and uncontrollable aversive events. American Scientist, 60, 457–465.
Herman-Stabl, M. A., Stemmler, M., & Petersen, A. C. (1995). Approach and avoidant coping: Implications for adolescent mental health. Journal of Youth and Adolescence, 24, 649–665. doi. org/10.1007/BF01536949
Holt-Lunstad, J., Smith, T. B., & Layton, J. B. (2010). Social relationships and mortality risk: A meta-analytic review. PLoS Medicine, 7(7), e1000316. https://doi.org/10.1371/journal.pmed.1000316
House, J. S., Landis, K. R., & Umberson, D. (1988). Social relationships and health. Science, 241, 540–545. doi.org/10.1126/ science.3399889
Iribarren, C., Sidney, S., Bild, D. E., Liu, K., Markovitz, J. H., Roseman,
J. M., & Matthews, K. (2000). Association of hostility with coronary artery calcification in young adults. Journal of the American Medical Association, 283, 2546–2551. doi.org/10.1001/jama. 283.19.2546
Kubzansky, L. D., Sparrow, D., Vokonas, P., & Kawachi, I. (2001). Is the glass half empty or half full? A prospective study of optimism and coronary heart disease in the normative aging study. Psychosomatic Medicine, 63, 910–916. doi.org/10.1097/ 00006842-200111000-00009
Matthews, K. A., Glass, D. C., Rosenman, R. H., & Bortner, R. W. (1977). Competitive drive, pattern A, and coronary heart disease: A further analysis of some data from the Western Collaborative Group Study. Journal of Chronic Diseases, 30, 489–498. doi.org/10.1016/ 0021-9681(77)90071-6
Meara, E., White, C., & Cutler, D. M. (2004). Trends in medical spending by age, 1963–2000. Health Affairs, 23, 176–183. https://doi.org/ 10.1377/hlthaff.23.4.176
Miller, T. Q., Smith, T. W., Turner, C. W., Guijarro, M. L., & Hallet, A. J. (1996). Meta-analytic review of research on hostility and physical health. Psychological Bulletin, 119, 322–348. https://doi.org/ 10.1037/0033-2909.119.2.322
Moravec, C. S. (2008). Biofeedback therapy in cardiovascular disease: Rationale and research overview. Cleveland Clinic Journal of Medicine, 75(Suppl. 2), S35–S38.
Nes, L. S., & Segerstrom, S. C. (2006). Dispositional optimism and coping: A meta-analytic review. Personality and Social Psychology Review, 10, 235–251. https://doi.org/10.1207/s15327957pspr1003_3
Oaten, M., & Cheng, K. (2005). Academic examination stress impairs self-control. Journal of Social and Clinical Psychology, 24, 254–279. https://doi.org/10.1521/jscp.24.2.254.62276
O’Leary, A. (1985). Self-efficacy and health. Behaviour Research and Therapy, 23, 437–451. https://doi.org/10.1016/0005-7967(85)90172-X
Patel, C., Marmot, M. G., & Terry, D. J. (1981). Controlled trial of biofeedback-aided behavioural methods in reducing mild hypertension. British Medical Journal (Clinical Research Ed.), 282, 2005–2008. doi.org/10.1136/bmj.282.6281.2005
Pressman, S. D., & Cohen, S. (2005). Does positive affect influence health? Psychological Bulletin, 131, 925–971. https://doi.org/ 10.1037/0033-2909.131.6.925
Pressman, S. D., Gallagher, M. W., & Lopez, S. J. (2013). Is the emotion- health connection a “first-world problem”? Psychological Science, 24, 544–549. https://doi.org/10.1177/0956797612457382
Richman, L. S., Kubzansky, L., Maselko, J., Kawachi, I., Choo, P., & Bauer, M. (2005). Positive emotion and health: Going beyond the negative. Health Psychology, 24, 422–429. doi.org/10.1037/ 0278-6133.24.4.422
Rodin, J., & Langer, E. J. (1977). Long-term effects of a control-relevant intervention with the institutionalized aged. Journal of Personality and Social Psychology, 35, 897–902. doi.org/10.1037/ 0022-3514.35.12.897
Rutter, M. (1985). Resilience in the face of adversity. British Journal of Psychiatry, 147, 598–611. https://doi.org/10.1192/bjp.147.6.598
Salmon, P. (2001). Effects of physical exercise on anxiety, depression, and sensitivity to stress: A unifying theory. Clinical Psychology Review, 21(1), 33–61. https://doi.org/10.1016/S0272-7358(99)00032-X
Scheier, M. F., & Carver, C. S. (1985). Optimism, coping, and health: assessment and implications of generalized outcome expectancies. Health Psychology, 4, 219–247. doi.org/10.1037/ 0278-6133.4.3.219
Schoenfeld, E. R., Greene, J. M., Wu, S. Y., & Leske, M. C. (2001). Patterns of adherence to diabetes vision care guidelines: Baseline findings from the Diabetic Retinopathy Awareness Program. Ophthalmology, 108, 563–571. https://doi.org/10.1016/S0161-6420(00)00600-X
Schulz, R., & Hanusa, B.H. (1978). Long-term effects of control and predictability-enhancing interventions: Findings and ethical issues. Journal of Personality and Social Psychology, 36, 1194–1202. https:// doi.org/10.1037/0022-3514.36.11.1194
Segerstrom, S. C., Taylor, S. E., Kemeny, M. E., & Fahey, J. L. (1998). Optimism is associated with mood, coping, and immune change in response to stress. Journal of Personality and Social Psychology, 74, 1646–1655. https://doi.org/10.1037/0022-3514.74.6.1646
Seligman, M. E. P. (2008). Positive health. Applied Psychology, 57, 3–18. doi.org/10.1111/j.1464-0597.2008.00351.x
Selye, H. (1946). The general adaptation syndrome and the diseases of adaptation. Journal of Clinical Endocrinology & Metabolism, 6, 117–230. https://doi.org/10.1210/jcem-6-2-117
Sieber, W. J., Rodin, J., Larson, L., Ortega, S., Cummings, N., Levy, S., Whiteside, T., & Herberman, R. (1992). Modulation of human natural
killer cell activity by exposure to uncontrollable stress. Brain, Behavior, and Immunity, 6, 141–156. doi.org/10.1016/ 0889-1591(92)90014-F
Spiegel, D., Kraemer, H., Bloom, J., & Gottheil, E. (1989). Effect of psychosocial treatment on survival of patients with metastatic breast cancer. The Lancet, 334, 888–891. doi.org/10.1016/ S0140-6736(89)91551-1
Taylor, S. E. (2012). Health psychology (8th ed.). McGraw-Hill.
Taylor, S. E., Klein, L. C., Lewis, B. P., Gruenewald, T. L., Gurung, R. A., &
Updegraff, J. A. (2000). Biobehavioral responses to stress in females: Tend-and-befriend, not fight-or-flight. Psychological Review, 107, 411–429. https://doi.org/10.1037/0033-295X.107.3.411
Twisk, J. W. R., Snel, J., Kemper, H. C. G., & van Mechelen, W. (1999). Changes in daily hassles and life events and the relationship with coronary heart disease risk factors: A 2-year longitudinal study in 27–29-year-old males and females. Journal of Psychosomatic Research, 46, 229–240. doi.org/10.1016/S0022- 3999(98)00088-9
Wallston, B. S., & Wallston, K. A. (1978). Locus of control and health: a review of the literature. Health Education & Behavior, 6, 107–117.
World Health Organization. (1946). Preamble to the Constitution of the World Health Organization. https://www.who.int/about/who-we-are/ constitution
World Health Organization. (2017). Cardiovascular diseases (CVDs). https://www.who.int/en/news-room/fact-sheets/detail/ cardiovascular-diseases-(cvds) | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/15%3A_The_Healthy_Life/15.08%3A_The_Future_Of_Health_Psychology.txt |
While we all experience altered states of consciousness in the form of sleep on a regular basis, some people use drugs and other substances that also result in altered states of consciousness. This section will present information relating to the use of various psychoactive drugs and problems associated with such use. This will be followed by brief descriptions of the effects of some of the more well-known drugs commonly used today.
16.02: Substance Use Disorders
The fifth edition of the Diagnostic and Statistical Manual of Mental Disorders (DSM-5) is used by clinicians to diagnose individuals suffering from various psychological disorders. Drug use disorders are addictive disorders, and the criteria for specific substance (drug) use disorders are described in DSM-5. A person who has a substance use disorder often uses more of the substance than they originally intended to and continues to use that substance despite experiencing significant adverse consequences. In individuals diagnosed with a substance use disorder, there is a compulsive pattern of drug use that is often associated with both physical and psychological dependence.
Physical dependence involves changes in normal bodily functions—the user will experience withdrawal from the drug upon cessation of use. In contrast, a person who has psychological dependence has an emotional, rather than physical, need for the drug and may use the drug to relieve psychological distress. Tolerance is linked to physiological dependence, and it occurs when a person requires more and more drug to achieve effects previously experienced at lower doses. Tolerance can cause the user to increase the amount of drug used to a dangerous level—even to the point of overdose and death.
Drug withdrawal includes a variety of negative symptoms experienced when drug use is discontinued. These symptoms usually are opposite the effects of the drug. For example, withdrawal from sedative drugs often produces unpleasant arousal and agitation. In addition to withdrawal, many individuals who are diagnosed with substance use disorders will also develop tolerance to these substances. Psychological dependence, or drug craving, is a recent addition to the diagnostic criteria for substance use disorder in DSM-5. This is an important factor because we can develop tolerance to and experience withdrawal from any number of drugs that we do not abuse. In other words, physical dependence in and of itself is of limited utility in determining whether or not some- one has a substance use disorder.
16.03: Drug Categories
The effects of all psychoactive drugs occur through their interactions with our endogenous neurotransmitter systems. Many of these drugs, and their relationships, are shown in Figure \(1\). As you have learned, drugs can act as agonists or antagonists of a given neurotransmitter system. An agonist facilitates the activity of a neurotransmitter system, and antagonists impede neurotransmitter activity.
Alcohol and Other Depressants
Ethanol, which we commonly refer to as alcohol, is in a class of psychoactive drugs known as depressants (Figure \(2\)). A depressant is a drug that tends to suppress central nervous system activity. Other depressants include barbiturates and benzodiazepines. These drugs share in common their ability to serve as agonists of the gamma-aminobutyric acid (GABA) neurotransmitter system. Because GABA has a quieting effect on the brain, GABA agonists also have a quieting effect; these types of drugs are often prescribed to treat both anxiety and insomnia.
Acute alcohol administration results in a variety of changes to consciousness. At rather low doses, alcohol use is associated with feelings of euphoria. As the dose increases, people report feeling sedated. Generally, alcohol is associated with decreases in reaction time and visual acuity, lowered levels of alertness, and reduction in behavioral control. With excessive alcohol use, a person might experience a complete loss of consciousness and/or difficulty remembering events that occurred during a period of intoxication (McKim & Hancock, 2013). In addition, if a pregnant woman consumes alcohol, her infant may be born with a cluster of birth defects and symptoms collectively called fetal alcohol spectrum disorder (FASD) or fetal alcohol syndrome (FAS).
With repeated use of many central nervous system depressants, such as alcohol, a person becomes physically dependent upon the substance and will exhibit signs of both tolerance and withdrawal. Psychological dependence on these drugs is also possible. Therefore, the abuse potential of central nervous system depressants is relatively high.
Drug withdrawal is usually an aversive experience, and it can be a life-threatening process in individuals who have a long history of very high doses of alcohol and/or barbiturates. This is of such concern that people who are trying to over- come addiction to these substances should only do so under medical supervision.
Stimulants
Stimulants are drugs that tend to increase overall levels of neural activity. Many of these drugs act as agonists of the dopamine neurotransmitter system. Dopamine activity is often associated with reward and craving; therefore, drugs that affect dopamine neurotransmission often have abuse liability. Drugs in this category include cocaine, amphetamines (including methamphetamine), cathinones (i.e., bath salts), MDMA (ecstasy), nicotine, and caffeine.
Cocaine can be taken in multiple ways. While many users snort cocaine, intravenous injection and ingestion are also common. The freebase version of cocaine, known as crack, is a potent, smokable version of the drug. Like many other stimulants, cocaine agonizes the dopamine neurotransmitter system by blocking the reuptake of dopamine in the neuronal synapse.
LINK TO LEARNING
Read this interesting newspaper article describing myths about crack cocaine.
Amphetamines have a mechanism of action quite similar to cocaine in that they block the reuptake of dopamine in addition to stimulating its release (Figure \(3\)). While amphetamines are often abused, they are also commonly pre- scribed to children diagnosed with attention deficit hyperactivity disorder (ADHD). It may seem counterintuitive that stimulant medications are prescribed to treat a disorder that involves hyperactivity, but the therapeutic effect comes from increases in neurotransmitter activity within certain areas of the brain associated with impulse control.
DIG BEEPER
Crack Cocaine
Crack (Figure \(4\)) is often considered to be more addictive than cocaine itself because it is smokable and reaches the brain very quickly. Crack is often less expensive than other forms of cocaine; therefore, it tends to be a more accessible drug for individuals from impoverished segments of society. During the 1980s, many drug laws were rewritten to punish crack users more severely than cocaine users. This led to discriminatory sentencing with low-income, inner-city minority populations receiving the harshest punishments. The wisdom of these laws has recently been called into question, especially given research that suggests crack may not be more addictive than other forms of cocaine, as previously thought (Haasen & krausz, 2001; Reinerman, 2007). ■
In recent years, methamphetamine (meth) use has become increasingly widespread. Methamphetamine is a type of amphetamine that can be made from ingredients that are readily available (e.g., medications containing pseudoephedrine, a compound found in many over-the-counter cold and flu remedies). Despite recent changes in laws designed to make obtaining pseudoephedrine more difficult, methamphetamine continues to be an easily accessible and relatively inexpensive drug option (Shukla et al., 2012).
Users of cocaine, amphetamine, cathinones, and MDMA seek a euphoric high—feelings of intense elation and plea- sure—especially those who take the drug via intravenous injection or smoking. Repeated use of these stimulants can have significant adverse consequences. Users can experience physical symptoms that include nausea, elevated blood pressure, and increased heart rate. In addition, these drugs can cause feelings of anxiety, hallucinations, and paranoia (Fiorentini et al., 2011). Normal brain functioning is altered after repeated use of these drugs. For example, repeated use can lead to over- all depletion among the monoamine neurotransmitters (dopamine, norepinephrine, and serotonin). People may engage in compulsive use of these stimulant substances in part to try to reestablish normal levels of these neurotransmitters (Jayanthi & Ramamoorthy, 2005; Rothman et al., 2007).
Caffeine is another stimulant drug. While it is probably the most commonly used drug in the world, the potency of this particular drug pales in comparison to the other stimulant drugs described in this section. Generally, people use caffeine to maintain increased levels of alertness and arousal. Caffeine is found in many common medicines (such as weight loss drugs), beverages, foods, and even cosmetics (Herman & Herman, 2013). While caffeine may have some indirect effects on dopamine neurotransmission, its primary mechanism of action involves antagonizing adenosine activity (Porkka-Heiskanen, 2011).
While caffeine is generally considered a relatively safe drug, high blood levels of caffeine can result in insomnia, agitation, muscle twitching, nausea, irregular heartbeat, and even death (Reissig et al., 2009; Wolk et al., 2012). In 2012, Kromann and Nielson reported on a case study of a 40-year- old woman who suffered significant ill effects from her use of caffeine. The woman had used caffeine in the past to boost her mood and to provide energy, but over the course of several years, she increased her caffeine consumption to the point that she was consuming three liters of soda each day. Although she had been taking a prescription antidepressant, her symptoms of depression continued to worsen and she began to suffer physically, displaying significant warning signs of cardiovascular disease and diabetes. Upon admission to an outpatient clinic for treatment of mood disorders, she met all of the diagnostic criteria for substance dependence and was advised to dramatically limit her caffeine intake. Once she was able to limit her use to less than 12 ounces of soda a day, both her mental and physical health gradually improved. Despite the prevalence of caffeine use and the large number of people who confess to suffering from caffeine addiction, this was the first published description of soda dependence appearing in scientific literature.
Nicotine is highly addictive, and the use of tobacco products is associated with increased risks of heart disease, stroke, and a variety of cancers. Nicotine exerts its effects through its interaction with acetylcholine receptors. Acetylcholine functions as a neurotransmitter in motor neurons. In the central nervous system, it plays a role in arousal and reward mechanisms. Nicotine is most commonly used in the form of tobacco products like cigarettes or chewing tobacco; there- fore, there is a tremendous interest in developing effective smoking cessation techniques. To date, people have used a variety of nicotine replacement therapies in addition to various psychotherapeutic options in an attempt to discontinue their use of tobacco products. In general, smoking cessation programs may be effective in the short term, but it is unclear whether these effects persist (Cropley et al., 2008; Levitt et al., 2007; Smedslund et al., 2004).
Opioids
An opioid is one of a category of drugs that includes heroin, morphine, methadone, and codeine. Opioids have analgesic properties; that is, they decrease pain. Humans have an endogenous opioid neurotransmitter system—the body makes small quantities of opioid compounds that bind to opioid receptors, reducing pain and producing euphoria. Thus, opioid drugs, which mimic this endogenous painkilling mechanism, have an extremely high potential for abuse. Natural opioids, called opiates, are derivatives of opium, which is a naturally occur- ring compound found in the poppy plant. There are now several synthetic versions of opiate drugs (correctly called opioids) that have very potent painkilling effects, and they are often abused. For example, the National Institutes of Drug Abuse has sponsored research that suggests the misuse and abuse of the prescription pain killers hydrocodone and oxycodone are significant public health concerns (Maxwell, 2006). In 2013, the U.S. Food and Drug Administration recommended tighter controls on their medical use.
Historically, heroin has been a major opioid drug of abuse (Figure \(5\)). Heroin can be snorted, smoked, or injected intravenously. Like the stimulants described earlier, the use of heroin is associated with an initial feeling of euphoria followed by periods of agitation. Because heroin is often administered via intravenous injection, users often bear needle track marks on their arms and, like all abusers of intravenous drugs, have an increased risk for contraction of both tuberculosis and HIV.
Aside from their utility as analgesic drugs, opioid-like compounds are often found in cough suppressants, anti-nausea, and anti-diarrhea medications. Given that withdrawal from a drug often involves an experience opposite to the effect of the drug, it should be no surprise that opioid withdrawal resembles a severe case of the flu. While opioid withdrawal can be extremely unpleasant, it is not life-threatening (Julien, 2005). Still, people experiencing opioid withdrawal may be given methadone to make withdrawal from the drug less difficult. Methadone is a synthetic opioid that is less euphorigenic than heroin and similar drugs. Methadone clinics help people who previously struggled with opioid addiction manage withdrawal symptoms through the use of methadone. Other drugs, including the opioid buprenorphine, have also been used to alleviate symptoms of opiate withdrawal.
Codeine is an opioid with relatively low potency. It is often prescribed for minor pain, and it is available over-the-counter in some other countries. Like all opioids, codeine does have abuse potential. In fact, abuse of prescription opioid medications is becoming a major concern worldwide (Aquina et al., 2009; Casati et al., 2012).
Hallucinogens
A hallucinogen is one of a class of drugs that results in pro- found alterations in sensory and perceptual experiences (Figure \(6\)). In some cases, users experience vivid visual hallucinations. It is also common for this type of drug to cause hallucinations of body sensations (e.g., feeling as if you are a giant) and a skewed perception of the passage of time.
As a group, hallucinogens are incredibly varied in terms of the neurotransmitter systems they affect. Mescaline and LSD are serotonin agonists, and PCP (angel dust) and ketamine (an animal anesthetic) act as antagonists of the NMDA glutamate receptor. In general, these drugs are not thought to possess the same sort of abuse potential as other classes of drugs discussed in this section.
DIG DEEPER
Medical Marijuana
While the possession and use of marijuana is illegal in most states, it is now legal in Washington and Colorado to possess limited quantities of marijuana for recreational use (Figure \(7\)). In contrast, medical marijuana use is now legal in nearly half of the United States and in the District of Columbia. Medical marijuana is marijuana that is prescribed by a doctor for the treatment of a health condition. For example, people who undergo chemotherapy will often be prescribed marijuana to stimulate their appetites and prevent excessive weight loss resulting from the side effects of chemotherapy treatment. Marijuana may also have some promise in the treatment of a variety of medical conditions (Mather et al., 2013; Robson, 2014; Schicho & Storr, 2014).
While medical marijuana laws have been passed on a state-by-state basis, federal laws still classify this as an illicit substance, making conducting research on the potentially beneficial medicinal uses of marijuana problematic. There is quite a bit of controversy within the scientific community as to the extent to which marijuana might have medicinal benefits due to a lack of large-scale, controlled research (Bostwick, 2012). As a result, many scientists have urged the federal government to allow for relaxation of current marijuana laws and classifications in order to facilitate a more widespread study of the drug’s effects (Aggarwal et al., 2009; Bostwick, 2012; kogan & Mechoulam, 2007).
Until recently, the United States Department of Justice routinely arrested people involved and seized marijuana used in medicinal settings. In the latter part of 2013, however, the United States Department of Justice issued statements indicating that they would not continue to challenge state medical marijuana laws. This shift in policy may be in response to the scientific community’s recommendations and/or reflect changing public opinion regarding marijuana. ■
LINK TO LEARNING
To learn more about some of the most commonly abused prescription and street drugs, check out the Commonly Abused Drugs Chart and the Commonly Abused Prescription Drugs Chart from the National Institute on Drug Abuse.
REFERENCES
Aggarwal, S. K., Carter, G. T., Sullivan, M. D., ZumBrunnen, C., Morrill, R., & Mayer, J. D. (2009). Medicinal use of cannabis in the United States: Historical perspectives, current trends, and future directions. Journal of Opioid Management, 5, 153–168. doi.org/10.5055/ jom.2009.0016
Aquina, C. T., Marques-Baptista, A., Bridgeman, P., & Merlin, M. A. (2009). Oxycontin® abuse and overdose. Postgraduate Medicine, 121, 163–167. https://doi.org/10.3810/pgm.2009.03.1988
Bostwick, J. M. (2012). Blurred boundaries: The therapeutics and politics of medical marijuana. Mayo Clinic Proceedings, 87, 172–186. https:// doi.org/10.1016/j.mayocp.2011.10.003
Casati, A., Sedefov, R., & Pfeiffer-Gerschel, T. (2012). Misuse of medicines in the European Union: A systematic review of the literature.
European Addiction Research, 18, 228–245. doi. org/10.1159/000337028
Cropley, M., Theadom, A., Pravettoni, G., & Webb, G. (2008). The effectiveness of smoking cessation interventions prior to surgery: A systematic review. Nicotine and Tobacco Research, 10, 407–412. https://doi.org/10.1080/14622200801888996
Fiorentini, A., Volonteri, L.S., Dragogna, F., Rovera, C., Maffini, M., Mauri, M. C., & Altamura, C. A. (2011). Substance-induced psychoses: A critical review of the literature. Current Drug Abuse Reviews, 4, 228–240. https://doi.org/10.2174/1874473711104040228
Haasen, C., & Krausz, M. (2001). Myths versus evidence with respect to cocaine and crack: Learning from the US experience. European Addiction Research, 7, 159–160. https://doi.org/10.1159/000050736
Herman, A., & Herman, A. P. (2013). Caffeine’s mechanisms of action and its cosmetic use. Skin Pharmacology and Physiology, 26, 8–14. https://doi.org/10.1159/000343174
Jayanthi, L. D., & Ramamoorthy, S. (2005). Regulation of monoamine transporters: Influence of psychostimulants and therapeutic antidepressants. The AAPS Journal, 7, E728–E738. https://doi.org/ 10.1208/aapsj070373
Julien, R. M. (2005). Opioid analgesics. In A primer of drug action: A comprehensive guide to the actions, uses, and side effects of psychoactive drugs (10th ed., pp. 461–500). Worth.
Kogan, N. M., & Mechoulam, R. (2007). Cannabinoids in health and disease. Dialogues in Clinical Neuroscience, 9, 413–430. https:// www.ncbi.nlm.nih.gov/pubmed/18286801
Levitt, C., Shaw, E., Wong, S., & Kaczorowski, J. (2007). Systematic review of the literature on postpartum care: Effectiveness of interventions for smoking relapse prevention, cessation, and reduction in postpartum women. Birth, 34, 341–347. https://doi.org/ 10.1111/j.1523-536X.2007.00194.x
Mather, L. E., Rauwendaal, E. R., Moxham-Hall, V. L., & Wodak, A. D. (2013). (Re)introducing medicinal cannabis. The Medical Journal of Australia, 199, 759–761. doi.org/10.5694/mja13.10728
Maxwell, J. C. (2006). Trends in the abuse of prescription drugs. Gulf Coast Addiction Technology Transfer Center. https://attchub.org/ userfiles/file/GulfCoast/PrescriptionTrends_Web.pdf
McKim, W. A., & Hancock, S. D. (2013). Drugs and behavior: An introduction to behavioral pharmacology (7th ed.). Pearson.
Porkka-Heiskanen, T. (2011). Methylxanthines and sleep. Handbook of Experimental Pharmacology, 200, 331–348. doi.org/10.1007/ 978-3-642-13443-2_12
Reinerman, C. (2007, October 14). 5 myths about that demon crack. Washington Post. www.washingtonpost.com/wp-dyn/content/ article/2007/10/09/AR2007100900751.html
Reissig, C. J., Strain, E. C., & Griffiths, R. R. (2009). Caffeinated energy drinks—A growing problem. Drug and Alcohol Dependence, 99, 1–10. https://doi.org/10.1016/j.drugalcdep.2008.08.001
Robson, P. J. (2014). Therapeutic potential of cannabinoid medicines. Drug Testing and Analysis, 6, 24–30. doi.org/10.1002/dta.1529 Rothman, R. B., Blough, B. E., & Baumann, M. H. (2007). Dual dopamine/ serotonin releasers as potential medications for stimulant and alcohol addictions. The AAPS Journal, 9, E1–E10. https://doi.org/ 10.1007/978-0-387-76678-2_19
Schicho, R., & Storr, M. (2014). Cannabis finds its way into treatment of Crohn’s disease. Pharmacology, 93, 1–3. doi.org/10.1159/ 000356512
Shukla, R. K, Crump, J. L., & Chrisco, E. S. (2012). An evolving problem: Methamphetamine production and trafficking in the United States. International Journal of Drug Policy, 23, 426–435. https://doi.org/ 10.1016/j.drugpo.2012.07.004
Smedslund, G., Fisher, K. J., Boles, S. M., & Lichtenstein, E. (2004). The effectiveness of workplace smoking cessation programmes: A meta-analysis of recent studies. Tobacco Control, 13, 197–204. doi.org/10.1136/tc.2002.002915
U.S. Food and Drug Administration. (2013, October 24). Statement on proposed hydrocodone reclassification from Janet Woodcock, M.D., Director, Center for Drug Evaluation and Research. wayback. archive-it.org/7993/20170112130311/http://www.fda.gov/Drugs/ DrugSafety/ucm372089.htm
Wolk, B. J., Ganetsky, M., & Babu, K. M. (2012). Toxicity of energy drinks. Current Opinion in Pediatrics, 24, 243–251. doi.org/10.1097/ MOP.0b013e3283506827 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/16%3A_Substance_Use_and_Abuse/16.01%3A_Introduction.txt |
One cannot overstate the significance of the contributions Ivan Pavlov made to the study of predictive learning. Pavlov introduced a level of rigor and precision of measurement of both the independent and dependent variables in animal learning that did not exist at the time. In 1904, Pavlov, a physiologist, was awarded the Nobel Prize in Medicine for his research investigating the digestive process in dogs. He became fascinated by an observation he and his laboratory assistants made while conducting this research. One of the digestive processes they studied was salivation. Saliva contains enzymes that initiate the process of breaking down what one eats into basic nutrients required to fuel and repair the body. The subjects frequently started salivating before being placed in the experimental apparatus. Pavlov described this salivation as a “psychic secretion” since it was not being directly elicited by food. He considered the phenomenon so important that within a few years he abandoned his research program in digestion and dedicated the rest of his professional career to systematically studying the details of this basic learning process.
This is a wonderful example of what has been described as serendipity, or accidental discovery in science. Dogs have been domesticated for thousands of years. A countless number of people probably observed dogs appearing to predict (i.e., anticipate or expect) food. Pavlov, however, recognized the significance of the observation as an example of a fundamental learning process. We often think of science as requiring new observations. Pavlov’s “discovery” of the classical conditioning process is an example of how this is not necessarily the case. One of the characteristics of an exceptional scientist is to recognize the significance of commonly occurring observations.
We will now review the apparatus, methods, and terminology Pavlov developed for studying predictive learning. He adapted an experimental apparatus designed for one scientific field of inquiry (the physiology of digestion) to an entirely different field (adaptive learning). Pavlov made a small surgical incision in the dog’s cheek and implanted a tube permitting saliva to be directly collected in a graduated test tube. The amount of saliva could then be accurately measured and graphed as depicted in Figure \(2\). Predictive learning was inferred when salivation occurred to a previously neutral stimulus as the result of appropriate experience.
VIDEO CLIP
Watch this video describing Pavlov and classical conditioning.
Animals inherit the tendency to make simple responses (i.e., reflexes) to specific types of stimulation. Pavlov’s salivation research was based on the reflexive eliciting of salivation by food (e.g., meat powder). This research was adapted to the study of predictive learning by including a neutral stimulus. By neutral, we simply mean that this stimulus did not initially elicit any behavior related to food. Pavlov demonstrated that if a neutral stimulus preceded a biologically significant stimulus on several occasions, one would see a new response occurring to the previously neutral stimulus. Figure \(3\) uses the most popular translation of Pavlov’s terminology (which he wrote in Russian). The reflexive behavior was referred to as the unconditioned response (UR). The stimulus that reflexively elicited this response was referred to as the unconditioned stimulus (US). A novel stimulus, by virtue of being paired in a predictive relation- ship with the food (US), acquires the capacity to elicit a food-related, conditioned response (CR). Once acquiring this capacity, the novel stimulus is considered a conditioned stimulus (CS). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/17%3A_Predictive_Learning/17.01%3A_Introduction.txt |
If predictive learning is a lawful process, controlled empirical investigation has the potential to establish reliable cause-effect relationships. We will see this is the case as we review several basic classical conditioning phenomena. Many of these phenomena were discovered and named by Pavlov himself, starting with the acquisition process described earlier.
Acquisition
The term acquisition refers to a procedure or process whereby one stimulus is presented in a predictive relationship with another stimulus. Predictive learning (classical conditioning) is inferred from the occurrence of a new response to the first stimulus. Keeping in mind that mentalistic terms are inferences based upon behavioral observations, it is as though the individual learns to predict if this happens, then that happens.
Extinction
The term extinction refers to a procedure or process whereby a previously established predictive stimulus is no longer followed by the second stimulus. This typically results in a weakening in the strength of the prior learned response. It is as though the individual learns what used to happen, doesn’t happen anymore. Extinction is commonly misused as a term describing only the result of the procedure or process. That is, it is often used like the term schizophrenia, which is defined exclusively on the dependent variable (symptom) side. Extinction is actually more like influenza, in that it is a true explanation standing for the relationship between a specific independent variable (the procedure) and dependent variable (the change in behavior).
VIDEO CLIP
Watch this video describing classical conditioning acquisition and extinction.
Spontaneous Recovery
The term spontaneous recovery refers to an increase in the strength of the prior learned response after an extended time period lapses between extinction trials. The individual acts as though, perhaps what used to happen, still does.
17.03: Is Extinction Unlearing Or Inhibitory Learning
Pavlov was an excellent example of someone who today would be considered a behavioral neuroscientist. In fact, the full title of his classic book (1927) is Conditioned Reflexes: An Investigation of the Physiological Activity of the Cerebral Cortex. Behavioral neuroscientists study behavior in order to infer underlying brain mechanisms. Thus, Pavlov did not perceive himself as converting from a physiologist into a psychologist when he abandoned his study of digestion to explore the intricacies of classical conditioning. As implied by his “psychic secretion” metaphor, he believed he was continuing to study physiology, turning his attention from studying the digestive system to studying the brain.
One question of interest to Pavlov was the nature of the extinction process. Pavlov assumed that acquisition produced a connection between a sensory neuron representing the conditioned stimulus and a motor neuron eliciting salivation. The reduction in responding resulting from the extinction procedure could result from either breaking this bond (i.e., unlearning) or counteracting it with a competing response. The fact that spontaneous recovery occurs indicates that the bond is not broken during the extinction process. Extinction must involve learning an inhibitory response counteracting the conditioned response. The individual appears to learn that one stimulus no longer predicts another. The conclusion that extinction does not permanently eliminate a previously learned association has important practical and clinical implications. It means that someone who has received treatment for a problem and improved is not the same as a person never requiring treatment in the first place (compare with Bouton, 2000; Bouton & Nelson, 1998). For example, even if someone has quit smoking, there is a greater likelihood of that person’s relapsing than a non-smoker’s acquiring the habit.
17.04: Stimulus Generalization And Discrimination
Imagine if you had to learn to make the same response over and over again to each new situation. Fortunately, this is often not necessary. Stimulus generalization refers to the fact that a previously acquired response will occur in the presence of stimuli other than the original one, the likelihood being a function of the degree of similarity. In Figure \(1\), we see that a response learned to a 500 Hz frequency tone occurs to other stimuli, the percentage of times depending upon how close the frequency is to 500 Hz. It is as though the individual predicts what happens after one event will happen after similar events.
The fact that generalization occurs significantly increases the efficiency of individual learning experiences. However, there are usually limits on the appropriateness of making the same response in different situations. For example, new fathers often beam the first time they hear their infant say “dada.” They are less thrilled when they hear their child call the mailman “dada!” Usually, it is necessary to conduct additional teaching so that the child only says “dada” in the presence of the father. Stimulus discrimination occurs when one stimulus (the S+, e.g., a tone or the father) is predictive of a second stimulus (e.g., food or the word “dada”) but a different stimulus (the S−, e.g., a light or the mailman) is never followed by that second stimulus. Eventually the individual responds to the S+ (tone or father) and not to the S− (light or mailman) as though learning if this happens then that happens, but if this other thing happens that does not happen. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/17%3A_Predictive_Learning/17.02%3A_Basic_Predictive_Learning_Phenomena.txt |
For most of the twentieth century, Pavlov’s originally pro- posed stimulus substitution model of classical conditioning was widely accepted. Pavlov viewed conditioning as a mechanistic (automatic) result of pairing neutral and biologically significant events in time. He believed that the established conditioned stimulus became a substitute for the original unconditioned stimulus. There were four assumptions under- lying this stimulus substitution model:
• Classical conditioning requires a biologically significant stimulus (i.e., US).
• Temporal contiguity between a neutral stimulus and an unconditioned stimulus is necessary for the neutral stimulus to become a conditioned stimulus.
• Temporal contiguity between a neutral stimulus and an unconditioned stimulus is sufficient for the neutral stimulus to become a conditioned stimulus.
• The conditioned response will always resemble if not be identical to the unconditioned response.
Does Classical Conditioning Require a Biologically Significant Stimulus?
Higher-order conditioning is a procedure or process whereby a previously neutral stimulus is presented in a predictive relationship with a second, previously established, predictive stimulus. Learning is inferred from the occurrence of a new response in the presence of this previously neutral stimulus. For example, after pairing the tone with food, it is possible to place the tone in the position of the US by presenting a light immediately before it occurs. Research indicates that a conditioned response (salivation, in this case) will occur to the light even though it was not paired with a biologically significant stimulus (food).
VIDEO CLIP
Watch this video for a demonstration of higher-order conditioning.
is Temporal Contiguity Necessary for Conditioning?
Human beings have speculated about the learning process since at least the time of the early Greek philosophers. Aristotle, in the fourth century bce, proposed three laws of association that he believed applied to human thought and memory. The law of contiguity stated that objects or events occurring close in time (temporal contiguity) or space (spatial contiguity) become associated. The law of similarity stated that we tend to associate objects or events having features in common such that observing one event will prompt recall of similar events. The law of frequency stated that the more often we experience objects or events, the more likely we will be to remember them. In a sense, Pavlov created a methodology permitting empirical testing of Aristotle’s laws. The law that applies in this section is the law of temporal contiguity. Timing effects, like many variables studied scientifically, lend themselves to parametric studies in which the independent variable consists of different values on a dimension. It has been demonstrated that human eyelid conditioning in which a light is followed by a puff of air to the eye is strongest when the puff occurs approximately 500 milliseconds (1⁄2 second) after the light. The strength of conditioning at shorter or longer intervals drops off within tenths of a second. Thus, temporal contiguity appears critical in human eyelid conditioning, consistent with Pavlov’s second assumption.
An Exception: Acquired Taste Aversion
Acquired taste aversion is the only apparent exception to the necessity of temporal contiguity in predictive learning (classical conditioning). This exception can be understood as an evolutionary adaptation to protect animals from food poisoning. Imagine if members of the Nukak—an indigenous tribe in Colombia that is said to be at imminent risk of extinction—got sick after eating a particular food and continued to eat the same substance. There is a good chance the tribe members (and tribe!) would not survive for long. It would be advantageous to avoid foods one ate prior to becoming ill, even if the symptoms did not appear for several minutes or even hours. The phenomenon of acquired taste aversion has been studied extensively. The time intervals used sometimes differ by hours rather than seconds or tenths of seconds. For example, rats were made sick by being exposed to X-rays after drinking sweet water (Smith & Roll, 1967). Rats have a strong preference for sweet water, drinking it approximately 80% of the time when given a choice with ordinary tap water. If the rat became sick within 1⁄2 hour, sweet-water drinking was totally eliminated. With intervals of 1 to 6 hours, it was reduced from 80% to 10%. There was even evidence of an effect after a 24-hour delay! Pavlov’s dogs would not associate a tone with presentation of food an hour later, let alone 24 hours. The acquired aversion to sweet water can be interpreted as either an exception to the law of temporal contiguity or contiguity must be considered on a time scale of different orders of magnitude (hours rather than seconds).
is Temporal Contiguity Sufficient for Conditioning?
Pavlov believed not only that temporal contiguity between CS and US was necessary for conditioning to occur but also that it was all that is necessary (i.e., that it was sufficient). Rescorla (1966, 1968, 1988) has demonstrated that the correlation between CS and US (i.e., the extent to which the CS predicted the US) was more important then temporal contiguity. For example, if the only time one gets shocked is in the presence of the tone, then the tone correlates with shock (i.e., is predictive of the shock). If one is shocked the same amount whether the tone is present or not, the tone does not correlate with shock (i.e., provides no predictive information). Rescorla demonstrated that despite temporal contiguity between tone and shock in both instances, classical conditioning would be strong in the first case and not occur in the second.
Another example of the lack of predictive learning despite temporal contiguity between two events is provided in a study by Leon Kamin (1969). A blocking group received a tone (CS 1) followed by shock (US) in the first phase, while a control group was simply placed in the chamber. The groups were identical from then on. During the second phase, a com- pound stimulus consisting of the light and a tone (CS 2) was followed by shock. During a test phase, each component was presented by itself to determine the extent of conditioning.
In the blocking group, conditioning occurred to the tone and not to the light. Conditioning occurred to both elements of the compound in the control group. It is as though the previous experience with the tone resulted in the blocking-group subjects not paying attention to the light in the second phase. The light was redundant. It did not provide additional information.
A novel and fun demonstration of blocking in college students involved a computerized video game (Arcediano et al., 1997). Subjects used a laser gun (the space bar) to try to protect the earth from invasion by Martians. Unfortunately, the enterprising Martians had developed an anti-laser shield. If the subject fired when the shield was in place, their laser-gun would be ineffective, permitting a bunch of Martians to land and do their mischief. A flashing light preceded implementation of the laser-shield for subjects in the blocking group. A control group did not experience a predictive stimulus for the laser-shield. Subsequently, both groups experienced a com- pound stimulus consisting of the flashing light and a complex tone. The control group associated the tone with activation of the laser-shield, whereas, due to their previous history with the light, the blocking group did not. For them, the tone was redundant.
VIDEO CLIP
Watch this video for a demonstration of blocking.
The blocking procedure demonstrates that temporal contiguity between events, even in a predictive relationship, is not sufficient for learning to occur. In the second phase of the blocking procedure, the compound stimulus precedes the US. According to Pavlov, since both components are contiguous with the US, both should become associated with it and eventually elicit CRs. The combination of Rescorla’s (1966) and Kamin’s (1969) findings lead to the conclusion that learn- ing occurs when individuals obtain new information enabling them to predict events they were unable to previously predict. Kamin suggested that this occurs only when we are surprised. That is, as long as events are proceeding as expected, we do not learn. Once something unexpected occurs, individuals search for relevant information. Many of our activities may be described as “habitual” (Kirsch et al., 2004) or “automatic” (Aarts & Dijksterhuis, 2000). We have all had the experience of riding a bike or driving as though we are on “autopilot.” We are not consciously engaged in steering as long as events are proceeding normally. Once something unexpected occurs we snap to attention and focus on the immediate environmental circumstances. This provides the opportunity to acquire new information. This is a much more active and adaptive under- standing of predictive learning than that provided by Pavlov’s stimulus substitution model (see Rescorla, 1988).
Must the Conditioned Response Resemble the Unconditioned Response?
We will now examine the fourth assumption of Pavlov’s model, that the conditioned response always resembles the unconditioned response. Meat powder reflexively elicits salivation, and Pavlov observed the same reaction to a conditioned stimulus predictive of meat powder. Puffs of air reflexively elicit eye blinks, and taps on the knee elicit knee jerks. The conditioned responses are similar to the unconditioned responses in research involving puffs of air and knee taps as unconditioned stimuli. It is understandable that Pavlov and others believed for so long that the conditioned response must resemble if not be identical to the unconditioned response. However, Zener (1937, p. 393) took movies of dogs undergoing salivary conditioning and disagreed with this conclusion. He observed, “Despite Pavlov’s assertions, the dog does not appear to be eating an imaginary food. . . . It is a different response, anthropomorphically describable as a looking for, expecting, the fall of food with a readiness to perform the eating behavior which will occur when the food falls.”
Kimble (1961, p. 54) offered the possible interpretation that “the function of the conditioned response is to prepare the organism for the occurrence of the unconditioned stimulus.” Research by Shepard Siegel (1975, 1977, 1984, 2005) has swung the pendulum toward widespread acceptance of this interpretation of the nature of the conditioned response. Siegel’s research involved administration of a drug as the unconditioned stimulus. For example, rats were injected with insulin in the presence of a novel stimulus (Siegel, 1975). Insulin is a drug that lowers blood sugar level and is often used to treat diabetics. Eventually, a conditioned response was developed to the novel stimulus (now a CS). However, rather than lowering blood sugar level, the blood sugar level increased to the CS. Siegel described this increase as a compensatory response in preparation for the effect of insulin. He argued that it was similar to other homeostatic mechanisms designed to maintain optimal levels of biological processes (e.g., temperature, white blood cell count, fluid levels, etc.). Similar compensatory responses have been demonstrated with morphine, a drug having analgesic properties (Siegel, 1977) and with caffeine (Siegel, 2005). Siegel (2008) has gone so far as to suggest that “the learning researcher is a homeostasis researcher.”
Siegel has developed a fascinating and influential model of drug tolerance and overdose effects based upon his findings concerning the acquisition of compensatory responses (Siegel, 1983). He suggested that many so-called heroin over- doses are actually the result of the same dosage being consumed differently or in a different environment. Such an effect has actually been demonstrated experimentally with rats. Whereas 34% of rats administered a higher than usual dosage of heroin in the same cage died, 64% administered the same dosage in a different cage died (Siegel et al., 1982). As an experiment, this study has high internal validity but obviously could not be replicated with human subjects. In a study with high external validity, Siegel (1984) interviewed survivors of suspected heroin overdoses. Most insisted they had taken the usual quantity but indicated that they had used a different technique or consumed the drug in a different environment. This combination of high external validity and high internal validity results makes a compelling case for Siegel’s learning model of drug tolerance and overdose effects.
Drug-induced compensatory responses are consistent with the interpretation that the conditioned response constitutes preparation for the unconditioned stimulus. Combining this interpretation with the conclusions reached regarding the necessity of predictiveness for classical conditioning to occur leads to the following alternative to Pavlov’s stimulus substitution model: Classical conditioning is an adaptive process whereby individuals acquire the ability to predict future events and prepare for their occurrence.
EXERCISES
1. Describe the basis for concluding that extinction is an inhibitory as opposed to an unlearning process.
2. Define and give examples of the following classical conditioning phenomena: acquisition, extinction, spontaneous recovery, stimulus generalization, and stimulus discrimination.
3. Explain how the fact that spontaneous recovery occurs indicates that the connection between a conditioned stimulus and conditioned response is not broken during the extinction process.
4. State the assumptions underlying Pavlov’s stimulus substitution model of classical conditioning. Describe the research findings addressing each of the assumptions. Show how the research findings are consistent with the description of classical conditioning as an adaptive learning process.
REFERENCES
Aarts, H., & Dijksterhuis, A. (2000). Habits as knowledge structures: Automaticity in goal-directed behavior. Journal of Personality and Social Psychology, 78, 53–63. doi.org/10.1037/ 0022-3514.78.1.53
Arcediano, F., Matute, H., & Miller, R. R. (1997). Blocking of Pavlovian conditioning in humans. Learning and Motivation, 28(2), 188–199. https://doi.org/10.1006/lmot.1996.0957
Bouton, M. E. (2000). A learning theory perspective on lapse, relapse, and the maintenance of behavior change. Health Psychology, 19(1, Suppl.), 57–63. https://doi.org/10.1037/0278-6133.19.Suppl1.57
Bouton, M. E., & Nelson, J. B. (1998). The role of context in classical conditioning: Some implications for cognitive behavior therapy. In W. T. O’Donohue (Ed.), Learning and behavior therapy (pp. 59–84). Allyn & Bacon.
Kamin, L. J. (1969). Predictability, surprise, attention, and conditioning. In B. A. Campbell & R. M. Church (Eds.), Punishment and aversive behavior. Appleton-Century-Crofts.
Kimble, G. A. (1961). Hilgard and Marquis’ conditioning and learning (2nd ed.). Appleton-Century-Crofts.
Kirsch, I., Lynn, S. J., Vigorito, M., & Miller, R. R. (2004). The role of cognition in classical and operant conditioning. Journal of Clinical Psychology, 60, 369–392. doi.org/10.1002/jclp.10251
Pavlov, I. P. (1927). Conditioned reflexes: An investigation of the physiological activity of the cerebral cortex. Oxford University Press.
Rescorla, R. A. (1966). Predictability and number of pairings in Pavlovian fear conditioning. Psychonomic Science, 4, 383–384. doi. org/10.3758/BF03342350
Rescorla, R. A. (1968). Probability of shock in the presence and absence of CS in fear conditioning. Journal of Comparative and Physiological Psychology, 66, 1–5. https://doi.org/10.1037/h0025984
Rescorla, R. A. (1988). Pavlovian conditioning: It’s not what you think it is. American Psychologist, 43(3), 151–160. doi.org/10.1037/ 0003-066X.43.3.151
Siegel, S. (1975). Conditioning insulin effects. Journal of Comparative and Physiological Psychology, 89(3), 189–199. doi.org/10.1037/ h0076811
Siegel, S. (1977). Morphine tolerance acquisition as an associative process. Journal of Experimental Psychology: Animal Behavior Processes, 3, 1–13. https://doi.org/10.1037/0097-7403.3.1.1
Siegel, S. (1983). Classical conditioning, drug tolerance, and drug dependence. In R. G. Smart, F. B. Glaser, Y. Israel, H. Kalant, R. E. Popham, & W. Schmidt (Eds.), Research advances in alcohol and drug problems (Vol. 7, pp. 202–246). Plenum.
Siegel, S. (1984). Pavlovian conditioning and heroin overdose: Reports by overdose victims. Bulletin of the Psychonomic Society, 22(5), 428–430. https://doi.org/10.3758/BF03333867
Siegel, S. (2005). Drug tolerance, drug addiction, and drug anticipation. Current Directions in Psychological Science, 14(6), 296–300. https:// doi.org/10.1111/j.0963-7214.2005.00384.x
Siegel, S. (2008). Learning and the wisdom of the body. Learning & Behavior, 36(3), 242–252. https://doi.org/10.3758/LB.36.3.242
Siegel, S., Hinson, R. E., Krank, M. D., & McCully, J. (1982). Heroin “overdose” death: The contribution of drug-associated environmental
cues. Science, 216(4544), 436–437. doi.org/10.1126/ science.7200260
Smith, J. C., & Roll, D. L. (1967). Trace conditioning with X-rays as an aversive stimulus. Psychonomic Science, 9(1), 11–12. doi. org/10.3758/BF03330734
Zener, K. (1937). The significance of behavior accompanying conditioned salivary secretion for theories of the conditioned response. American Journal of Psychology, 50(1/4), 384–403. doi.org/10.2307/ 1416644 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/17%3A_Predictive_Learning/17.05%3A_Pavlov%27s_Stimulus_Substitution_Model_Of_Classical_Conditioning.txt |
The previous reading focused on the type of associative learning known as classical conditioning. Remember that in classical conditioning, something in the environment triggers a reflex automatically, and researchers train the organism to react to a different stimulus. Now we turn to the second type of associative learning, operant conditioning. In operant conditioning, organisms learn to associate a behavior and its consequence (Table \(1\)). A pleasant consequence makes that behavior more likely to be repeated in the future. For example, Spirit, a dolphin at the National Aquarium in Baltimore, does a flip in the air when her trainer blows a whistle. The consequence is that she gets a fish.
Table \(1\): Classical and Operant Conditioning Compared
Classical Conditioning
Operant Conditioning
Conditioning approach
An unconditioned stimulus (such as food) is paired with a neutral stimulus (such as a bell). The neutral stimulus eventually becomes the conditioned stimulus, which brings about the conditioned response (salivation).
The target behavior is followed by reinforcement or punishment to either strengthen or weaken it, so that the learner is more likely to exhibit the desired behavior in the future.
Stimulus timing
The stimulus occurs immediately before the response.
The stimulus (either reinforcement or punishment) occurs soon after the response.
Psychologist B. F. Skinner saw that classical conditioning is limited to existing behaviors that are reflexively elicited, and it doesn’t account for new behaviors such as riding a bike. He proposed a theory about how such behaviors come about. Skinner believed that behavior is motivated by the consequences we receive for the behavior: the reinforcements and punishments. His idea that learning is the result of consequences is based on the law of effect, which was first proposed by psychologist Edward Thorndike.
According to the law of effect, behaviors that are followed by consequences that are satisfying to the organism are more likely to be repeated, and behaviors that are followed by unpleasant consequences are less likely to be repeated (Thorndike, 1898). Essentially, if an organism does some- thing that brings about a desired result, the organism is more likely to do it again. If an organism does something that does not bring about a desired result, the organism is less likely to do it again. An example of the law of effect is in employment. One of the reasons (and often the main reason) we show up for work is because we get paid to do so. If we stop getting paid, we will likely stop showing up—even if we love our job.
Working with Thorndike’s law of effect as his foundation, Skinner began conducting scientific experiments on animals (mainly rats and pigeons) to determine how organisms learn through operant conditioning (Skinner, 1938). He placed these animals inside an operant conditioning chamber, which has come to be known as a “Skinner box” (Figure \(1\)). A Skinner box contains a lever (for rats) or disk (for pigeons) that the animal can press or peck for a food reward via the dispenser. Speakers and lights can be associated with certain behaviors. A recorder counts the number of responses made by the animal.
LINK TO LEARNING
Watch this brief video clip to learn more about operant conditioning: Skinner is interviewed, and operant conditioning of pigeons is demonstrated.
In discussing operant conditioning, we use several every- day words—positive, negative, reinforcement, and punish- ment—in a specialized manner. In operant conditioning, positive and negative do not mean good and bad. Instead, positive means you are adding something, and negative means you are taking something away. Reinforcement means you are increasing a behavior, and punishment means you are decreasing a behavior. Reinforcement can be positive or negative, and punishment can also be positive or negative. All reinforces (positive or negative) increase the likelihood of a behavioral response. All punishes (positive or negative) decrease the likelihood of a behavioral response. Now let’s combine these four terms: positive reinforcement, negative reinforcement, positive punishment, and negative punishment (Table \(2\)).
Table \(2\): Positive and Negative Reinforcement and Punishment
Reinforcement
Punishment
Positive
Something is added to increase the likelihood of a behavior.
Something is added to decrease the likelihood of a behavior.
Negative
Something is removed to increase the likelihood of a behavior.
Something is removed to decrease the likelihood of a behavior.
18.02: Reinforcement
The most effective way to teach a person or animal a new behavior is with positive reinforcement. In positive reinforcement, a desirable stimulus is added to increase a behavior.
For example, you tell your 5-year-old son, Jerome, that if he cleans his room, he will get a toy. Jerome quickly cleans his room because he wants a new art set. Let’s pause for a moment. Some people might say, “Why should I reward my child for doing what is expected?” But in fact we are constantly and consistently rewarded in our lives. Our paychecks are rewards, as are high grades and acceptance into our preferred school. Being praised for doing a good job and for passing a driver’s test is also a reward. Positive reinforcement as a learning tool is extremely effective. It has been found that one of the most effective ways to increase achievement in school districts with below-average reading scores was to pay the children to read. Specifically, second-grade students in Dallas were paid \$2 each time they read a book and passed a short quiz about the book. The result was a significant increase in reading comprehension (Fryer, 2010). What do you think about this program? If Skinner were alive today, he would probably think this was a great idea. He was a strong proponent of using operant conditioning principles to influence students’ behavior at school. In fact, in addition to the Skinner box, he also invented what he called a teaching machine that was designed to reward small steps in learning (Skinner, 1961)—an early forerunner of computer-assisted learning. His teaching machine tested students’ knowledge as they worked through various school subjects. If students answered questions correctly, they received immediate positive reinforcement and could continue; if they answered incorrectly, they did not receive any reinforcement. The idea was that students would spend additional time studying the material to increase their chance of being reinforced the next time (Skinner, 1961).
In negative reinforcement, an undesirable stimulus is removed to increase a behavior. For example, car manufacturers use the principles of negative reinforcement in their seatbelt systems, which go “beep, beep, beep” until you fasten your seatbelt. The annoying sound stops when you exhibit the desired behavior, increasing the likelihood that you will buckle up in the future. Negative reinforcement is also used frequently in horse training. Riders apply pressure—by pulling the reins or squeezing their legs—and then remove the pressure when the horse performs the desired behavior, such as turning or speeding up. The pressure is the negative stimulus that the horse wants to remove. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/18%3A_Operant_Conditioning/18.01%3A_Introduction.txt |
Many people confuse negative reinforcement with punishment in operant conditioning, but they are two very different mechanisms. Remember that reinforcement, even when it is negative, always increases a behavior. In contrast, punishment always decreases a behavior. In positive punishment, you add an undesirable stimulus to decrease a behavior. An example of positive punishment is scolding a student to get the student to stop texting in class. In this case, a stimulus (the reprimand) is added in order to decrease the behavior (texting in class). In negative punishment, you remove an aversive stimulus to decrease behavior. For example, when a child misbehaves, a parent can take away a favorite toy. In this case, a stimulus (the toy) is removed in order to decrease the behavior.
Punishment, especially when it is immediate, is one way to decrease undesirable behavior. For example, imagine your 6-year-old son, Brandon, hit his younger brother. You have Brandon write 100 times “I will not hit my brother” (positive punishment). Chances are he won’t repeat this behavior. While strategies like this are common today, in the past children were often subject to physical punishment, such as spanking. It’s important to be aware of some of the drawbacks in using physical punishment on children. First, punishment may teach fear. If Brandon is spanked for running into the street, he may become fearful of the street, but he also may become fearful of the person who delivered the punishment—you, his parent. Similarly, children who are punished by teachers may come to fear the teacher and try to avoid school (Gershoff et al., 2010). Consequently, most schools in the United States have banned corporal punishment. Second, punishment may cause children to become more aggressive and prone to antisocial behavior and delinquency (Gershoff, 2002). They see their parents resort to spanking when they become angry and frustrated, so, in turn, they may act out this same behavior when they become angry and frustrated. For example, because you spank Brenda when you are angry with her for her misbehavior, she might start hitting her friends when they won’t share their toys.
While positive punishment can be effective in some cases, Skinner suggested that the use of punishment should be weighed against the possible negative effects. Today’s psychologists and parenting experts favor reinforcement over pun- ishment—they recommend that you catch your child doing something good and reward her for it.
Shaping
In his operant conditioning experiments, Skinner often used an approach called shaping. Instead of rewarding only the target behavior, in shaping, we reward successive approximations of a target behavior. Why is shaping needed? Remember that in order for reinforcement to work, the organism must first display the behavior. Shaping is needed because it is extremely unlikely that an organism will display anything but the simplest of behaviors spontaneously. In shaping, behaviors are broken down into many small, achievable steps. The specific steps used in the process are the following:
1. Reinforce any response that resembles the desired behavior.
2. Then reinforce the response that more closely resembles the desired behavior. You will no longer reinforce the previously reinforced response.
3. Next, begin to reinforce the response that even more closely resembles the desired behavior.
4. Continue to reinforce closer and closer approximations of the desired behavior.
5. Finally, only reinforce the desired behavior
Shaping is often used in teaching a complex behavior or chain of behaviors. Skinner used shaping to teach pigeons not only such relatively simple behaviors as pecking a disk in a Skinner box, but also many unusual and entertaining behaviors, such as turning in circles, walking in figure eights, and even playing ping pong; the technique is commonly used by animal trainers today. An important part of shaping is stimulus discrimination. Recall Pavlov’s dogs—he trained them to respond to the tone of a bell, and not to similar tones or sounds. This discrimination is also important in operant conditioning and in shaping behavior.
LINK TO LEARNING
Here is a brief video of Skinner’s pigeons playing ping pong.
It’s easy to see how shaping is effective in teaching behaviors to animals, but how does shaping work with humans? Let’s consider parents whose goal is to have their child learn to clean his room. They use shaping to help him master steps toward the goal. Instead of performing the entire task, they set up these steps and reinforce each step. First, he cleans up one toy. Second, he cleans up five toys. Third, he chooses whether to pick up ten toys or put his books and clothes away. Fourth, he cleans up everything except two toys. Finally, he cleans his entire room. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/18%3A_Operant_Conditioning/18.03%3A_Punishment.txt |
Rewards such as stickers, praise, money, toys, and more can be used to reinforce learning. Let’s go back to Skinner’s rats again. How did the rats learn to press the lever in the box? They were rewarded with food each time they pressed the lever. For animals, food would be an obvious reinforcer.
What would be a good reinforcer for humans? For your daughter Sydney, it was the promise of a toy if she cleaned her room. How about Joaquin, the soccer player? If you gave Joaquin a piece of candy every time he made a goal, you would
be using a primary reinforcer. Primary reinforcers are reinforcers that have innate reinforcing qualities. These kinds of reinforcers are not learned. Water, food, sleep, shelter, sex, and touch, among others, are primary reinforcers. Pleasure is also a primary reinforcer. Organisms do not lose their drive for these things. For most people, jumping in a cool lake on a very hot day would be reinforcing and the cool lake would be innately reinforcing—the water would cool the person off (a physical need), as well as provide pleasure.
Everyday Connection
Behavior Modification in Children
Parents and teachers often use behavior modification to change a child’s behavior. Behavior modification uses the principles of operant conditioning to accomplish behavior change so that undesirable behaviors are switched for more socially acceptable ones. Some teachers and parents create a sticker chart, in which several behaviors are listed (Figure \(1\)). Sticker charts are a form of token economies, as described in the text. Each time children perform the behavior, they get a sticker, and after a certain number of stickers, they get a prize, or reinforcer. The goal is to increase accept- able behaviors and decrease misbehavior. Remember, it is best to reinforce desired behaviors, rather than to use punishment. In the classroom, the teacher can reinforce a wide range of behaviors in students, from raising their hands, to walking quietly in the hall, to turning in their homework. At home, parents might create a behavior chart that rewards children for things such as putting away toys, brushing their teeth, and helping with dinner. In order for behavior modification to be effective, the reinforcement needs to be connected with the behavior; the reinforcement must matter to the child and be done consistently.
Time-out is another popular technique used in behavior modification with children. It operates on the principle of negative punishment. When a child demonstrates an undesirable behavior, she is removed from the desirable activity at hand (Figure \(2\)). For example, say that Sophia and her brother Mario are playing with building blocks. Sophia throws some blocks at her brother, so you give her a warning that she will go to time-out if she does it again. A few minutes later, she throws more blocks at Mario. you remove Sophia from the room for a few minutes. When she comes back, she doesn’t throw blocks.
There are several important points that you should know if you plan to implement time-out as a behavior modification technique. First, make sure the child is being removed from a desirable activity and placed in a less desirable location. If the activity is something undesirable for the child, this technique will backfire because it is more enjoyable for the child to be removed from the activity. Second, the length of the time-out is important. The general rule of thumb is one minute for each year of the child’s age. Sophia is five; there- fore, she sits in a time-out for five minutes. Setting a timer helps children know how long they have to sit in time-out. Finally, as a caregiver, keep several guidelines in mind over the course of a time-out: remain calm when directing your child to time-out; ignore your child during time-out (because caregiver attention may reinforce misbehavior); and give the child a hug or a kind word when time-out is over. ■
A secondary reinforcer has no inherent value and only has reinforcing qualities when linked with a primary rein- forcer. Praise, linked to affection, is one example of a secondary reinforcer, as when you called out “Great shot!” every time Joaquin made a goal. Another example, money, is only worth something when you can use it to buy other things— either things that satisfy basic needs (food, water, shelter—all primary reinforcers) or other secondary reinforcers. If you were on a remote island in the middle of the Pacific Ocean and you had stacks of money, the money would not be useful if you could not spend it. What about stickers on a behavior chart? They also are secondary reinforcers.
Sometimes, instead of stickers on a sticker chart, a token is used. Tokens, which are also secondary reinforcers, can then be traded in for rewards and prizes. Entire behavior management systems, known as token economies, are built around the use of these kinds of token reinforcers. Token economies have been found to be very effective at modifying behavior in a variety of settings such as schools, prisons, and mental hospitals. For example, a study by Cangi and Daly (2013) found that use of a token economy increased appropriate social behaviors and reduced inappropriate behaviors in a group of autistic school children. Autistic children tend to exhibit disruptive behaviors such as pinching and hitting. When the children in the study exhibited appropriate behavior (not hitting or pinching), they received a “quiet hands” token. When they hit or pinched, they lost a token. The children could then exchange specified amounts of tokens for minutes of playtime. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/18%3A_Operant_Conditioning/18.04%3A_Primary_And_Secondary_Reinforcers.txt |
Remember, the best way to teach a person or animal a behavior is to use positive reinforcement. For example, Skinner used positive reinforcement to teach rats to press a lever in a Skinner box. At first, the rat might randomly hit the lever while exploring the box, and out would come a pellet of food. After eating the pellet, what do you think the hungry rat did next? It hit the lever again, and received another pellet of food. Each time the rat hit the lever, a pellet of food came out. When an organism receives a reinforcer each time it displays a behavior, it is called continuous reinforcement. This reinforcement schedule is the quickest way to teach someone a behavior, and it is especially effective in training a new behavior. Let’s consider a dog that is learning to sit on command. Now, each time he sits, you give him a treat. Timing is important here: you will be most successful if you present the reinforcer immediately after he sits, so that he can make an association between the target behavior (sitting) and the consequence (getting a treat).
LINK TO LEARNING
Watch this video clip where veterinarian Dr. Sophia Yin shapes a dog’s behavior using the steps outlined above.
Once a behavior is trained, researchers and trainers often turn to another type of reinforcement schedule—partial reinforcement. In partial reinforcement, also referred to as intermittent reinforcement, the person or animal does not get reinforced every time they perform the desired behavior. There are several different types of partial reinforcement schedules (able \(1\)). These schedules are described as either fixed or variable, and as either interval or ratio. Fixed refers to the number of responses between reinforcements, or the amount of time between reinforcements, which is set and unchanging. Variable refers to the number of responses or amount of time between reinforcements, which varies or changes. Interval means the schedule is based on the time between reinforcements, and ratio means the schedule is based on the number of responses between reinforcements.
Reinforcement Schedule
Description
Result
Example
Fixed interval
Reinforcement is delivered at predictable time intervals (e.g., after 5, 10, 15, and 20 minutes).
Moderate response rate with significant pauses after reinforcement
Hospital patient uses patient- controlled, doctor-timed pain relief
Variable interval
Reinforcement is delivered at unpredictable time intervals (e.g., after 5, 7, 10, and 20 minutes).
Moderate yet steady response rate
Checking Facebook
Fixed ratio
Reinforcement is delivered after a predictable number of responses (e.g., after 2, 4, 6, and 8 responses).
High response rate with pauses after reinforcement
Piecework—factory worker getting paid for every x number of items manufactured
Variable ratio
Reinforcement is delivered after an unpredictable number of responses (e.g., after 1, 4, 5, and 9 responses).
High and steady response rate
Gambling
Table \(1\): Reinforcement Schedules
Now let’s combine these four terms. A fixed interval reinforcement schedule is when behavior is rewarded after a set amount of time. For example, June undergoes major surgery in a hospital. During recovery, she is expected to experience pain and will require prescription medications for pain relief. June is given an IV drip with a patient-controlled painkiller. Her doctor sets a limit: one dose per hour. June pushes a but- ton when pain becomes difficult, and she receives a dose of medication. Since the reward (pain relief) only occurs on a fixed interval, there is no point in exhibiting the behavior when it will not be rewarded.
With a variable interval reinforcement schedule, the person or animal gets the reinforcement based on varying amounts of time, which are unpredictable. Say that Manuel is the manager at a fast-food restaurant. Every once in a while someone from the quality control division comes to Manuel’s restaurant. If the restaurant is clean and the service is fast, everyone on that shift earns a \$20 bonus. Manuel never knows when the quality control person will show up, so he always tries to keep the restaurant clean and ensures that his employees provide prompt and courteous service. His productivity regarding prompt service and keeping a clean restaurant are steady because he wants his crew to earn the bonus.
With a fixed ratio reinforcement schedule, there are a set number of responses that must occur before the behavior is rewarded. Carla sells glasses at an eyeglass store, and she earns a commission every time she sells a pair of glasses. She always tries to sell people more pairs of glasses, including prescription sunglasses or a backup pair, so she can increase her commission. She does not care if the person really needs the prescription sun- glasses, Carla just wants her bonus. The quality of what Carla sells does not matter because her commission is not based on quality; it’s only based on the number of pairs sold. This distinction in the quality of performance can help determine which reinforcement method is most appropriate for a particular situation. Fixed ratios are better suited to optimize the quantity of output, whereas a fixed interval, in which the reward is not quantity based, can lead to a higher quality of output.
In a variable ratio reinforcement schedule, the number of responses needed for a reward varies. This is the most powerful partial reinforcement schedule. An example of the variable ratio reinforcement schedule is gambling. Imagine that Sarah—generally a smart, thrifty woman—visits Las Vegas for the first time. She is not a gambler, but out of curiosity she puts a quarter into the slot machine, and then another, and another. Nothing happens. Two dollars in quarters later, her curiosity is fading, and she is just about to quit. But then, the machine lights up, bells go off, and Sarah gets 50 quarters back. That’s more like it! Sarah gets back to inserting quarters with renewed interest, and a few minutes later she has used up all her gains and is \$10 in the hole. Now might be a sensible time to quit. And yet, she keeps putting money into the slot machine because she never knows when the next reinforcement is coming. She keeps thinking that with the next quarter she could win \$50, or \$100, or even more. Because the reinforcement schedule in most types of gambling has a variable ratio schedule, people keep trying and hoping that the next time they will win big. This is one of the reasons that gambling is so addictive—and so resistant to extinction.
In operant conditioning, extinction of a reinforced behavior occurs at some point after reinforcement stops, and the speed at which this happens depends on the reinforcement schedule. In a variable ratio schedule, the point of extinction comes very slowly, as described above in the slot-machine example. But in the other reinforcement schedules, extinction may come quickly. For example, if June presses the button for the pain relief medication before the allotted time her doctor has approved, no medication is administered. She is on a fixed interval reinforcement schedule (dosed hourly), so extinction occurs quickly when reinforcement doesn’t come at the expected time. Among the reinforcement schedules, variable ratio is the most productive and the most resistant to extinction. Fixed interval is the least productive and the easiest to extinguish (Figure \(1\)). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/18%3A_Operant_Conditioning/18.05%3A_Reinforcement_Schedules.txt |
Although strict behaviorists such as Skinner and Watson refused to believe that cognition (such as thoughts and expectations) plays a role in learning, another behaviorist, Edward C. Tolman, had a different opinion. Tolman’s experiments with rats demonstrated that organisms can learn even if they do not receive immediate reinforcement (Tolman & Honzik, 1930; Tolman et al., 1946). This finding was in conflict with the prevailing idea at the time that reinforcement must be immediate in order for learning to occur, thus suggesting a cognitive aspect to learning.
CONNECT THE CONCEPTS
Gambling and the Brain
Skinner (1953) stated, “If the gambling establishment cannot persuade a patron to turn over money with no return, it may achieve the same effect by returning part of the patron’s money on a variable-ratio schedule” (p. 397).
Skinner uses gambling as an example of the power and effectiveness of conditioning behavior based on a variable ratio reinforcement schedule. In fact, Skinner was so confident in his knowledge of gambling addiction that he even claimed he could turn a pigeon into a pathological gambler (“Skinner’s Utopia,” 1971). Beyond the power of variable ratio reinforcement, gambling seems to work on the brain in the same way as some addictive drugs. The Illinois Institute for Addiction Recovery (n.d.) reports evidence suggesting that pathological gambling is an addiction similar to a chemical addiction (Figure \(1\)). Specifically, gambling may activate the reward centers of the brain, much like cocaine does. Research has shown that some pathological gamblers have lower levels of the neurotransmitter (brain chemical) known as norepinephrine than do normal gamblers (Roy et al., 1988). According to a study conducted by Alec Roy and colleagues, norepinephrine is secreted when a person feels stress, arousal, or thrill; pathological gamblers use gambling to increase their levels of this neurotransmitter. Another researcher, neuroscientist Hans Breiter, has done extensive research on gambling and its effects on the brain. Breiter (as cited in Franzen, 2001) reports that “Monetary reward in a gambling-like experiment produces brain activation very similar to that observed in a cocaine addict receiving an infusion of cocaine” (para. 1). Deficiencies in serotonin (another neurotransmitter) might also contribute to compulsive behavior, including a gambling addiction.
It may be that pathological gamblers’ brains are different than those of other people, and perhaps this difference may somehow have led to their gambling addiction, as these studies seem to suggest. However, it is very difficult to ascertain the cause because it is impossible to conduct a true experiment (it would be unethical to try to turn randomly assigned participants into problem gamblers). There- fore, it may be that causation actually moves in the opposite direction—perhaps the act of gambling somehow changes neurotransmitter levels in some gamblers’ brains. It also is possible that some overlooked factor, or confounding variable, played a role in both the gambling addiction and the differences in brain chemistry. ■
In the experiments, Tolman placed hungry rats in a maze with no reward for finding their way through it. He also studied a comparison group that was rewarded with food at the end of the maze. As the unreinforced rats explored the maze, they developed a cognitive map: a mental picture of the lay- out of the maze (Figure \(2\)). After 10 sessions in the maze without reinforcement, food was placed in a goal box at the end of the maze. As soon as the rats became aware of the food, they were able to find their way through the maze quickly, just as quickly as the comparison group, which had been rewarded with food all along. This is known as latent learning: learning that occurs but is not observable in behavior until there is a reason to demonstrate it.
Latent learning also occurs in humans. Children may learn by watching the actions of their parents but only demonstrate it at a later date, when the learned material is needed. For example, suppose that Ravi’s dad drives him to school every day. In this way, Ravi learns the route from his house to his school, but he’s never driven there himself, so he has not had a chance to demonstrate that he’s learned the way. One morning Ravi’s dad has to leave early for a meeting, so he can’t drive Ravi to school. Instead, Ravi follows the same route on his bike that his dad would have taken in the car. This demonstrates latent learning. Ravi had learned the route to school, but had no need to demonstrate this knowledge earlier.
LINK TO LEARNING
Watch this video to learn more about Carlson’s studies on cognitive maps and navigation in buildings.
EVERYDAY CONNECTION
This Place is Like a Maze
Have you ever gotten lost in a building and couldn’t find your way back out? While that can be frustrating, you’re not alone. At one time or another we’ve all gotten lost in places like a museum, hospital, or university library. Whenever we go someplace new, we build a mental representation—or cognitive map—of the location, as Tolman’s rats built a cognitive map of their maze. However, some buildings are con- fusing because they include many areas that look alike or have short lines of sight. Because of this, it’s often difficult to predict what’s around a corner or decide whether to turn left or right to get out of a building. Psychologist laura Carlson (2010) suggests that what we place in our cognitive map can impact our success in navigating through the environment. She suggests that paying attention to specific features upon entering a building, such as a picture on the wall, a fountain, a statue, or an escalator, adds information to our cognitive map that can be used later to help find our way out of the building. ■
REFERENCES
Cangi, K., & Daly, M. (2013). The effects of token economies on the occurrence of appropriate and inappropriate behaviors by children with autism in a social skills setting. West Chester University: Journal of Undergraduate Research.
Carlson, L., Holscher, C., Shipley, T., & Conroy Dalton, R. (2010). Getting lost in buildings. Current Directions in Psychological Science, 19(5), 284–289. https://doi.org/10.1177/0963721410383243
Franzen, H. (2001, May 24). Gambling—like food and drugs—produces feelings of reward in the brain. Scientific American. www. scientificamerican.com/article.cfm?id=gamblinglike-food-and-dru
Fryer, R. G., Jr. (2010, April). Financial incentives and student achievement: Evidence from randomized trials. National Bureau of Economic Research [NBER] Working Paper, No. 15898. www. nber.org/papers/w15898
Gershoff, E. T. (2002). Corporal punishment by parents and associated child behaviors and experiences: A meta-analytic and theoretical review. Psychological Bulletin, 128(4), 539–579. https://doi.org/ 10.1037/0033-2909.128.4.539
Gershoff, E. T., Grogan-Kaylor, A., Lansford, J. E., Chang, L., Zelli, A., Deater-Deckard, K., & Dodge, K. A. (2010). Parent discipline practices in an international sample: Associations with child behaviors and moderation by perceived normativeness. Child Development, 81(2), 487–502. doi.org/10.1111/j.1467-8624.2009.01409.x
Illinois Institute for Addiction Recovery. (n.d.). WTVP on gambling. http:// www.addictionrecov.org/InTheNews/Gambling/
Roy, A., Adinoff, B., Roehrich, L., Lamparski, D., Custer, R., Lorenz, V., Barbaccia, M., Guidotti, A., Costa, E., & Linnoila, M. (1988). Pathological gambling: A psychobiological study. Archives of General Psychiatry, 45(4), 369–373. doi.org/10.1001/archpsyc.1988. 01800280085011
Skinner, B. F. (1938). The behavior of organisms: An experimental analysis. Appleton-Century-Crofts.
Skinner, B. F. (1953). Science and human behavior. Macmillan. Skinner, B. F. (1961). Cumulative record: A selection of papers.
Appleton-Century-Crofts.
Skinner’s utopia: Panacea, or path to hell? (1971, September 20). Time. http://www.wou.edu/~girodm/611/Skinner%27s_utopia.pdf
Thorndike, E. L. (1898). Animal intelligence: An experimental study of the associative processes in animals. The Psychological Review: Monograph Supplements, 2(4), i–109. doi.org/10.1037/ h0092987
Tolman, E. C., & Honzik, C. H. (1930). Degrees of hunger, reward, and non-reward, and maze learning in rats. University of California Publications in Psychology, 4, 241–256.
Tolman, E. C., Ritchie, B. F., & Kalish, D. (1946). Studies in spatial learning: II. Place learning versus response learning. Journal of Experimental Psychology, 36, 221–229. https://doi.org/10.1037/h0060262 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/18%3A_Operant_Conditioning/18.06%3A_Cognition_And_Latent_Learning.txt |
As you can see in TABLE \(1\), psychologists conceptualize memory in terms of types, stages, and processes. In this section we will consider the two types of memory, explicit memory and implicit memory, and then the three major memory stages: sensory, short-term, and long-term (Atkinson & Shiffrin, 1968). Then, in the next section, we will consider the nature of long-term memory, with a particular emphasis on the cognitive techniques we can use to improve our memories. Our discussion will focus on the three processes that are central to long-term memory: encoding, storage, and retrieval.
Table \(1\): Memory Conceptualized in Terms of Types, Stages, and Processes
As types
Explicit memory
Implicit memory
As stages
Sensory memory
Short-term memory
Long-term memory
As processes
Encoding
Storage
Retrieval
19.02: Explicit memory
When we assess memory by asking a person to consciously remember things, we are measuring explicit memory. Explicit memory refers to knowledge or experiences that can be consciously remembered. As you can see in Figure \(1\), there are two types of explicit memory: episodic and semantic. Episodic memory refers to the firsthand experiences that we have had (e.g., recollections of our high school graduation day or of the fantastic dinner we had in New York last year). Semantic memory refers to our knowledge of facts and concepts about the world (e.g., that the absolute value of −90 is greater than the absolute value of 9 and that one definition of the word “affect” is “the experience of feeling or emotion”).
Explicit memory is assessed using measures in which the individual being tested must consciously attempt to remember the information. A recall memory test is a measure of explicit memory that involves bringing from memory information that has previously been remembered. We rely on our recall memory when we take an essay test, because the test requires us to generate previously remembered information. A multiple-choice test is an example of a recognition memory test, a measure of explicit memory that involves determining whether information has been seen or learned before.
Your own experiences taking tests will probably lead you to agree with the scientific research finding that recall is more difficult than recognition. Recall, such as required on essay tests, involves two steps: first generating an answer and then determining whether it seems to be the correct one. Recognition, as on multiple-choice tests, only involves determining which item from a list seems most correct (Haist et al., 1992). Although they involve different processes, recall and recognition memory measures tend to be correlated. Students who do better on a multiple-choice exam will also, by and large, do better on an essay exam (Bridgeman & Morgan, 1996).
A third way of measuring memory is known as relearning (Nelson, 1985). Measures of relearning (or savings) assess how much more quickly information is processed or learned when it is studied again after it has already been learned but then forgotten. If you have taken some French courses in the past, for instance, you might have forgotten most of the vocabulary you learned. But if you were to work on your French again, you’d learn the vocabulary much faster the second time around. Relearning can be a more sensitive measure of memory than either recall or recognition because it allows assessing memory in terms of “how much” or “how fast” rather than simply “correct” versus “incorrect” responses. Relearning also allows us to measure memory for procedures like driving a car or playing a piano piece, as well as memory for facts and figures. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/19%3A_Memories_as_Types_and_Stages/19.01%3A_Introduction.txt |
While explicit memory consists of the things that we can consciously report that we know, implicit memory refers to knowledge that we cannot consciously access. However, implicit memory is nevertheless exceedingly important to us because it has a direct effect on our behavior. Implicit memory refers to the influence of experience on behavior, even if the individual is not aware of those influences. As you can see in Figure \(1\), there are three general types of implicit memory: procedural memory, classical conditioning effects, and priming.
Procedural memory refers to our often unexplainable knowledge of how to do things. When we walk from one place to another, speak to another person in English, dial a cell phone, or play a video game, we are using procedural memory. Procedural memory allows us to perform complex tasks, even though we may not be able to explain to others how we do them. There is no way to tell someone how to ride a bicycle; a person has to learn by doing it. The idea of implicit memory helps explain how infants are able to learn. The ability to crawl, walk, and talk are procedures, and these skills are easily and efficiently developed while we are children despite the fact that as adults we have no conscious memory of having learned them.
A second type of implicit memory is classical conditioning effects, in which we learn, often without effort or aware- ness, to associate neutral stimuli (such as a sound or a light) with another stimulus (such as food), which creates a naturally occurring response, such as enjoyment or salivation. The memory for the association is demonstrated when the conditioned stimulus (the sound) begins to create the same response as the unconditioned stimulus (the food) did before the learning.
The final type of implicit memory is known as priming, or changes in behavior as a result of experiences that have happened frequently or recently. Priming refers both to the activation of knowledge (e.g., we can prime the concept of “kindness” by presenting people with words related to kind- ness) and to the influence of that activation on behavior (people who are primed with the concept of kindness may act more kindly).
One measure of the influence of priming on implicit memory is the word fragment test, in which a person is asked to fill in missing letters to make words. You can try this your- self: First, try to complete the following word fragments, but work on each one for only three or four seconds. Do any words pop into mind quickly?
_ ib _ a _ y
_ h _ s _ _ i_ n _ o_ k
_ h _ is _
Now read the following sentence carefully:
“He got his materials from the shelves, checked them out, and then left the building.”
Then try again to make words out of the word fragments.
I think you might find that it is easier to complete fragments 1 and 3 as “library” and “book,” respectively, after you read the sentence than it was before you read it. However, reading the sentence didn’t really help you to complete fragments 2 and 4 as “physician” and “chaise.” This difference in implicit memory probably occurred because as you read the sentence, the concept of “library” (and perhaps “book”) was primed, even though they were never mentioned explicitly. Once a concept is primed it influences our behaviors, for instance, on word fragment tests.
RESEARCH FOCUS
Priming Outside Awareness influences Behavior
One of the most important characteristics of implicit memories is that they are frequently formed and used automatically, without much effort or awareness on our part. In one demonstration of the automaticity and influence of priming effects, John Bargh and his colleagues (1996) conducted a study in which they showed college students lists of five scrambled words, each of which they were to make into a sentence. Furthermore, for half of the research participants, the words were related to stereotypes of the elderly. These participants saw words such as the following:
in Florida retired live people bingo man the forgetful plays
The other half of the research participants also made sentences, but from words that had nothing to do with elderly stereotypes. The purpose of this task was to prime stereo- types of elderly people in memory for some of the participants but not for others.
The experimenters then assessed whether the priming of elderly stereotypes would have any effect on the students’ behavior—and indeed it did. When the research participant had gathered all of his or her belongings, thinking that the experiment was over, the experimenter thanked him or her for participating and gave directions to the closest elevator. Then, without the participants knowing it, the experimenters recorded the amount of time that the participant spent walking from the doorway of the experimental room toward the elevator. As you can see in Figure \(2\), participants who had made sentences using words related to elderly stereo- types took on the behaviors of the elderly—they walked significantly more slowly as they left the experimental room.
To determine if these priming effects occurred out of the awareness of the participants, Bargh and his colleagues asked yet another group of students to complete the priming task and then to indicate whether they thought the words they had used to make the sentences had any relationship to each other, or could possibly have influenced their behavior in any way. These students had no awareness of the possibility that the words might have been related to the elderly or could have influenced their behavior. ■
Our everyday behaviors are influenced by priming in a wide variety of situations. Seeing an advertisement for cigarettes may make us start smoking, seeing the flag of our home country may arouse our patriotism, and seeing a student from a rival school may arouse our competitive spirit. And these influences on our behaviors may occur without our being aware of them. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/19%3A_Memories_as_Types_and_Stages/19.03%3A_Implicit_memory.txt |
Another way of understanding memory is to think about it in terms of stages that describe the length of time that information remains available to us. According to this approach (see Figure \(1\)), information begins in sensory memory, moves to short-term memory, and eventually moves to long- term memory. But not all information makes it through all three stages; most of it is forgotten. Whether the information moves from shorter-duration memory into longer-duration memory or whether it is lost from memory entirely depends on how the information is attended to and processed.
Sensory Memory
Sensory memory refers to the brief storage of sensory information. Sensory memory is a memory buffer that lasts only very briefly and then, unless it is attended to and passed on for more processing, is forgotten. The purpose of sensory memory is to give the brain some time to process the incoming sensations and to allow us to see the world as an unbroken stream of events rather than as individual pieces.
Visual sensory memory is known as iconic memory. Iconic memory was first studied by the psychologist George Sperling (1960). In his research, Sperling showed participants a display of letters in rows, similar to that shown in Figure \(2\). However, the display lasted only about 50 milliseconds (one-twentieth of a second). Then, Sperling gave his participants a recall test in which they were asked to name all the letters they could remember. On average, the participants could remember only about one-quarter of the letters they had seen.
Sperling reasoned that the participants had seen all the letters but could remember them only very briefly, making it impossible for them to report them all. To test this idea, in his next experiment he first showed the same letters, but then after the display had been removed, he signaled to the participants to report the letters from either the first, second, or third row. In this condition, the participants now reported almost all the letters in that row. This finding confirmed Sperling’s hunch: Participants had access to all of the letters in their iconic memories, and if the task was short enough, they were able to report on the part of the display he asked them to. The “short enough” is the length of iconic memory, which turns out to be about 250 milliseconds (1⁄4 of a second).
Auditory sensory memory is known as echoic memory. In contrast to iconic memories, which decay very rapidly, echoic memories can last as long as 4 seconds (Cowan et al., 1990). This is convenient as it allows you—among other things—to remember the words that you said at the beginning of a long sentence when you get to the end of it, and to take notes on your psychology professor’s most recent statement even after he or she has finished saying it.
In some people iconic memory seems to last longer, a phenomenon known as eidetic imagery (or “photographic memory”) in which people can report details of an image over long periods of time. These people, who often suffer from psychological disorders such as autism, claim that they can “see” an image long after it has been presented, and can often report accurately on that image. There is also some evidence for eidetic memories in hearing; some people report that their echoic memories persist for unusually long periods of time. The composer Wolfgang Amadeus Mozart may have possessed eidetic memory for music, because even when he was very young and had not yet had a great deal of musical training, he could listen to long compositions and then play them back almost perfectly (Solomon, 1995).
Short-Term Memory
Most of the information that gets into sensory memory is for- gotten, but information that we turn our attention to, with the goal of remembering it, may pass into short-term memory. Short-term memory (STM) is the place where small amounts of information can be temporarily kept for more than a few seconds but usually for less than one minute (Baddeley et al., 1990). Information in short-term memory is not stored permanently but rather becomes available for us to process, and the processes that we use to make sense of, modify, interpret, and store information in STM are known as working memory.
Although it is called “memory,” working memory is not a store of memory like STM but rather a set of memory procedures or operations. Imagine, for instance, that you are asked to participate in a task such as this one, which is a mea- sure of working memory (Unsworth & Engle, 2007). Each of the following questions appears individually on a computer screen and then disappears after you answer the question:
Is 10 × 2 − 5 = 15?
(Answer YES or NO)
Then remember “S”
Is 12 ÷ 6 − 2 = 1?
(Answer YES or NO) Then remember “R”
Is 10 × 2 = 5?
(Answer YES or NO) Then remember “P”
Is 8 ÷ 2 − 1 = 1?
(Answer YES or NO) Then remember “I”
Is 6 × 2 − 1 = 8?
(Answer YES or NO) Then remember “U”
Is 2 × 3 − 3 = 0?
(Answer YES or NO) Then remember “Q”
To successfully accomplish the task, you have to answer each of the math problems correctly and at the same time remember the letter that follows the task. Then, after the six questions, you must list the letters that appeared in each of the trials in the correct order (in this case S, R, P, T, U, Q).
To accomplish this difficult task you need to use a variety of skills. You clearly need to use STM, as you must keep the letters in storage until you are asked to list them. But you also need a way to make the best use of your available attention and processing. For instance, you might decide to use a strategy of “repeat the letters twice, then quickly solve the next problem, and then repeat the letters twice again including the new one.” Keeping this strategy (or others like it) going is the role of working memory’s central executive—the part of working memory that directs attention and processing. The central executive will make use of whatever strategies seem to be best for the given task. For instance, the central executive will direct the rehearsal process, and at the same time direct the visual cortex to form an image of the list of letters in memory. You can see that although STM is involved, the processes that we use to operate on the material in memory are also critical.
Short-term memory is limited in both the length and the amount of information it can hold. Peterson and Peterson (1959) found that when people were asked to remember a list of three-letter strings and then were immediately asked to perform a distracting task (counting backward by threes), the material was quickly forgotten (see Figure \(3\)), such that by 18 seconds it was virtually gone.
One way to prevent the decay of information from short- term memory is to use working memory to rehearse it. Maintenance rehearsal is the process of repeating information mentally or out loud with the goal of keeping it in memory. We engage in maintenance rehearsal to keep something that we want to remember (e.g., a person’s name, e-mail address, or phone number) in mind long enough to write it down, use it, or potentially transfer it to long-term memory.
If we continue to rehearse information it will stay in STM until we stop rehearsing it, but there is also a capacity limit to STM. Try reading each of the following rows of numbers, one row at a time, at a rate of about one number each second. Then when you have finished each row, close your eyes and write down as many of the numbers as you can remember.
``` 019
3586
10295
861059
1029384
75674834
657874104
6550423897
```
If you are like the average person, you will have found that on this test of working memory, known as a digit span test, you did pretty well up to about the fourth line, and then you started having trouble. I bet you missed some of the numbers in the last three rows, and did pretty poorly on the last one.
The digit span of most adults is between five and nine dig- its, with an average of about seven. The cognitive psychologist George Miller (1956) referred to “seven plus or minus two” pieces of information as the “magic number” in short-term memory. But if we can only hold a maximum of about nine digits in short-term memory, then how can we remember larger amounts of information than this? For instance, how can we ever remember a 10-digit phone number long enough to dial it?
One way we are able to expand our ability to remember things in STM is by using a memory technique called chunking. Chunking is the process of organizing information into smaller groupings (chunks), thereby increasing the number of items that can be held in STM. For instance, try to remember this string of 12 letters:
XOFCBANNCVTM
You probably won’t do that well because the number of letters is more than the magic number of seven. Now try again with this one:
MTVCNNABCFOX
Would it help you if I pointed out that the material in this string could be chunked into four sets of three letters each? I think it would, because then rather than remembering 12 letters, you would only have to remember the names of four television stations. In this case, chunking changes the number of items you have to remember from 12 to only four.
Experts rely on chunking to help them process complex information. Herbert Simon and William Chase (1973) showed chess masters and chess novices various positions of pieces on a chessboard for a few seconds each. The experts did a lot better than the novices in remembering the positions because they were able to see the “big picture.” They didn’t have to remember the position of each of the pieces individually, but chunked the pieces into several larger layouts. But when the researchers showed both groups random chess positions—positions that would be very unlikely to occur in real games—both groups did equally poorly, because in this situation the experts lost their ability to organize the layouts (see Figure \(4\)). The same occurs for basketball. Basketball players recall actual basketball positions much better than do nonplayers, but only when the positions make sense in terms of what is happening on the court, or what is likely to happen in the near future, and thus can be chunked into bigger units (Didierjean & Marmèche, 2005).
Long-Term Memory
If information makes it past short term-memory it may enter long-term memory (LTM), memory storage that can hold information for days, months, and years. The capacity of long-term memory is large, and there is no known limit to what we can remember (Wang et al., 2003). Although we may forget at least some information after we learn it, other things will stay with us forever. In the next section we will discuss the principles of long-term memory.
REFERENCES
Atkinson, R. C., & Shiffrin, R. M. (1968). Human memory: A proposed system and its control processes. In K. Spence (Ed.), The psychology of learning and motivation (Vol. 2). Academic Press.
Baddeley, A. D., Vallar, G., & Shallice, T. (1990). The development of the concept of working memory: Implications and contributions of neuropsychology. In G. Vallar & T. Shallice (Eds.), Neuropsychological impairments of short-term memory (pp. 54–73). Cambridge University Press.
Bargh, J. A., Chen, M., & Burrows, L. (1996). Automaticity of social behavior: Direct effects of trait construct and stereotype activation on action. Journal of Personality and Social Psychology, 71(2), 230–244. https://doi.org/10.1037/0022-3514.71.2.230
Bridgeman, B., & Morgan, R. (1996). Success in college for students with discrepancies between performance on multiple-choice and essay tests. Journal of Educational Psychology, 88(2), 333–340. doi. org/10.1037/0022-0663.88.2.333
Cowan, N., Lichty, W., & Grove, T. R. (1990). Properties of memory for unattended spoken syllables. Journal of Experimental Psychology: Learning, Memory, and Cognition, 16(2), 258–268. https://doi.org/ 10.1037/0278-7393.16.2.258
Didierjean, A., & Marmèche, E. (2005). Anticipatory representation of visual basketball scenes by novice and expert players. Visual Cognition, 12(2), 265–283. doi.org/10.1080/ 13506280444000021A
Haist, F., Shimamura, A. P., & Squire, L. R. (1992). On the relationship between recall and recognition memory. Journal of Experimental Psychology: Learning, Memory, and Cognition, 18(4), 691–702. https://doi.org/10.1037/0278-7393.18.4.691
Miller, G. A. (1956). The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychological Review, 63(2), 81–97. https://doi.org/10.1037/h0043158
Nelson, T. O. (1985). Ebbinghaus’s contribution to the measurement of retention: Savings during relearning. Journal of Experimental Psychology: Learning, Memory, and Cognition, 11(3), 472–478. https://doi.org/10.1037/0278-7393.11.3.472
Peterson, L., & Peterson, M. J. (1959). Short-term retention of individual verbal items. Journal of Experimental Psychology, 58(3), 193–198. https://doi.org/10.1037/h0049234
Simon, H. A., & Chase, W. G. (1973). Skill in chess. American Scientist, 61(4), 394–403.
Solomon, M. (1995). Mozart: A life. Harper Perennial.
Sperling, G. (1960). The information available in brief visual presentation.
Psychological Monographs: General and Applied, 74(11), 1–29.
https://doi.org/10.1037/h0093759
Unsworth, N., & Engle, R. W. (2007). On the division of short-term and working memory: An examination of simple and complex span and their relation to higher order abilities. Psychological Bulletin, 133(6), 1038–1066. https://doi.org/10.1037/0033-2909.133.6.1038
Wang, Y., Liu, D., & Wang, Y. (2003). Discovering the capacity of human memory. Brain and Mind, 4(2), 189–198.
19.05: Key Takeaways
• Memory refers to the ability to store and retrieve information over time.
• For some things our memory is very good, but our active cognitive processing of information assures that memory is never an exact replica of what we have experienced.
• Explicit memory refers to experiences that can be intentionally and consciously remembered, and it is measured using recall, recognition, and relearning. Explicit memory includes episodic and semantic memories.
• Measures of relearning (also known as savings) assess how much more quickly information is learned when it is studied again after it has already been learned but then forgotten.
• Implicit memory refers to the influence of experience on behavior, even if the individual is not aware of those influences. The three types of implicit memory are procedural memory, classical conditioning, and priming.
• Information processing begins in sensory memory, moves to short- term memory, and eventually moves to long-term memory.
• Maintenance rehearsal and chunking are used to keep information in short-term memory.
• The capacity of long-term memory is large, and there is no known limit to what we can remember.
EXERCISES AND CRITICAL THINKING
1. List some situations in which sensory memory is useful for you. What do you think your experience of the stimuli would be like if you had no sensory memory?
2. Describe a situation in which you need to use working memory to perform a task or solve a problem. How do your working memory skills help you? | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/19%3A_Memories_as_Types_and_Stages/19.04%3A_Stages_Of_Memory-_Sensory_Short-Term_And_Long-Term_Memory.txt |
Although it is useful to hold information in sensory and short-term memory, we also rely on our long-term memory (LTM). We want to remember the name of the new boy in the class, the name of the movie we saw last week, and the material for our upcoming psychology test. Psychological research has produced a great deal of knowledge about long- term memory, and this research can be useful as you try to learn and remember new material (see Table \(1\)). In this section we will consider this question in terms of the types of processing that we do on the information we want to remember. To be successful, the information that we want to remember must be encoded and stored, and then retrieved.
Table \(1\): Helpful Memory Techniques Based on Psychological Research
Technique
Description
Useful example
Use elaborative encoding.
Material is better remembered if it is processed more fully.
Think, for instance, “Proactive interference is like retroactive interference but it occurs in a forward manner.”
Make use of the self-reference effect.
Material is better remembered if it is linked to thoughts about the self.
Think, for instance, “I remember a time when
I knew the answer to an exam question but couldn’t quite get it to come to mind. This was an example of the tip-of-the-tongue phenomenon.”
Be aware of the forgetting curve.
Information that we have learned drops off rapidly with time.
Review the material that you have already studied right before the exam to increase the likelihood it will remain in memory.
Make use of the spacing effect.
Information is learned better when it is studied in shorter periods spaced over time.
Study a little bit every day; do not cram at the last minute.
Rely on overlearning.
We can continue to learn even after we think we know the information perfectly.
Keep studying, even if you think you already have it down.
Use context-dependent retrieval.
We have better retrieval when it occurs in the same situation in which we learned the material.
If possible, study under conditions similar to the conditions in which you will take the exam.
Use state-dependent retrieval.
We have better retrieval when we are in the same psychological state as we were when we learned the material.
Many possibilities, but don’t study under the influence of drugs or alcohol, unless you plan to use them on the day of the exam (which is not recommended).
20.02: Encoding And Storage- How Our Perceptions Become Memories
Encoding is the process by which we place the things that we experience into memory. Unless information is encoded, it cannot be remembered. I’m sure you’ve been to a party where you’ve been introduced to someone and then—maybe only seconds later—you realize that you do not remember the per- son’s name. Of course it’s not really surprising that you can’t remember the name, because you probably were distracted and you never encoded the name to begin with.
Not everything we experience can or should be encoded. We tend to encode things that we need to remember and not bother to encode things that are irrelevant. Look at Figure \(1\), which shows different images of U.S. pennies. Can you tell which one is the real one? Nickerson and Adams (1979) found that very few of the U.S. participants they tested could identify the right one. We see pennies a lot, but we don’t bother to encode their features.
One way to improve our memory is to use better encoding strategies. Some ways of studying are more effective than others. Research has found that we are better able to remember information if we encode it in a meaningful way. When we engage in elaborative encoding we process new information in ways that make it more relevant or meaningful (Craik & Lockhart, 1972; Harris & Qualls, 2000).
Imagine that you are trying to remember the characteristics of the different schools of psychology. Rather than simply trying to remember the schools and their characteristics, you might try to relate the information to things you already know. For instance, you might try to remember the fundamentals of the cognitive school of psychology by linking the characteristics to the computer model. The cognitive school focuses on how information is input, processed, and retrieved, and you might think about how computers do pretty much the same thing. You might also try to organize the information into meaningful units. For instance, you might link the cognitive school to structuralism because both were concerned with mental processes. You also might try to use visual cues to help you remember the information. You might look at the image of Freud and imagine what he looked like as a child. That image might help you remember that childhood experiences were an important part of Freudian theory. Each person has his or her unique way of elaborating on information; the important thing is to try to develop unique and meaningful associations among the materials.
RESEARCH FOCUS
Elaboration and Memory
In an important study showing the effectiveness of elaborative encoding, Rogers et al. (1977) studied how people recalled information that they had learned under different processing conditions. All the participants were presented with the same list of 40 adjectives to learn, but through the use of random assignment, the participants were given one of four different sets of instructions about how to process the adjectives.
Participants assigned to the structural task condition were asked to judge whether the word was printed in upper- case or lowercase letters. Participants in the phonemic task condition were asked whether or not the word rhymed with another given word. In the semantic task condition, the participants were asked if the word was a synonym of another word. And in the self-reference task condition, participants were asked to indicate whether or not the given adjective was or was not true of themselves. After completing the specified task, each participant was asked to recall as many adjectives as he or she could remember.
Rogers and his colleagues hypothesized that different types of processing would have different effects on memory. As you can see in Figure \(2\), the students in the self-reference task condition recalled significantly more adjectives than did students in any other condition. This finding, known. as the self-reference effect, is powerful evidence that the self-concept helps us organize and remember information. The next time you are studying for an exam, you might try relating the material to your own experiences. The self-reference effect suggests that doing so will help you better remember the information (Symons & Johnson, 1997). ■
Using the Contributions of Hermann Ebbinghaus to improve Your Memory
Hermann Ebbinghaus (1850–1909) was a pioneer of the study of memory. In this section we consider three of his most important findings, each of which can help you improve your memory. In his research, in which he was the only research participant, Ebbinghaus practiced memorizing lists of non- sense syllables, such as the following:
DIF, LAJ, LEQ, MUV, WYC, DAL, SEN, KEP, NUD
You can imagine that because the material he was trying to learn was not at all meaningful, it was not easy to do. Ebbinghaus plotted how many of the syllables he could remember against the time that had elapsed since he had studied them. He discovered an important principle of memory: Memory decays rapidly at first, but the amount of decay levels off with time (Figure \(3\)). Although Ebbinghaus looked at forgetting after days had elapsed, the same effect occurs on longer and shorter time scales. Bahrick (1984) found that students who took a Spanish language course forgot about half of the vocabulary they had learned within three years, but that after that time their memory remained pretty much constant. Forget- ting also drops off quickly on a shorter time frame. This suggests that you should try to review the material that you have already studied right before you take an exam; that way, you will be more likely to remember the material during the exam.
Ebbinghaus also discovered another important principle of learning, known as the spacing effect. The spacing effect refers to the fact that learning is improved when the same amount of study is spread out over periods of time than it is when it occurs closer together or at the same time. This means that even if you have only a limited amount of time to study, you’ll learn more if you study continually throughout the semester (a little bit every day is best) than if you wait to cram at the last minute before your exam (Figure \(4\)). Another good strategy is to study and then wait as long as you can before you forget the material. Then review the information and again wait as long as you can before you forget it. (This probably will be a longer period of time than the first time.) Repeat and repeat again. The spacing effect is usually considered in terms of the difference between distributed practice (practice that is spread out over time) and massed practice (practice that comes in one block), with the former approach producing better memory.
Ebbinghaus also considered the role of overlearning— that is, continuing to practice and study even when we think that we have mastered the material. Ebbinghaus and other researchers have found that overlearning helps encoding (Driskell et al., 1992). Students frequently think that they have already mastered the material but then discover when they get to the exam that they have not. The point is clear: Try to keep studying and reviewing, even if you think you already know all the material.
RETRIEVAL DEMONSTRATION
Try this test of the ability to retrieve information with a classmate. The instructions are in the text.
We’ve all experienced retrieval failure in the form of the frustrating tip-of-the-tongue phenomenon, in which we are certain that we know something we are trying to recall but cannot quite come up with it. You can try this one on your friends as well. Read your friend the names of the 10 states listed in the States and Capital Cities Demonstration, and ask him to name the capital city of each state. Now, for the capital cities that your friend can’t name, give him just the first letter of the capital city. You’ll probably find that having the first letters of the cities helps with retrieval. The tip-of-the-tongue experience is a very good example of the inability to retrieve information that is actually stored in memory.
STATES AND CAPITAL CITIES DEMONSTRATION
Try this demonstration of the tip-of-the-tongue phenomenon with a classmate. Instructions are in the text.
We are more likely to be able to retrieve items from memory when conditions at retrieval are similar to the conditions under which we encoded them. Context-dependent learning refers to an increase in retrieval when the external situation in which information is learned matches the situation in which it is remembered. Godden and Baddeley (1975) conducted a study to test this idea using scuba divers. They asked the divers to learn a list of words either when they were on land or when they were underwater. Then they tested the divers on their memory, either in the same or the opposite situation. As you can see in Figure \(5\), the divers’ memory was better when they were tested in the same context in which they had learned the words than when they were tested in the other context.
You can see that context-dependent learning might also be important in improving your memory. For instance, you might want to try to study for an exam in a situation that is similar to the one in which you are going to take the exam.
Whereas context-dependent learning refers to a match in the external situation between learning and remembering, state-dependent learning refers to superior retrieval of memories when the individual is in the same physiological or psychological state as during encoding. Research has found, for instance, that animals that learn a maze while under the influence of one drug tend to remember their learning better when they are tested under the influence of the same drug than when they are tested without the drug (Jackson et al., 1992). And research with humans finds that bilinguals remember better when tested in the same language in which they learned the material (Marian & Kaushanskaya, 2007). Mood states may also produce state-dependent learning. People who learn information when they are in a bad (rather than a good) mood find it easier to recall these memories when they are tested while they are in a bad mood, and vice versa. It is easier to recall unpleasant memories than pleasant ones when we’re sad, and easier to recall pleasant memories than unpleasant ones when we’re happy (Bower, 1981; Eich, 2008).
Variations in the ability to retrieve information are also seen in the serial position curve. When we give people a list of words one at a time (e.g., on flashcards) and then ask them to recall them, the results look something like those in Figure \(6\). People are able to retrieve more words that were presented to them at the beginning and the end of the list than they are words that were presented in the middle of the list. This pattern, known as the serial position curve, is caused by two retrieval phenomena: The primacy effect refers to a tendency to better remember stimuli that are presented early in a list; the recency effect refers to the tendency to better remember stimuli that are presented later in a list.
There are a number of explanations for primacy and recency effects, but one of them is in terms of the effects of rehearsal on short-term and long-term memory (Baddeley et al., 2009). Because we can keep the last words we learned in the presented list in short-term memory by rehearsing them before the memory test begins, they are relatively easily remembered. So the recency effect can be explained in terms of maintenance rehearsal in short-term memory. And the pri- macy effect may also be due to rehearsal—when we hear the first word in the list we start to rehearse it, making it more likely that it will be moved from short-term to long-term memory. And the same is true for the other words that come early in the list. But for the words in the middle of the list, this rehearsal becomes much harder, making them less likely to be moved to LTM.
In some cases our existing memories influence our new learning. This may occur either in a backward way or a for- ward way. Retroactive interference occurs when learning something new impairs our ability to retrieve information that was learned earlier. For example, if you have learned to program in one computer language, and then you learn to program in another similar one, you may start to make mistakes programming the first language that you never would have made before you learned the new one. In this case the new memories work backward (retroactively) to influence retrieval from memory that is already in place.
In contrast to retroactive interference, proactive interference works in a forward direction (see Figure \(7\) for a comparison). Proactive interference occurs when earlier learning impairs our ability to encode information that we try to learn later. For example, if we have learned French as a second language, this knowledge may make it more difficult, at least in some respects, to learn a third language (say Spanish), which involves similar but not identical vocabulary. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/20%3A_How_We_Remember-_Cues_to_Improving_Memory/20.01%3A_Introduction.txt |
Memories that are stored in LTM are not isolated but rather are linked together into categories—networks of associated memories that have features in common with each other. Forming categories, and using categories to guide behavior, is a fundamental part of human nature. Associated concepts within a category are connected through spreading activation, which occurs when activating one element of a category activates other associated elements. For instance, because tools are associated in a category, reminding people of the word “screwdriver” will help them remember the word “wrench.” And, when people have learned lists of words that come from different categories (e.g., as in the Retrieval Demonstration), they do not recall the information haphazardly. If they have just remembered the word “wrench,” they are more likely to remember the word “screwdriver” next than they are to remember the word “dahlia,” because the words are organized in memory by category and because “dahlia” is activated by spreading activation from “wrench” (Srull & Wyer, 1989).
Some categories have defining features that must be true of all members of the category. For instance, all members of the category “triangles” have three sides, and all members of the category “birds” lay eggs. But most categories are not so well-defined; the members of the category share some common features, but it is impossible to define which are or are not members of the category. For instance, there is no clear definition of the category “tool.” Some examples of the category, such as a hammer and a wrench, are clearly and easily identified as category members, whereas other members are not so obvious. Is an ironing board a tool? What about a car?
Members of categories (even those with defining features) can be compared to the category prototype, which is the member of the category that is most average or typical of the category (see Figure \(1\)). Some category members are more prototypical of, or similar to, the category than others. For instance, some category members (robins and sparrows) are highly prototypical of the category “birds,” whereas other category members (penguins and ostriches) are less prototypical. We retrieve information that is prototypical of a category faster than we retrieve information that is less prototypical (Rosch, 1975).
Mental categories are sometimes referred to as schemas— patterns of knowledge in long-term memory that help us organize information (see Figure \(2\)). We have schemas about objects (that a triangle has three sides and may take on different angles), about people (that Sam is friendly, likes to golf, and always wears sandals), about events (the particular steps involved in ordering a meal at a restaurant), and about social groups (we call these group schemas stereotypes).
Schemas are important in part because they help us remember new information by providing an organizational structure for it. Read the following paragraph (Bransford & Johnson, 1972) and then try to write down everything you can remember.
The procedure is actually quite simple. First you arrange things into different groups. Of course, one pile may be sufficient depending on how much there is to do. If you have to go some- where else due to lack of facilities, that is the next step; otherwise you are pretty well set. It is important not to overdo things. That is, it is better to do too few things at once than too many. In the short run this may not seem important, but complications can easily arise. A mistake can be expensive as well. At first the whole procedure will seem complicated. Soon, however, it will become just another facet of life. It is difficult to foresee any end to the necessity for this task in the immediate future, but then one never can tell. After the procedure is completed, one arranges the materials into different groups again. Then they can be put into their appropriate places. Eventually they will be used once more and the whole cycle will then have to be repeated. However, that is part of life.
It turns out that people’s memory for this information is quite poor, unless they have been told ahead of time that the information describes “doing the laundry,” in which case their memory for the material is much better. This demonstration of the role of schemas in memory shows how our existing knowledge can help us organize new information, and how this organization can improve encoding, storage, and retrieval. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/20%3A_How_We_Remember-_Cues_to_Improving_Memory/20.03%3A_The_Structure_Of_LTM-_Categories_Prototypes_And_Schemas.txt |
Just as information is stored on digital media such as DVDs and flash drives, the information in LTM must be stored in the brain. The ability to maintain information in LTM involves a gradual strengthening of the connections among the neurons in the brain. When pathways in these neural networks are frequently and repeatedly fired, the synapses become more efficient in communicating with each other, and these changes create memory. This process, known as long-term potentiation (LTP), refers to the strengthening of the synaptic connections between neurons as result of frequent stimulation (Lynch, 2002). Drugs that block LTP reduce learning, whereas drugs that enhance LTP increase learning (Lynch et al., 1991). Because the new patterns of activation in the synapses take time to develop, LTP happens gradually. The period of time in which LTP occurs and in which memories are stored is known as the period of consolidation.
Memory is not confined to the cortex; it occurs through sophisticated interactions between new and old brain structures (Figure \(1\)). One of the most important brain regions in explicit memory is the hippocampus, which serves as a preprocessor and elaborator of information (Squire, 1992). The hippocampus helps us encode information about spatial relationships, the context in which events were experienced, and the associations among memories (Eichenbaum, 1999). The hippocampus also serves in part as a switching point that holds the memory for a short time and then directs the information to other parts of the brain, such as the cortex, to actually do the rehearsing, elaboration, and long-term storage (Jonides et al., 2005). Without the hippocampus, which might be described as the brain’s “librarian,” our explicit memories would be inefficient and disorganized.
While the hippocampus is handling explicit memory, the cerebellum and the amygdala are concentrating on implicit and emotional memories, respectively. Research shows that the cerebellum is more active when we are learning associations and in priming tasks, and animals and humans with damage to the cerebellum have more difficulty in classical conditioning studies (Krupa et al., 1993; Woodruff-Pak et al., 2000). The storage of many of our most important emotional memories, and particularly those related to fear, is initiated and controlled by the amygdala (Sigurdsson et al., 2007).
Evidence for the role of different brain structures in different types of memories comes in part from case studies of patients who suffer from amnesia, a memory disorder that involves the inability to remember information. As with memory interference effects, amnesia can work in either a forward or a backward direction, affecting retrieval or encoding. For people who suffer damage to the brain, for instance, as a result of a stroke or other trauma, the amnesia may work backward. The outcome is retrograde amnesia, a memory disorder that produces an inability to retrieve events that occurred before a given time. Demonstrating the fact that LTP takes time (the process of consolidation), retrograde amnesia is usually more severe for memories that occurred just prior to the trauma than it is for older memories, and events that occurred just before the event that caused memory loss may never be recovered because they were never completely encoded.
Organisms with damage to the hippocampus develop a type of amnesia that works in a forward direction to affect encoding, known as anterograde amnesia. Anterograde amnesia is the inability to transfer information from short-term into long-term memory, making it impossible to form new memories. One well-known case study was a man named Henry Gustav Molaison (before he died in 2008, he was referred to only as H. M.) who had parts of his hippocampus removed to reduce severe seizures (Corkin et al., 1997). Following the operation, Molaison developed virtually complete anterograde amnesia. Although he could remember most of what had happened before the operation, and particularly what had occurred early in his life, he could no longer create new memories. Molaison was said to have read the same magazines over and over again without any awareness of having seen them before.
Cases of anterograde amnesia also provide information about the brain structures involved in different types of memory (Bayley & Squire, 2005; Helmuth, 1999; Paller, 2004). Although Molaison’s explicit memory was compromised because his hippocampus was damaged, his implicit memory was not (because his cerebellum was intact). He could learn to trace shapes in a mirror, a task that requires procedural memory, but he never had any explicit recollection of having performed this task or of the people who administered the test to him.
Although some brain structures are particularly important in memory, this does not mean that all memories are stored in one place. The American psychologist Karl Lashley (1929) attempted to determine where memories were stored in the brain by teaching rats how to run mazes, and then lesioning different brain structures to see if they were still able to complete the maze. This idea seemed straightforward, and Lashley expected to find that memory was stored in certain parts of the brain. But he discovered that no matter where he removed brain tissue, the rats retained at least some memory of the maze, leading him to conclude that memory isn’t located in a single place in the brain, but rather is distributed around it.
Long-term potentiation occurs as a result of changes in the synapses, which suggests that chemicals, particularly neurotransmitters and hormones, must be involved in memory. There is quite a bit of evidence that this is true. Glutamate, a neurotransmitter and a form of the amino acid glutamic acid, is perhaps the most important neurotransmitter in memory (McEntee & Crook, 1993). When animals, including people, are under stress, more glutamate is secreted, and this glutamate can help them remember (McGaugh, 2003). The neurotransmitter serotonin is also secreted when animals learn, and epinephrine may also increase memory, particularly for stressful events (Maki & Resnick, 2000; Sherwin, 1998). Estrogen, a female sex hormone, also seems critical, because women who are experiencing menopause, along with a reduction in estrogen, frequently report memory difficulties (Chester, 2001).
Our knowledge of the role of biology in memory suggests that it might be possible to use drugs to improve our memories, and Americans spend several hundred million dollars per year on memory supplements with the hope of doing just that. Yet controlled studies comparing memory enhancers, including Ritalin, methylphenidate, ginkgo biloba, and amphetamines, with placebo drugs find very little evidence for their effectiveness (Gold et al., 2002; McDaniel et al., 2002). Memory supplements are usually no more effective than drinking a sugared soft drink, which also releases glucose and thus improves memory slightly. This is not to say that we cannot someday create drugs that will significantly improve our memory. It is likely that this will occur in the future, but the implications of these advances are as yet unknown (Farah et al., 2004; Turner & Sahakian, 2006).
Although the most obvious potential use of drugs is to attempt to improve memory, drugs might also be used to help us forget. This might be desirable in some cases, such as for those suffering from posttraumatic stress disorder (PTSD) who are unable to forget disturbing memories. Although there are no existing therapies that involve using drugs to help people forget, it is possible that they will be available in the future. These possibilities will raise some important ethical issues: Is it ethical to erase memories, and if it is, is it desirable to do so? Perhaps the experience of emotional pain is a part of being a human being. And perhaps the experience of emotional pain may help us cope with the trauma.
REFERENCES
Baddeley, A., Eysenck, M. W., & Anderson, M. C. (2009). Memory. Psychology Press.
Bahrick, H. P. (1984). Semantic memory content in permastore: Fifty years of memory for Spanish learned in school. Journal of Experimental Psychology: General, 113(1), 1–29. doi.org/10.1037/ 0096-3445.113.1.1
Bayley, P. J., & Squire, L. R. (2005). Failure to acquire new semantic knowledge in patients with large medial temporal lobe lesions. Hippocampus, 15(2), 273–280. doi.org/10.1002/hipo.20057
Bower, G. H. (1981). Mood and memory. American Psychologist, 36(2), 129–148. https://doi.org/10.1037/0003-066X.36.2.129
Bransford, J. D., & Johnson, M. K. (1972). Contextual prerequisites for understanding: Some investigations of comprehension and recall. Journal of Verbal Learning and Verbal Behavior, 11(6), 717–726. https://doi.org/10.1016/S0022-5371(72)80006-9
Chester, B. (2001). Restoring remembering: Hormones and memory. McGill Reporter, 33(10). https://web.archive.org/ web/20180810052901/https://www.mcgill.ca/reporter/33/10/sherwin/
Corkin, S., Amaral, D. G., González, R. G., Johnson, K. A., & Hyman, B. T. (1997). H. M.’s medial temporal lobe lesion: Findings from magnetic resonance imaging. The Journal of Neuroscience, 17(10), 3964–3979. https://doi.org/10.1523/JNEUROSCI.17-10-03964.1997
Craik, F. I., & Lockhart, R. S. (1972). Levels of processing: A framework for memory research. Journal of Verbal Learning and Verbal Behavior, 11(6), 671–684. https://doi.org/10.1016/S0022-5371(72)80001-X
Driskell, J. E., Willis, R. P., & Copper, C. (1992). Effect of overlearning on retention. Journal of Applied Psychology, 77(5), 615–622. doi. org/10.1037/0021-9010.77.5.615
Eich, E. (2008). Mood and memory at 26: Revisiting the idea of mood mediation in drug-dependent and place-dependent memory. In M. A. Gluck, J. R. Anderson, & S. M. Kosslyn (Eds.), Memory and mind: A festschrift for Gordon H. Bower (pp. 247–260). Lawrence Erlbaum Associates.
Eichenbaum, H. (1999). Conscious awareness, memory, and the hippocampus. Nature Neuroscience, 2(9), 775–776. doi. org/10.1038/12137
Farah, M. J., Illes, J., Cook-Deegan, R., Gardner, H., Kandel, E., King, P., Parens, E., Sahakian, B., & Wolpe, P. R. (2004). Neurocognitive enhancement: What can we do and what should we do? Nature Reviews Neuroscience, 5(5), 421–425. https://doi.org/10.1038/ nrn1390
Godden, D. R., & Baddeley, A. D. (1975). Context-dependent memory in two natural environments: On land and underwater. British Journal of Psychology, 66(3), 325–331. doi.org/10.1111/j.2044-8295.1975. tb01468.x
Gold, P. E., Cahill, L., & Wenk, G. L. (2002). Ginkgo biloba: A cognitive enhancer? Psychological Science in the Public Interest, 3(1), 2–11. https://doi.org/10.1111/1529-1006.00006
Harris, J. L., & Qualls, C. D. (2000). The association of elaborative or maintenance rehearsal with age, reading comprehension and verbal working memory performance. Aphasiology, 14(5–6), 515–526. https://doi.org/10.1080/026870300401289
Helmuth, L. (1999). New role found for the hippocampus. Science, 285,
1339–1341. https://doi.org/10.1126/science.285.5432.1339b Jackson, A., Koek, W., & Colpaert, F. (1992). NMDA antagonists make
learning and recall state-dependent. Behavioural Pharmacology, 3(4), 415. doi.org/10.1097/00008877-199208000-00018
Jonides, J., Lacey, S. C., & Nee, D. E. (2005). Processes of working memory in mind and brain. Current Directions in Psychological Science, 14(1), 2–5. https://doi.org/10.1111/j.0963-7214.2005.00323.x
Krupa, D. J., Thompson, J. K., & Thompson, R. F. (1993). Localization of a memory trace in the mammalian brain. Science, 260(5110), 989–991. https://doi.org/10.1126/science.8493536
Lashley, K. S. (1929). The effects of cerebral lesions subsequent to the formation of the maze habit: Localization of the habit. In Brain mechanisms and intelligence: A quantitative study of injuries to the brain (pp. 86–108). University of Chicago Press.
Lynch, G. (2002). Memory enhancement: The search for mechanism- based drugs. Nature Neuroscience, 5(Suppl.), 1035–1038. doi. org/10.1038/nn935
Lynch, G., Larson, J., Staubli, U., Ambros-Ingerson, J., Granger, R., Lister, R. G., . . . Weingartner, H. J. (1991). Long-term potentiation and memory operations in cortical networks. In C. A. Wickliffe, M. Corballis, & G. White (Eds.), Perspectives on cognitive neuroscience (pp. 110–131). Oxford University Press.
Maki, P. M., & Resnick, S. M. (2000). Longitudinal effects of estrogen replacement therapy on PET cerebral blood flow and cognition. Neurobiology of Aging, 21(2), 373–383. doi.org/10.1016/ S0197-4580(00)00123-8
Marian, V. & Kaushanskaya, M. (2007). Language context guides memory content. Psychonomic Bulletin and Review, 14(5), 925–933. https:// doi.org/10.3758/BF03194123
McDaniel, M. A., Maier, S. F., & Einstein, G. O. (2002). “Brain-specific” nutrients: A memory cure? Psychological Science in the Public Interest, 3(1), 12–38. https://doi.org/10.1111/1529-1006.00007
McEntee, W., & Crook, T. (1993). Glutamate: Its role in learning, memory, and the aging brain. Psychopharmacology, 111(4), 391–401. https:// doi.org/10.1007/BF02253527
McGaugh, J. L. (2003). Memory and emotion: The making of lasting memories. Columbia University Press.
Nickerson, R. S., & Adams, M. J. (1979). Long-term memory for a common object. Cognitive Psychology, 11(3), 287–307. doi.org/10.1016/ 0010-0285(79)90013-6
Paller, K. A. (2004). Electrical signals of memory and of the awareness of remembering. Current Directions in Psychological Science, 13(2), 49–55. https://doi.org/10.1111/j.0963-7214.2004.00273.x
Rogers, T. B., Kuiper, N. A., & Kirker, W. S. (1977). Self-reference and the encoding of personal information. Journal of Personality and Social Psychology, 35(9), 677–688. doi.org/10.1037/ 0022-3514.35.9.677
Rosch, E. (1975). Cognitive representations of semantic categories. Journal of Experimental Psychology: General, 104(3), 192–233. https://doi.org/10.1037/0096-3445.104.3.192
Sherwin, B. B. (1998). Estrogen and cognitive functioning in women. Experimental Biology and Medicine, 217, 17–22. https://doi.org/ 10.3181/00379727-217-44200
Sigurdsson, T., Doyère, V., Cain, C. K., & LeDoux, J. E. (2007). Long-term potentiation in the amygdala: A cellular mechanism of fear learning and memory. Neuropharmacology, 52(1), 215–227. https://doi.org/ 10.1016/j.neuropharm.2006.06.022
Squire, L. R. (1992). Memory and the hippocampus: A synthesis from findings with rats, monkeys, and humans. Psychological Review, 99(2), 195–231. https://doi.org/10.1037/0033-295X.99.2.195
Srull, T., & Wyer, R. (1989). Person memory and judgment. Psychological Review, 96(1), 58–83. https://doi.org/10.1037/0033-295X.96.1.58
Symons, C. S., & Johnson, B. T. (1997). The self-reference effect in memory: A meta-analysis. Psychological Bulletin, 121(3), 371–394. https://doi.org/10.1037/0033-2909.121.3.371
Turner, D. C., & Sahakian, B. J. (2006). Analysis of the cognitive enhancing effects of modafinil in schizophrenia. In J. L. Cummings (Ed.), Progress in neurotherapeutics and neuropsychopharmacology (pp. 133–147). Cambridge University Press.
Woodruff-Pak, D. S., Goldenberg, G., Downey-Lamb, M. M., Boyko, O. B., & Lemieux, S. K. (2000). Cerebellar volume in humans related to magnitude of classical conditioning. NeuroReport, 11(3), 609–615. doi.org/10.1097/00001756-200002280-00035
20.05: Key Takeways
• Information is better remembered when it is meaningfully elaborated.
• Hermann Ebbinghaus made important contributions to the study of learning, including modeling the forgetting curve, and studying the spacing effect and the benefits of overlearning.
• Context- and state-dependent learning, as well as primacy and recency effects, influence long-term memory.
• Memories are stored in connected synapses through the process of long-term potentiation (LTP). In addition to the cortex, other parts of the brain, including the hippocampus, the cerebellum, and the amygdala, are also important in memory.
• Damage to the brain may result in retrograde amnesia or anterograde amnesia. Case studies of patients with amnesia can provide information about the brain structures involved in different types of memory.
• Memory is influenced by chemicals including glutamate, serotonin, epinephrine, and estrogen.
• Studies comparing memory enhancers with placebo drugs find very little evidence for their effectiveness.
EXERCISES AND CRITICAL THINKING
1. Plan a course of action to help you study for your next exam, incorporating as many of the techniques mentioned in this section as possible. Try to implement the plan.
2. Make a list of some the schemas that you have stored in your memory. What are the contents of each schema, and how might you use the schema to help you remember new information?
3. In the film Eternal Sunshine of the Spotless Mind, the characters undergo a medical procedure designed to erase their memories of a painful romantic relationship. Would you engage in such a procedure if it was safely offered to you? | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/20%3A_How_We_Remember-_Cues_to_Improving_Memory/20.04%3A_The_Biology_Of_Memory.txt |
The typical outcome of social influence is that our beliefs and behaviors become more similar to those of others around us. At times this change occurs in a spontaneous and automatic sense, without any obvious intent of one person to change the other. Perhaps you learned to like jazz or rap music because your roommate was playing a lot of it. You didn’t really want to like the music, and your roommate didn’t force it on you—your preferences changed in a passive way. Robert Cialdini and his colleagues (1990) found that col- lege students were more likely to throw litter on the ground when they had just seen another person throw some paper on the ground and were least likely to litter when they had just seen another person pick up and throw paper into a trash can. The researchers interpreted this as a kind of sponta- neous conformity—a tendency to follow the behavior of oth- ers, often entirely out of our awareness. Even our emotional states become more similar to those we spend more time with (Anderson et al., 2003).
21.02: Informational Conformity- Conforming To Be Accurate
Although mimicry represents the more subtle side, conformity also occurs in a more active and thoughtful sense, for instance, when we actively look to our friends’ opinions to determine appropriate behavior, when a car salesperson attempts to make a sale, or even when a powerful dictator uses physical aggression to force the people in his country to engage in the behaviors that he desires. In these cases, the influence is obvious. We know we are being influenced and we may attempt—sometimes successfully, and sometimes less so—to counteract the pressure.
Influence sometimes occurs because we believe that other people have valid knowledge about an opinion or issue, and we use that information to help us make good decisions. When we take our winter coat to school because the weatherman says it’s going to be cold, this is because we think that the weatherman has some good information that we can use. Informational conformity is the change in opinions or behavior that occurs when we conform to people whom we believe have accurate information. We base our beliefs on those presented to us by reporters, scientists, doctors, and lawyers because we believe they have more expertise in certain fields than we have. But we also use our friends and colleagues for information; when we choose a prom gown on the basis of our friends’ advice about what looks good on us, we are using informational conformity—we believe that our friends have good judgment about the things that matter to us.
Informational conformity is often the end result of social comparison, the process of comparing our opinions with those of others to gain an accurate appraisal of the validity of an opinion or behavior (Festinger et al., 1950; Hardin & Higgins, 1996; Turner, 1991). Informational conformity leads to real, long-lasting changes in beliefs. The result of informational influence is normally private acceptance: real change in opinions on the part of the individual. We believe that taking the winter coat was the right thing to do and that the prom gown really looks good on us. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/21%3A_The_Many_Varieties_of_Conformity/21.01%3A_Introduction.txt |
In other cases we conform not because we want to have valid knowledge but rather to meet the goal of belonging to and being accepted by a group we care about (Deutsch & Gerard, 1955). When we start smoking cigarettes or buy shoes we cannot really afford in order to impress others, we do these things not so much because we think they are the right things to do but rather because we want to be liked.
RESEARCH FOCUS
Imitation as Subtle Conformity
Perhaps you have noticed in your own behavior a type of very subtle conformity—the tendency to imitate other people who are around you. Have you ever found yourself talking, smiling, or frowning in the same way that a friend does? Tanya Chartrand and John Bargh (1999) investigated whether the tendency to imitate others would occur even for strangers, and even in very short periods of time.
In their first experiment, students worked on a task with another student, who was actually an experimental confederate. The two worked together to discuss photographs taken from current magazines. While they were working together, the confederate engaged in some unusual behaviors to see if the research participant would mimic them. Specifically, the confederate either rubbed his or her face or shook his or her foot. It turned out that the students did mimic the behavior of the confederate, by themselves either rubbing their own faces or shaking their own feet. And when the experimenters asked the participants if they had noticed anything unusual about the behavior of the other person during the experiment, none of them indicated awareness of any face rubbing or foot shaking.
It is said that imitation is a form of flattery, and we might therefore expect that we would like people who imitate us. Indeed, in a second experiment, Chartrand and Bargh found exactly this. Rather than creating the behavior to be mimicked, in this study the confederate imitated the behaviors of the participant. While the participant and the confederate discussed the magazine photos, the confederate mirrored the posture, movements, and mannerisms displayed by the participant.
As you can see in Figure \(1\), the participants who had been mimicked liked the other person more and indicated that they thought the interaction had gone more smoothly, in comparison with the participants who had not been imitated.
Imitation is an important part of social interaction. We easily and frequently mimic others without being aware that we are doing so. We may communicate to others that we agree with their viewpoints by mimicking their behaviors, and we tend to get along better with people with whom we are well “coordinated.” We even expect people to mimic us in social interactions, and we become distressed when they do not (Dalton et al., 2010). This unconscious conformity may help explain why we hit it off immediately with some people and never get it together with others (Chartrand & Dalton, 2009; Tickle-Degnen & Rosenthal, 1990, 1992). ■
Normative conformity occurs when we express opinions or behave in ways that help us to be accepted or that keep us from being isolated or rejected by others. When we engage in normative conformity we conform to social norms—socially accepted beliefs about what we do or should do in particular social contexts (Cialdini, 1993; Sherif, 1936; Sumner, 1906).
In contrast to informational conformity, in which the attitudes or opinions of the individual change to match that of the influencers, the outcome of normative conformity often represents public conformity rather than private acceptance. Public conformity is a superficial change in behavior (including the public expression of opinions) that is not accompanied by an actual change in one’s private opinion. Conformity may appear in our public behavior even though we may believe something completely different in private. We may obey the speed limit or wear a uniform to our job (behavior) to conform to social norms and requirements, even though we may not necessarily believe that it is appropriate to do so (opinion). We may use drugs with our friends without really wanting to, and without believing it is really right, because our friends are all using drugs. However, behaviors that are originally performed out of a desire to be accepted (normative conformity) may frequently produce changes in beliefs to match them, and the result becomes private acceptance. Perhaps you know someone who started smoking to please his friends but soon convinced himself that it was an acceptable thing to do.
Although in some cases conformity may be purely informational or purely normative, in most cases the goals of being accurate and being accepted go hand-in-hand, and therefore informational and normative conformity often occur at the same time. When soldiers obey their commanding officers, they probably do it both because others are doing it (normative conformity) and because they think it is the right thing to do (informational conformity). It has been argued that the distinction between informational and normative conformity is more apparent than real and that it may not be possible to fully differentiate them (Turner, 1991). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/21%3A_The_Many_Varieties_of_Conformity/21.03%3A_Normative_Conformity-_Conforming_To_Be_Liked_And_To_Avoid_Rejection.txt |
Although conformity occurs whenever group members change their opinions or behaviors as a result of their perceptions of others, we can divide such influence into two types. Majority influence occurs when the beliefs held by the larger number of individuals in the current social group prevail. In contrast, minority influence occurs when the beliefs held by the smaller number of individuals in the current social group prevail. Not surprisingly, majority influence is more common, and we will consider it first.
In a series of important studies on conformity, Muzafer Sherif (1936) used a perceptual phenomenon known as the autokinetic effect to study the outcomes of conformity on the development of group norms. The autokinetic effect is caused by the rapid, small movements of our eyes that occur as we view objects and that allow us to focus on stimuli in our environment. However, when individuals are placed in a dark room that contains only a single, small, stationary pinpoint of light, these eye movements produce an unusual effect for the perceiver—they make the point of light appear to move.
Sherif took advantage of this effect to study how group norms develop in ambiguous situations. In his studies, college students were placed in a dark room with the point of light and were asked to indicate, each time the light was turned on, how much it appeared to move. Some participants first made their judgments alone. Sherif found that although each participant who was tested alone made estimates that were within a relatively narrow range (as if they had their own “individual” norm), there were wide variations in the size of these judgments among the different participants he studied.
Sherif also found that when individuals who initially had made very different estimates were then placed in groups along with one or two other individuals, and in which all the group members gave their responses on each trial aloud (each time in a different random order), the initial differences in judgments among the participants began to disappear, such that the group members eventually made very similar judgments. You can see that this pattern of change, which is shown in Figure \(1\), illustrates the fundamental principle of social influence—over time, people come more and more to share their beliefs with each other. Sherif ’s study is thus a powerful example of the development of group norms.
Furthermore, and indicating that Sherif had created private acceptance, the new group norms continued to influence judgments when the individuals were again tested alone. The participants did not revert back to their initial opinions, even though they were quite free to do so; rather, they stayed with the new group norms. And these conformity effects appear to have occurred entirely out of the awareness of most participants. Sherif (1936, p. 108) reported that the majority of the participants indicated after the experiment was over that their judgments had not been influenced by the judgments made by the other group members.
Sherif also found that the norms that were developed in groups could continue over time. When the original research participants were moved into groups with new people, their opinions subsequently influenced the judgments of the new group members (Jacobs & Campbell, 1961). The norms persisted through several “generations” (MacNeil & Sherif, 1976) and could influence individual judgments up to a year after the individual was last tested (Rohrer et al., 1954).
When Solomon Asch (1952, 1955) heard about Sherif’s studies, he responded in perhaps the same way that you might have: “Well of course people conformed in this situation, because after all the right answer was very unclear,” you might have thought. Since the study participants didn’t know the right answer (or indeed the “right” answer was no movement at all), it is perhaps not that surprising that people conformed to the beliefs of others.
Asch conducted studies in which, in complete contrast to the autokinetic effect experiments of Sherif, the correct answers to the judgments were entirely unambiguous. In these studies, the research participants were male college students who were told that they were to be participating in a test of visual abilities. The men were seated in a small semicircle in front of a board that displayed the visual stimuli that they were going to judge. The men were told that there would be 18 trials during the experiment, and on each trial they would see two cards. The standard card had a single line that was to be judged, and the test card had three lines that varied in length between about 2 and 10 inches:
The men’s task was simply to indicate which line on the test card was the same length as the line on the standard card. As you can see from the Asch card sample above, there is no question that the correct answer is line 1. In fact, Asch found that people made virtually no errors on the task when they made their judgments alone.
On each trial, each person answered out loud, beginning with one end of the semicircle and moving to the other end. Although the participant did not know it, the other group members were not true participants but experimental confederates who gave predetermined answers on each trial. Because the participant was seated next to last in the row, he always made his judgment after most of the other group members made theirs. Although on the first two trials the confederates each gave the correct answer, on the third trial, and on 11 of the subsequent trials, they all had been instructed to give the same incorrect answer. For instance, even though the correct answer was Line 1, they would all say it was Line 2. Thus when it became the participant’s turn to answer, he could either give the clearly correct answer or conform to the incorrect responses of the confederates.
Asch found that about 76% of the 123 men who were tested gave at least one incorrect response when it was their turn, and 37% of the responses, overall, were conforming. This is indeed evidence for the power of conformity because the research participants were giving clearly incorrect answers out loud. However, conformity was not absolute—in addition to the 24% of the men who never conformed, only 5% of the men conformed on all 12 of the critical trials.
You might think that Asch’s study shows primarily normative conformity, in the sense that the participants gave the wrong answer not because they thought it was correct but because they wanted to avoid being seen as different. But is it not possible that the participants were also a bit unsure of the correct answer? When I look at the picture of one of the participants in Asch’s studies, it looks to me like he is really unsure about what he is seeing. Perhaps, then, even in Asch’s study some informational influence is occurring. Maybe the participants became unsure of their responses and used the opinions of others to help them determine what was correct. Normative and informational conformity often operate together. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/21%3A_The_Many_Varieties_of_Conformity/21.04%3A_Majority_Influence-_Conforming_To_The_Group.txt |
The research that we have discussed to this point involves conformity in which the opinions and behaviors of individuals become more similar to the opinions and behaviors of the majority of the people in the group—majority influence. But we do not always blindly conform to the beliefs of the majority. Although more unusual, there are nevertheless cases in which a smaller number of individuals are able to influence the opinions or behaviors of the group—this is minority influence.
It is a good thing that minorities can be influential; other- wise, the world would be pretty boring. When we look back on history we find that it is the unusual, divergent, innovative minority groups or individuals who—although frequently ridiculed at the time for their unusual ideas—end up being respected for producing positive changes. The work of scientists, religious leaders, philosophers, writers, musicians, and artists who go against group norms by expressing new and unusual ideas frequently is not liked at first. These novel thinkers may be punished—in some cases even killed—for their beliefs. In the end, however, if the ideas are interesting and important, the majority may conform to these new ideas, producing social change. In short, although conformity to majority opinions is essential to provide a smoothly working society, if individuals only conformed to others there would be few new ideas and little social change.
The French social psychologist Serge Moscovici was particularly interested in the situations under which minority influence might occur. In fact, he argued that all members of all groups are able, at least in some degree, to influence others, regardless of whether they are in the majority or the minority. To test whether minority group members could indeed produce influence, he and his colleagues (Moscovici et al., 1969) created the opposite of Asch’s line perception study, such that there was now a minority of confederates in the group (two) and a majority of experimental participants (four). All six individuals viewed a series of slides depicting colors, supposedly as a study of color perception, and as in Asch’s research, each voiced out loud an opinion about the color of the slide.
Although the color of the slides varied in brightness, they were all clearly blue. Moreover, demonstrating that the slides were unambiguous, just as the line judgments of Asch had been, participants who were asked to make their judgments alone called the slides a different color than blue less than 1% of the time. (When it happened, they called the slides green.)
In the experiment, the two confederates had been instructed to give one of two patterns of answers that were different from the normal responses. In the consistent-minority condition, the two confederates gave the unusual response (green) on every trial. In the inconsistent-minority condition the confederates called the slides “green” on two thirds of their responses and called them blue on the other third.
The minority of two was able to change the beliefs of the majority of four, but only when they were unanimous in their judgments. As shown in Figure \(2\), Moscovici found that the presence of a minority who gave consistently unusual responses influenced the judgments made by the experimental participants. When the minority was consistent, 32% of the majority group participants said green at least once and 18% of the responses of the majority group were green. How- ever, as you can see in Figure \(2\), the inconsistent minority had virtually no influence on the judgments of the majority.
On the basis of this research, Moscovici argued that minorities could have influence over majorities, provided they gave consistent, unanimous responses. Subsequent research has found that minorities are most effective when they express consistent opinions over time and with each other, when they show that they are invested in their position by making significant personal and material sacrifices, and when they seem to be acting out of principle rather than from ulterior motives (Hogg, 2010). Although they may want to adopt a relatively open-minded and reasonable negotiating style on issues that are less critical to the attitudes they are trying to change, successful minorities must be absolutely consistent with regard to their core arguments (Mugny & Papastamou, 1981).
When minorities are successful at producing influence, they are able to produce strong and lasting attitude change— true private acceptance—rather than simply public conformity. People conform to minorities because they think that they are right, and not because they think it is socially acceptable. Minorities have another, potentially even more important, outcome on the opinions of majority group members—the presence of minority groups can lead majorities to engage in fuller, as well as more divergent, innovative and creative thinking about the topics being discussed (Martin & Hewstone, 2003; Martin et al., 2007).
Nemeth and Kwan (1987) had participants work in groups of four on a creativity task in which they were presented with letter strings such as tdogto and asked to indicate which word came to their mind first as they looked at the letters. The judgments were made privately, which allowed the experimenters to provide false feedback about the responses of the other group members. All participants indicated the most obvious word (in this case, dog) as their response on each of the initial trials. However, the participants were told (according to experimental condition) either that three of the other group members had also reported seeing dog and that one had reported seeing god or that three out of the four had reported seeing god whereas only one had reported dog. Participants then completed other similar word strings on their own, and their responses were studied.
Results showed that when the participants thought that the unusual response (for instance, god rather than dog) was given by a minority of one individual in the group rather than by a majority of three individuals, they subsequently answered more of the new word strings using novel solutions, such as finding words made backward or using a random order of the letters. On the other hand, the individuals who thought that the majority of the group had given the novel response did not develop more creative ideas. The idea is when the participants thought that the novel response came from a group minority (one person), they thought about the responses more carefully, in comparison with the same behaviors performed by majority group members, and this led them to adopt new and creative ways to think about the problems. This result, along with other research showing similar findings, suggests that messages that come from minority groups lead us to think more fully about the decision, which can produce innovative, creative thinking in majority group members (Crano & Chen, 1998).
In summary, we can conclude that minority influence, although not as likely as majority influence, does sometimes occur. The few are able to influence the many when they are consistent and confident in their judgments but are less able to have influence when they are inconsistent or act in a less confident manner. Furthermore, although minority influence is difficult to achieve, if it does occur it is powerful. When majorities are influenced by minorities they really change their beliefs—the outcome is deeper thinking about the message, private acceptance of the message, and in some cases even more creative thinking. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/21%3A_The_Many_Varieties_of_Conformity/21.05%3A_Minority_Influence-_Resisting_Group_Presure.txt |
The studies of Asch, Sherif, and Moscovici demonstrate the extent to which individuals—both majorities and minorities— can create conformity in others. Furthermore, these studies provide information about the characteristics of the social situation that are important in determining the extent to which we conform to others. Let’s consider some of those variables.
The Size of the Majority
As the number of people in the majority increases, relative to the number of persons in the minority, pressure on the minority to conform also increases (Latané, 1981; Mullen, 1983). Asch conducted replications of his original line-judging study in which he varied the number of confederates (the majority subgroup members) who gave initial incorrect responses from 1 to 16 people, while holding the number in the minority subgroup constant at 1 (the single research par- ticipant). You may not be surprised to hear the results of this research: When the size of the majorities got bigger, the lone participant was more likely to give the incorrect answer.
Increases in the size of the majority increase conformity, regardless of whether the conformity is informational or normative. In terms of informational conformity, if more people express an opinion, their opinions seem more valid. Thus bigger majorities should result in more informational conformity. But larger majorities will also produce more normative conformity because being different will be harder when the majority is bigger. As the majority gets bigger, the individual giving the different opinion becomes more aware of being different, and this produces a greater need to conform to the prevailing norm.
Although increasing the size of the majority does increase conformity, this is only true up to a point. The increase in the amount of conformity that is produced by adding new members to the majority group (known as the social impact of each group member) is greater for initial majority members than it is for later members (Latané, 1981). This pattern is shown in Figure \(1\), which presents data from a well- known experiment by Stanley Milgram and his colleagues (1969) that studied how people are influenced by the behavior of others on the streets of New York City.
Milgram had confederates gather in groups on 42nd Street in New York City, in front of the Graduate Center of the City University of New York, each looking up at a window on the sixth floor of the building. The confederates were formed into groups ranging from one to 15 people. A video camera in a room on the sixth floor above recorded the behavior of 1,424 pedestrians who passed along the sidewalk next to the groups.
As you can see in Figure \(1\), larger groups of confederates increased the number of people who also stopped and looked up, but the influence of each additional confederate was generally weaker as size increased. Groups of three con- federates produced more conformity than did a single person, and groups of five produced more conformity than groups of three. But after the group reached about six people, it didn’t really matter very much. Just as turning on the first light in an initially dark room makes more difference in the brightness of the room than turning on the second, third, and fourth lights does, adding more people to the majority tends to pro- duce diminishing returns—less effect on conformity.
Group size is an important variable that influences a wide variety of behaviors of the individuals in groups. People leave proportionally smaller tips in restaurants as the number in their party increases, and people are less likely to help as the number of bystanders to an incident increases (Latané, 1981). The number of group members also has an important influence on group performance: As the size of a working group gets larger, the contributions of each individual member to
the group effort become smaller. In each case, the influence of group size on behavior is found to be similar to that shown in Figure \(1\).
As you can see in Figure \(1\), the effect of adding new individuals to the group eventually levels off entirely, such that adding more people to the majority after that point makes no meaningful difference in the amount of conformity. This peak usually occurs when the majority has about four or five persons. One reason that the impact of new group members decreases so rapidly is because as the number in the group increases, the individuals in the majority are soon seen more as a group rather than as separate individuals. When there are only a couple of individuals expressing opinions, each per- son is likely to be seen as an individual, holding his or her own unique opinions, and each new individual adds to the impact. As a result, two people are more influential than one, and three more influential than two. However, as the number of individuals grows, and particularly when those individuals are perceived as being able to communicate with each other, the individuals are more likely to be seen as a group rather than as individuals. At this point, adding new members does not change the perception; regardless of whether there are four, five, six, or more members, the group is still just a group. As a result, the expressed opinions or behaviors of the group members no longer seem to reflect their own characteristics, so much as they do that of the group as a whole, and thus increasing the number of group members is less effective in increasing influence (Wilder, 1977).
The Unanimity of the Majority
Although the number of people in the group is an important determinant of conformity, it cannot be the only thing—if it were, minority influence would be impossible. It turns out that the consistency or unanimity of the group members is even more important. In Asch’s study, as an example, conformity occurred not so much because many confederates gave a wrong answer but rather because each of the confederates gave the same wrong answer. In one follow-up study that he conducted, Asch increased the number of confederates to 16 but had just one of those confederates give the correct answer. He found that in this case, even though there were 15 incorrect and only one correct answer given by the confederates, conformity was nevertheless sharply reduced—to only about 5% of the participants’ responses. And you will recall that in the minority influence research of Moscovici, the same thing occurred; conformity was only observed when the minority group members were completely consistent in their expressed opinions.
Although you might not be surprised to hear that conformity decreases when one of the group members gives the right answer, you may be more surprised to hear that conformity is reduced even when the dissenting confederate gives a different wrong answer. For example, conformity is reduced dramatically in Asch’s line-judging situation, such that virtually all participants give the correct answer (assume it is line 3 in this case) even when the majority of the con- federates have indicated that line 2 is the correct answer and a single confederate indicates that line 1 is correct. In short, conformity is reduced when there is any inconsistency among the members of the majority group—even when one member of the majority gives an answer that is even more incorrect than that given by the other majority group members (Allen & Levine, 1968).
Why should unanimity be such an important determinant of conformity? For one, when there is complete agreement among the majority members, the individual who is the target of influence stands completely alone and must be the first to break ranks by giving a different opinion. Being the only person who is different is potentially embarrassing, and people who wish to make a good impression on, or be liked by, others may naturally want to avoid this. If you can convince your friend to wear blue jeans rather than a coat and tie to a wedding, then you’re naturally going to feel a lot less conspicuous when you wear jeans too.
RESEARCH FOCUS
How Task importance and Confidence influence Conformity
The joint influence of an individual’s confidence in his or her beliefs and the importance of the task was demonstrated in an experiment conducted by Baron et al. (1996) that used a slight modification of the Asch procedure to assess conformity. Participants completed the experiment along with two other students, who were actually experimental confederates. The participants worked on several different types of trials, but there were 26 that were relevant to the conformity predictions. On these trials, a photo of a single individual was presented first, followed immediately by a “lineup” photo of four individuals, one of whom had been viewed in the initial slide (but who might have been dressed differently).
The participants’ task was to call out which person in the lineup was the same as the original individual using a number between 1 (the person on the left) and 4 (the person on the right). In each of the critical trials the two confederates went before the participant and they each gave the same wrong response.
Two experimental manipulations were used. First, the researchers manipulated task importance by telling some participants (the high importance condition) that their performance on the task was an important measure of eyewitness ability and that the participants who performed most accu- rately would receive \$20 at the end of the data collection. (A lottery using all the participants was actually held at the end of the semester, and some participants were paid the
Second, when there is complete agreement—remember the consistent minority in the studies by Moscovici—the participant may become less sure of his or her own perceptions. Because everyone else is holding the exact same opinion, it seems that they must be correctly responding to the external reality. When such doubt occurs, the individual may be likely to conform due to informational conformity. Finally, when one or more of the other group members gives a different answer than the rest of the group (so that the unanimity of the majority group is broken), that person is no longer part of the group that is doing the influencing and becomes (along with the participant) part of the group being influenced. You can see that another way of describing the effect of unanimity is to say that as soon as the individual has someone who agrees with him or her that the others may not be correct (a supporter or ally), then the pressure to con- form is reduced. Having one or more supporters who challenge the status quo validates one’s own opinion and makes disagreeing with the majority more likely (Allen, 1975; Boyanowsky & Allen, 1973).
The importance of the Task
Still another determinant of conformity is the perceived importance of the decision. The studies of Sherif, Asch, and Moscovici may be criticized because the decisions that the participants made—for instance, judging the length of lines or the colors of objects—seem rather trivial. But what would happen when people were asked to make an important decision? Although you might think that conformity would be less when the task becomes more important (perhaps because people would feel uncomfortable relying on the judgments of others and want to take more responsibility for their own decisions), the influence of task importance actually turns out to be more complicated than that.
\$20.) Participants in the low-importance condition, on the other hand, were told that the test procedure was part of a pilot study and that the decisions were not that important. Second, task difficulty was varied by showing the test and the lineup photos for 5 and 10 seconds, respectively (easy condition) or for only 1⁄2 and 1 second, respectively (difficult condition). The conformity score was defined as the number of trials in which the participant offered the same (incorrect) response as the confederates.
As you can see in Figure \(3\), an interaction between task difficulty and task importance was observed. On easy tasks, participants conformed less to the incorrect judg- ments of others when the decision had more important con- sequences for them. In these cases, they seemed to rely more on their own opinions (which they were convinced were correct) when it really mattered, but were more likely to go along with the opinions of the others when things were not that critical (probably normative conformity).
On the difficult tasks, however, results were the oppo- site. In this case, participants conformed more when they thought the decision was of high, rather than low, impor- tance. In these cases in which they were more unsure of their opinions and yet they really wanted to be correct, they used the judgments of others to inform their own views (informational conformity). ■
REFERENCES
Allen, V. L. (1975). Social support for nonconformity. In L. Berkowitz (Ed.), Advances in experimental social psychology (Vol. 8). Academic Press.
Allen, V. L., & Levine, J. M. (1968). Social support, dissent and conformity. Sociometry, 31(2), 138–149. doi.org/10.2307/2786454
Anderson, C., Keltner, D., & John, O. P. (2003). Emotional convergence between people over time. Journal of Personality and Social Psychology, 84(5), 1054–1068. doi.org/10.1037/ 0022-3514.84.5.1054
Asch, S. E. (1952). Social psychology. Prentice-Hall.
Asch, S. E. (1955). Opinions and social pressure. Scientific American, 11,
32. doi.org/10.1038/scientificamerican1155-31
Baron, R. S., Vandello, J. A., & Brunsman, B. (1996). The forgotten
variable in conformity research: Impact of task importance on social influence. Journal of Personality and Social Psychology, 71(5), 915–927. https://doi.org/10.1037/0022-3514.71.5.915
Boyanowsky, E. O., & Allen, V. L. (1973). Ingroup norms and self-identity as determinants of discriminatory behavior. Journal of Personality and Social Psychology, 25(3), 408–418. doi.org/10.1037/ h0034212
Chartrand, T. L., & Bargh, J. A. (1999). The chameleon effect: The perception-behavior link and social interaction. Journal of Personality and Social Psychology, 76(6), 893–910. doi.org/10.1037/ 0022-3514.76.6.893
Chartrand, T. L., & Dalton, A. N. (2009). Mimicry: Its ubiquity, importance, and functionality. In E. Morsella, J. A. Bargh, & P. M. Gollwitzer (Eds.), Oxford handbook of human action (pp. 458–483). Oxford University Press.
Cialdini, R. B. (1993). Influence: Science and practice (3rd ed.). Harper Collins.
Cialdini, R. B., Reno, R. R., & Kallgren, C. A. (1990). A focus theory of normative conduct: Recycling the concept of norms to reduce littering in public places. Journal of Personality and Social Psychology, 58(6), 1015–1026. doi.org/10.1037/ 0022-3514.58.6.1015
Crano, W. D., & Chen, X. (1998). The leniency contract and persistence of majority and minority influence. Journal of Personality and Social Psychology, 74(6), 1437–1450. https://doi.org/10.1037/0022-3514.74.6.1437
Dalton, A. N., Chartrand, T. L., & Finkel, E. J. (2010). The schema-driven chameleon: How mimicry affects executive and self-regulatory resources. Journal of Personality and Social Psychology, 98(4), 605–617. https://doi.org/10.1037/a0017629
Deutsch, M., & Gerard, H. B. (1955). A study of normative and informational social influences upon individual judgment. Journal of Abnormal and Social Psychology, 51(3), 629–636. https://doi.org/ 10.1037/h0046408
Festinger, L., Schachter, S., & Back, K. (1950). Social pressures in informal groups. Harper.
Hardin, C., & Higgins, T. (1996). Shared reality: How social verification makes the subjective objective. In R. M. Sorrentino & E. T. Higgins (Eds.), Handbook of motivation and cognition: Foundations of social behavior (Vol. 3, pp. 28–84). Guilford.
Hogg, M. A. (2010). Influence and leadership. In S. F. Fiske, D. T. Gilbert, & G. Lindzey (Eds.), Handbook of social psychology (Vol. 2, pp. 1166– 1207). Wiley.
Jacobs, R. C., & Campbell, D. T. (1961). The perpetuation of an arbitrary tradition through several generations of a laboratory microculture. Journal of Abnormal and Social Psychology, 62(3), 649–658. https:// doi.org/10.1037/h0044182
Latané, B. (1981). The psychology of social impact. American Psychologist, 36(4), 343–356. doi.org/10.1037/ 0003-066X.36.4.343
MacNeil, M. K., & Sherif, M. (1976). Norm change over subject generations as a function of arbitrariness of prescribed norms. Journal of Personality and Social Psychology, 34(5), 762–773. https:// doi.org/10.1037/0022-3514.34.5.762
Martin, R., & Hewstone, M. (2003). Majority versus minority influence: When, not whether, source status instigates heuristic or systematic processing. European Journal of Social Psychology, 33(3), 313–330. doi.org/10.1002/ejsp.146
Martin, R., Martin, P. Y., Smith, J. R., & Hewstone, M. (2007). Majority versus minority influence and prediction of behavioral intentions and behavior. Journal of Experimental Social Psychology, 43(5), 763–771. https://doi.org/10.1016/j.jesp.2006.06.006
Milgram, S., Bickman, L., & Berkowitz, L. (1969). Note on the drawing power of crowds of different size. Journal of Personality and Social Psychology, 13(2), 79–82. https://doi.org/10.1037/h0028070
Moscovici, S., Lage, E., & Naffrechoux, M. (1969). Influence of a consistent minority on the responses of a majority in a color perception task. Sociometry, 32(4), 365–379. https://doi.org/ 10.2307/2786541
Mugny, G., & Papastamou, S. (1981). When rigidity does not fail: Individualization and psychologicalization as resistance to the diffusion of minority innovation. European Journal of Social Psychology, 10, 43–62. doi.org/10.1002/ejsp.2420100104
Mullen, B. (1983). Operationalizing the effect of the group on the individual: A self-attention perspective. Journal of Experimental Social Psychology, 19(4), 295–322. doi.org/10.1016/ 0022-1031(83)90025-2
Nemeth, C. J., & Kwan, J. L. (1987). Minority influence, divergent thinking, and the detection of correct solutions. Journal of Applied Social Psychology, 17(9), 788–799. doi.org/10.1111/j.1559-1816.1987. tb00339.x
Rohrer, J. H., Baron, S. H., Hoffman, E. L., & Swander, D. V. (1954). The stability of autokinetic judgments. The Journal of Abnormal and Social Psychology, 67(4, Pt. 1), 143–146. doi.org/10.1037/ h0060827
Sherif, M. (1936). The psychology of social norms. Harper & Row. Sumner, W. G. (1906). Folkways. Ginn.
Tickle-Degnen, L., & Rosenthal, R. (1990). The nature of rapport and its nonverbal correlates. Psychological Inquiry, 1(4), 285–293. https:// doi.org/10.1207/s15327965pli0104_1
Tickle-Degnen, L., & Rosenthal, R. (Eds.). (1992). Nonverbal aspects of therapeutic rapport. Lawrence Erlbaum.
Turner, J. C. (1991). Social influence. Brooks Cole.
Wilder, D. A. (1977). Perception of groups, size of opposition, and social influence. Journal of Experimental Social Psychology, 13(3), 253–268. https://doi.org/10.1016/0022-1031(77)90047-6
21.07: Key Takeways
• Social influence creates conformity.
• Influence may occur in more passive or more active ways.
• We conform both to gain accurate knowledge (informational conformity) and to avoid being rejected by others (normative conformity).
• Both majorities and minorities may create social influence, but they do so in different ways.
• The characteristics of the social situation, including the number of people in the majority and the unanimity of the majority, have a strong influence on conformity.
EXERCISES AND CRITICAL THINKING
1. Describe a time when you conformed to the opinions or behaviors of others. Interpret the conformity in terms of the concepts discussed in this section.
2. Imagine you were serving on a jury trial in which you found yourself the only person who believed that the defendant was innocent. What strategies might you use to convince the majority? | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/21%3A_The_Many_Varieties_of_Conformity/21.06%3A_Situational_Determinants_Of_Conformity.txt |
One of the fundamental aspects of social interaction is that some individuals have more influence than others. Social power can be defined as the ability of a person to create conformity even when the people being influenced may attempt to resist those changes (Fiske, 1993; Keltner et al., 2003). Bosses have power over their workers, parents have power over their children, and, more generally, we can say that those in authority have power over their subordinates. In short, power refers to the process of social influence itself— those who have power are those who are most able to influ- ence others.
22.02: Milgram's Studies On Obedience To Authority
The powerful ability of those in authority to control others was demonstrated in a remarkable set of studies performed by Stanley Milgram (1974). Milgram was interested in under- standing the factors that lead people to obey the orders given by people in authority. He designed a study in which he could observe the extent to which a person who presented himself as an authority would be able to produce obedience, even to the extent of leading people to cause harm to others.
MILGRAM’S OBEDIENCE TO AUTHORITY EXPERIMENT
Watch this video about Milgram’s experiment.
Like many other social psychologists, Milgram’s interest in conformity stemmed in part from his desire to under- stand how the presence of a powerful person—particularly the German dictator Adolph Hitler, who ordered the killing of millions of people during World War II—could produce obedience. Under Hitler’s direction, the German SS troops oversaw the execution of 6 million Jews as well as other “undesirables,” including political and religious dissidents, homosexuals, the mentally and physically disabled, and prisoners of war.
Milgram used newspaper ads to recruit men (and, in one study, women) from a wide variety of backgrounds to participate in his research. When the research participant arrived at the lab, he or she was introduced to a man whom the participant believed was another research participant but who was actually an experimental confederate. The experimenter explained that the goal of the research was to study the effects of punishment on learning. After the participant and the confederate both consented to be in the study, the researcher explained that one of them would be the teacher and the other the learner. They were each given a slip of paper and asked to open it and to indicate what it said. In fact, both papers read teacher, which allowed the confederate to pretend that he had been assigned to be the learner and thus to assure that the actual participant was always the teacher.
While the research participant (now the teacher) looked on, the learner was taken into the adjoining shock room and strapped to an electrode that was to deliver the punishment. The experimenter explained that the teacher’s job would be to sit in the control room and to read a list of word pairs to the learner. After the teacher read the list once, it would be the learner’s job to remember which words went together. For instance, if the word pair was blue-sofa, the teacher would say the word blue on the testing trials and the learner would have to indicate which of four possible words (house, sofa, cat, or carpet) was the correct answer by pressing one of four buttons in front of him.
After the experimenter gave the “teacher” a sample shock (which was said to be at 45 volts) to demonstrate that the shocks really were painful, the experiment began. The research participant first read the list of words to the learner and then began testing him on his learning. The shock panel was presented in front of the teacher, as shown in Figure \(1\), and the learner was not visible in the shock room. The experimenter sat behind the teacher and explained to him that each time the learner made a mistake the teacher was to press one of the shock switches to administer the shock. Moreover, which switch was to be pressed increased by one level with each mistake, such that each mistake required a stronger shock.
Once the learner (who was, of course, actually an experimental confederate) was alone in the shock room, he unstrapped himself from the shock machine and brought out a tape recorder that he used to play a prerecorded series of responses that the teacher could hear through the wall of the room.
As you can see in Table \(1\), the teacher heard the learner say “ugh!” after the first few shocks. After the next few mistakes, when the shock level reached 150 volts, the learner was heard to exclaim, “Let me out of here. I have heart trouble!” As the shock reached about 270 volts the learner’s protests became more vehement, and after 300 volts the learner pro- claimed that he was not going to answer any more questions.
Table \(1\): The Confederate’s Schedule of Protest in the Milgram Experiments
75 volts
Ugh!
90 volts
Ugh!
105 volts
Ugh! (louder)
120 volts
Ugh! Hey, this really hurts.
135 volts
Ugh!!
150 volts
Ugh!! Experimenter! That’s all. Get me out of here.
I told you I had heart trouble. My heart’s starting to bother me now. Get me out of here, please. My heart’s starting to bother me. I refuse to go on. Let me out!
165 volts
Ugh! Let me out! (shouting)
180 volts
Ugh! I can’t stand the pain. Let me out of here! (shouting)
195 volts
Ugh! Let me out of here! Let me out of here! My heart’s bothering me. Let me out of here! You have no right to keep me here! Let me out! Let me out of here! Let me out! Let me out of here! My heart’s bothering me. Let me out! Let me out!
210 volts
Ugh!! Experimenter! Get me out of here. I’ve had enough. I won’t be in the experiment any more.
225 volts
Ugh!
240 volts
Ugh!
255 volts
Ugh! Get me out of here.
270 volts
(agonized scream) Let me out of here. Let me out of here. Let me out of here. Let me out. Do you hear? Let me out of here.
285 volts
(agonized scream)
300 volts
(agonized scream) I absolutely refuse to answer any more. Get me out of here. You can’t hold me here. Get me out. Get me out of here.
315 volts
(intensely agonized scream) Let me out of here. Let me out of here. My heart’s bothering me. Let me out, I tell you. (hysterically) Let me out of here. Let me out of here. You have no right to hold me here. Let me out! Let me out! Let me out! Let me out of here! Let me out! Let me out!
From 330 volts and up the learner was silent. The experimenter responded to participants’ questions at this point, if they asked any, with a scripted response indicating that they should continue reading the questions and applying increasing shock when the learner did not respond.
The results of Milgram’s research were themselves quite shocking. Although all of the participants gave the initial mild levels of shock, responses varied after that. Some refused to continue after about 150 volts, despite the insistence of the experimenter to continue to increase the shock level. Still others, however, continued to present the questions, and to administer the shocks, under the pressure of the experimenter, who demanded that they continue. In the end, 65% of the participants continued giving the shock to the learner all the way up to the 450-volt maximum, even though that shock was marked as “danger: severe shock” and there had been no response heard from the participant for several trials. In sum, well over half of the men who participated had, as far as they knew, shocked another person to death, all as part of a supposed experiment on learning.
In case you are thinking that such high levels of obedience would not be observed in today’s modern culture, there is, in fact, evidence that they would be. Studies similar to Milgram’s findings have been conducted all over the world (Blass, 1999) and have found similar levels of conformity, with obedience rates ranging from a high of 90% in Spain and the Nether- lands (Meeus & Raaijmakers, 1986) to a low of 16% among Australian women (Kilham & Mann, 1974).
Recently, Milgram’s results were almost exactly replicated, using men and women from a wide variety of ethnic groups, in a study conducted by Jerry Burger at Santa Clara University. In this replication of the Milgram experiment, 65% of the men and 73% of the women agreed to administer increasingly painful electric shocks when they were ordered to by an authority figure (Borge, 2007). In the replication, however, the participants were not allowed to go beyond the 150 volt shock switch.
Although it might be tempting to conclude that Milgram’s experiments demonstrate that people are innately evil creatures who are ready to shock others to death, Milgram did not believe that this was the case. Rather, he felt that it was the social situation, and not the people themselves, that was responsible for the behavior. To demonstrate this, Milgram conducted research that explored a number of variations on his original procedure, each of which demonstrated that changes in the situation could dramatically influence the amount of conformity. These variations are summarized in Figure \(2\), which presents the percentage of participants who were maximally obedient (that is, who gave all 450 volts of shock) in some of the variations that he conducted.
In the initial study, the authority’s status and power was maximized—the experimenter had been introduced as a respected scientist at a respected university. However, in replications of the study in which the experimenter’s authority was decreased, obedience also declined. In one replication, the status of the experimenter was reduced by having the experiment take place in a building located in Bridgeport, Connecticut, rather than at the labs on the Yale University campus, and the research was ostensibly sponsored by a private commercial research firm instead of by the university. In this study, less obedience was observed (only 48% of the participants delivered the maximum shock). Full obedience was also reduced (to 20%) when the experimenter’s ability to express his authority was limited by having him sit in an adjoining room and communicate to the teacher by telephone. And when the experimenter left the room and had another student (actually a confederate) give the instructions for him, conformity was also reduced to 20%.
In addition to the role of authority, Milgram’s studies also confirmed the role of unanimity in producing conformity. When another research participant (again an experimental confederate) began by giving the shocks but then later refused to continue and the participant was asked to take over, only 10% were obedient. And if two experimenters were present but only one proposed shocking while the other argued for stopping the shocks, all the research participants took the more benevolent advice and did not shock. Finally, in another condition the teacher sat next to the learner and was forced to wear an insulated glove and to hold the learner’s hand on the shock pad as he administered the shock. Again, obedience was reduced, this time to only 30%.
Perhaps most telling regarding the important role of the situation were the studies in which Milgram allowed the participants to choose their own shock levels or in which one of the experimenters suggested that they should not actually use the shock machine. In these situations, there was virtually no shocking. These conditions show that people do not like to harm others, and when given a choice they will not. On the other hand, the social situation can create powerful, and potentially deadly, social influence.
SOCIAL PSYCHOLOGY IN THE PUBLIC INTERSEST
The Zimbardo Prison Studies and Abu Ghraib
In Milgram’s research we can see a provocative demonstration of how people who have power can control the behavior of others. Can our understanding of the social psychological factors that produce conformity and obedience help us explain the events that occurred in 2004 at Abu Ghraib, the Iraqi prison in which U.S. soldiers physically and psychologically tortured their Iraqi prisoners? The social psychologist Philip Zimbardo thinks so. He notes the parallels between the events that occurred at Abu Ghraib and the events that occurred in the “prison study” that he conducted in 1971 (Zimbardo, n.d.).
In that study, Zimbardo and his colleagues set up a mock prison. They selected 23 student volunteers and divided them into two groups. One group was chosen to be the “prisoners.” They were picked up at their homes by actual police officers, “arrested,” and brought to the prison to be guarded by the other group of students—the “guards.” The two groups were placed in a setting that was designed to look like a real prison, and the role-play began.
THE ZIMBARDO PRISON STUDY
Watch this video about Zimbardo’s study.
The study was expected to run for two weeks. However, on the second day, the prisoners tried to rebel against the guards. The guards quickly moved to stop the rebellion by using both psychological punishment and physical abuse. In the ensuing days, the guards denied the prisoners food, water, and sleep; shot them with fire-extinguisher spray; threw their blankets into the dirt; and stripped them naked. On the fifth night the experimenters witnessed the guards putting bags over the prisoners’ heads, chaining their legs, and marching them around. At this point the researchers stopped the experiment early. The conclusions of Zimbardo’s research were clear: People may be so profoundly influenced by their social situation that they become coldhearted jail masters who torture their victims.
Zimbardo’s research can help us understand the events that occurred at Abu Ghraib. Zimbardo acted as an expert witness in the trial of Staff Sergeant Ivan “Chip” Frederick, who was sentenced to eight years in prison for his role in the abuse at Abu Ghraib. Frederick was the Army reservist who was put in charge of the night shift at Tier 1A, where the detainees were abused. During his trial, Frederick said, “What I did was wrong, and I don’t understand why I did it.” Zimbardo believes that Frederick acted exactly like the students in the prison study did. He worked in a prison that was overcrowded, filthy, and dangerous, and where he was expected to maintain control over the Iraqi prisoners—in short, the situation he found himself in was very similar to that of Zimbardo’s prison study.
In a recent interview, Zimbardo argued (you can tell that he is a social psychologist) that “human behavior is more influenced by things outside of us than inside.” He believes that, despite our moral and religious beliefs and despite the inherent goodness of people, there are times when external circumstances can overwhelm us, and we do things we never thought we were capable of doing. He argued that “if you’re not aware that this can happen, you can be seduced by evil. We need inoculations against our own potential for evil. We have to acknowledge it. Then we can change it” (Driefus, 2007).
you may wonder whether the extreme behavior of the guards and prisoners in Zimbardo’s prison study was unique to the particular social context that he created. Recent research by Stephen Reicher and Alex Haslam (2006) suggests that this is indeed the case. In their research, they recreated Zimbardo’s prison experiment while making some small, but important, changes. For one, the prisoners were not “arrested” before the study began, and the setup of the jail was less realistic. Furthermore, the researchers in this experiment told the “guards” and the “prisoners” that the groups were arbitrary and could change over time (that is, that some prisoners might be able to be promoted to guards). The results of this study were entirely different than those found by Zimbardo. Although this study was also stopped early, this was more because the guards felt uncomfortable in their superior position than because the prisoners were being abused.
This “prison” simply did not feel like a real prison to the participants, and as a result they did not take on the roles they were assigned. Again, the conclusions are clear—the specifics of the social situation, more than the people themselves, are often the most important determinants of behavior. ■ | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/22%3A_Obedience_Power_and_Leadership/22.01%3A_Introduction.txt |
One of the most influential theories of power was developed by Bertram Raven and John French (French & Raven, 1959; Raven, 1992). Raven identified five different types of power— reward power, coercive power, legitimate power, referent power, and expert power (shown in Table \(1\)), arguing that each type of power involves a different type of social influence and that the different types vary in terms of whether their use will create public conformity or private acceptance. Understanding the types of power is important because it allows us to see more clearly the many ways that people can influence others. Let’s consider these five types of power, beginning with those that are most likely to produce public conformity only and moving on to those that are more likely to produce private acceptance.
Table \(1\): Types of Power
Note. French and Raven (1959) proposed five types of power, which differ in their likelihood of producing public conformity or private acceptance.
Reward power
The ability to distribute positive or negative rewards
Coercive power
The ability to dispense punishments
Legitimate power
Authority that comes from a belief on the part of those being influenced that the person has a legitimate right to demand conformity.
Referent power
Influence based on identification with, attraction to, or respect for the power-holder
Expert power
Power that comes from other’s beliefs that the power-holder possesses superior skills and abilities
Reward Power
Reward power occurs when one person is able to influence others by providing them with positive outcomes. Bosses have reward power over employees because they are able to increase employees’ salary and job benefits, and teachers have reward power over students because they can assign high marks. The variety of rewards that can be used by the powerful is almost endless and includes verbal praise or approval, the awarding of status or prestige, and even direct financial payment.
The ability to wield reward power over those we want to influence is contingent on the needs of the person being influenced. Power is greater when the person being influenced has a strong desire to obtain the reward, and power is weaker when the individual does not need the reward. A boss will have more influence on an employee who has no other job prospects than on one who is being sought after by other corporations, and expensive presents will be more effective in persuading those who cannot buy the items with their own money. Because the change in behavior that results from reward power is driven by the reward itself, its use is usually more likely to produce public conformity than private acceptance.
Coercive Power
Coercive power is power that is based on the ability to create negative outcomes for others, for instance by bullying, intimidating, or otherwise punishing. Bosses have coercive power over employees if they are able (and willing) to punish employees by reducing their salary, demoting them to a lower position, embarrassing them, or firing them. And friends can coerce each other through teasing, humiliation, and ostracism. As with reward power, coercive power is effective only when the person being influenced is dependent upon the power-holder. Furthermore, coercion works only if the person being influenced does not leave the group entirely—people who are punished too much are likely to look for other situations that provide more positive outcomes.
In many cases power-holders use reward and coercive power at the same time—for instance, by increasing salaries as a result of positive performance but also threatening to reduce them if the performance drops. Because the use of coercion has such negative consequences, authorities are generally more likely to use reward than coercive power (Molm, 1997). Coercion is usually more difficult to use, since it often requires energy to keep the person from avoiding the punishment by leaving the situation altogether. And coercive power is less desirable for both the power-holder and the person being influenced because it creates an environment of negative feelings and distrust that is likely to make interactions difficult, undermine satisfaction, and lead to retaliation against the power-holder (Tepper et al., 2009).
As with reward power, coercive power is more likely to produce public conformity than private acceptance. Further- more, in both cases the effective use of the power requires that the power-holder continually monitor the behavior of the target to be sure that he or she is complying. This monitoring may itself lead to a sense of mistrust between the two individuals in the relationship. The power-holder feels (perhaps unjustly) that the target is only complying due to the monitoring, whereas the target feels (again perhaps unjustly) that the power-holder does not trust him or her.
Legitimate Power
Whereas reward and coercive power are likely to produce conformity, other types of power, which are not so highly focused around reward and punishment, are more likely to create changes in attitudes as well as behavior (private acceptance). In many ways, then, these sources of power are stronger because they produce real belief change. Legitimate power is power vested in those who are appointed or elected to positions of authority, such as teachers, politicians, policemen, and judges, and their power is successful because members of the group accept it as appropriate. We accept that governments can levy taxes and that judges can decide the outcomes of court cases because we see these groups and individuals as valid parts of our society. Individuals with legitimate power can exert substantial influence on their followers. Those with legitimate power may not only create changes in the behavior of others but also have the power to create and change the social norms of the group.
In some cases legitimate power is given to the authority figure as a result of laws or elections, or as part of the norms, traditions, and values of the society. The power that the experimenter had over the research participants in Milgram’s study on obedience seems to have been primarily the result of his legitimate power as a respected scientist at an important university. In other cases, legitimate power comes more informally, as a result of being a respected group member. People who contribute to the group process and follow group norms gain status within the group and therefore earn legitimate power.
In some cases, legitimate power can even be used successfully by those who do not seem to have much power. You may recall that after Hurricane Katrina hit the city of New Orleans in 2005, the people there demanded that the federal government help them rebuild the city. Although these people did not have much reward or coercive power, they were nevertheless perceived as good and respected citizens of the United States. Many U.S. citizens tend to believe that people who do not have as much as others (for instance, those who are very poor) should be treated fairly and that these people may legitimately demand resources from those who have more. This might not always work, but to the extent that it does it represents a type of legitimate power—power that comes from a belief in the appropriateness or obligation to respond to the requests of others with legitimate standing.
RESEARCH FOCUS
Does Power Corrupt?
Having power provides some benefits for those who have it. In comparison to those with less power, people who have more power over others are more confident and more attuned to potential opportunities in their environment (Anderson & Berdahl, 2002). They are also more likely than are people with less power to take action to meet their goals (Anderson & Galinsky, 2006; Galinsky et al., 2003).
Despite these advantages of having power, a little power goes a long way and having too much can be dangerous, for both the targets of the power and the powerholder himself or herself. In an experiment by David kipnis (1972), college students played the role of “supervisors” who were supposedly working on a task with other students (the “workers”). According to random assignment to experimental conditions, one half of the supervisors were able to influence the workers through legitimate power only, by sending them messages attempting to persuade them to work harder. The other half of the supervisors were given increased power. In addition to being able to persuade the workers to increase their out- put through the messages, they were also given both reward power (the ability to give small monetary rewards) and coercive power (the ability to take away earlier rewards).
Although the workers (who were actually preprogrammed) performed equally well in both conditions, the participants who were given more power took advantage of it by more frequently contacting the workers and more frequently threatening them. The students in this condition relied almost exclusively on coercive power rather than attempting to use their legitimate power to develop positive relations with the subordinates.
Although it did not increase the workers’ performance, having the extra power had a negative effect on the powerholders’ images of the workers. At the end of the study, the supervisors who had been given extra power rated the workers more negatively, were less interested in meeting them, and felt that the only reason the workers did well was to obtain the rewards.
The conclusion of these researchers is clear: Having power may lead people to use it, even though it may not be necessary, which may then lead them to believe that their subordinates are performing only because of the threats. Although using excess power may be successful in the short run, power that is based exclusively on reward and coercion is not likely to produce a positive environment for either the power-holder or the subordinate. People with power may also be more likely to stereotype people with less power than they have (Dépret & Fiske, 1999) and may be less likely to help other people who are in need (van kleef et al.,2008).
Although this research suggests that people may use power when it is available to them, other research has found that this is not equally true for all people—still another case of a person-situation interaction. Serena Chen and her colleagues (2001) found that students who had been classified as more self-oriented (in the sense that they considered relationships in terms of what they could and should get out of them for themselves) were more likely to misuse their power, whereas students who were classified as other-oriented were more likely to use their power to help others. ■
Referent Power
People with referent power have an ability to influence others because they can lead those others to identify with them. In this case, the person who provides the influence is (a) a member of an important reference group—someone we personally admire and attempt to emulate; (b) a charismatic, dynamic, and persuasive leader; or (c) a person who is particularly attractive or famous (Heath et al., 1994; Henrich & Gil-White, 2001; Kamins, 1989; Wilson & Sherrell, 1993). A young child who mimics the opinions or behaviors of an older sibling or a famous baseball player, or a churchgoer who follows the advice of a respected church leader, is influenced by referent power. Referent power generally produces private acceptance rather than public conformity (Kelman, 1961).
The influence brought on by referent power may occur in a passive sense because the person being emulated does not necessarily attempt to influence others, and the person who is being influenced may not even realize that the influence is occurring. In other cases, however, the person with referent power (such as the leader of a cult) may make full use of his or her status as the target of identification or respect to produce change. In either case, referent power is a particularly strong source of influence because it is likely to result in the acceptance of the opinions of the important other.
Expert Power
French and Raven’s final source of power is expert power. Experts have knowledge or information, and conforming to those whom we perceive to be experts is useful for making decisions about issues for which we have insufficient expertise. Expert power thus represents a type of informational influence based on the fundamental desire to obtain valid and accurate information, and where the outcome is likely to be private acceptance. Conformity to the beliefs or instructions of doctors, teachers, lawyers, and computer experts is an example of expert influence; we assume that these individuals have valid information about their areas of expertise, and we accept their opinions based on this perceived expertise (particularly if their advice seems to be successful in solving problems). Indeed, one method of increasing one’s power is to become an expert in a domain. Expert power is increased for those who possess more information about a relevant topic than others do because the others must turn to this individual to gain the information. You can see, then, that if you want to influence others, it can be useful to gain as much information about the topic as you can. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/22%3A_Obedience_Power_and_Leadership/22.03%3A_Types_Of_Power.txt |
One type of person who has power over others, in the sense that they are able to influence them, is leaders. Leaders are in a position in which they can exert leadership, which is the ability to direct or inspire others to achieve goals (Chemers, 2001; Hogg, 2010). Leaders have many different influence techniques at their disposal. In some cases, they may give commands and enforce them with reward or coercive power, resulting in public conformity with the commands. In other cases, they may rely on well-reasoned technical arguments or inspirational appeals, making use of legitimate, referent, or expert power, with the goal of creating private acceptance and leading their followers to achieve.
Leadership is a classic example of the combined effects of the person and the social situation. Let’s consider first the person part of the equation and then turn to consider how the person and the social situation work together to create effective leadership.
Personality and Leadership
One approach to understanding leadership is to focus on person variables. Personality theories of leadership are explanations of leadership based on the idea that some people are simply “natural leaders” because they possess personality characteristics that make them effective (Zaccaro, 2007).
One personality variable that is associated with effective leadership is intelligence. Being intelligent improves leader- ship, as long as the leader is able to communicate in a way that is easily understood by his or her followers (Simonton, 1994, 1995). Other research has found that a leader’s social skills, such as the ability to accurately perceive the needs and goals of the group members, are also important to effective leadership. People who are more sociable, and therefore better able to communicate with others, tend to make good leaders (Kenny & Zaccaro, 1983; Sorrentino & Boutillier, 1975). Other variables that relate to leadership effectiveness include verbal skills, creativity, self-confidence, emotional stability, conscientiousness, and agreeableness (Cronshaw & Lord, 1987; Judge et al., 2002; Yukl, 2002). And of course the individual’s skills at the task at hand are important. Leaders who have expertise in the area of their leadership will be more effective than those who do not.
Because so many characteristics seem to be related to leader skills, some researchers have attempted to account for leadership not in terms of individual traits but in terms of a package of traits that successful leaders seem to have. Some have considered this in terms of charisma (Beyer, 1999; Conger & Kanungo, 1998). Charismatic leaders are leaders who are enthusiastic, committed, and self-confident; who tend to talk about the importance of group goals at a broad level; and who make personal sacrifices for the group. Charismatic leaders express views that support and validate existing group norms but that also contain a vision of what the group could or should be. Charismatic leaders use their referent power to motivate, uplift, and inspire others. And research has found a positive relationship between a leader’s charisma and effective leadership performance (Simonton, 1988).
Another trait-based approach to leadership is based on the idea that leaders take either transactional or transformational leadership styles with their subordinates (Avolio & Yammarino, 2003; Podsakoff et al., 1990). Transactional leaders are the more regular leaders who work with their subordinates to help them understand what is required of them and to get the job done. Transformational leaders, on the other hand, are more like charismatic leaders—they have a vision of where the group is going and attempt to stimulate and inspire their workers to move beyond their present status and to create a new and better future. Transformational leaders are those who can reconfigure or transform the group’s norms (Reicher & Hopkins, 2003).
the Person and the Situation
Despite the fact that there appear to be at least some personality traits that relate to leadership ability, the most important approaches to understanding leadership take into consideration both the personality characteristics of the leader and the situation in which the leader is operating. In some cases, the situation itself is important. For instance, you might remember that President George W. Bush’s ratings as a leader increased dramatically after the September 11, 2001, terrorist attacks on the World Trade Center. This is a classic example of how a situation can influence the perceptions of a leader’s skill. In other cases, however, both the situation and the per- son are critical.
One well-known person-situation approach to under- standing leadership effectiveness was developed by Fred Fiedler and his colleagues (Ayman et al., 1995). The contingency model of leadership effectiveness is a model of leadership effectiveness that focuses on both person variables and situational variables. Fiedler conceptualized the leadership style of the individual as a relatively stable personality variable and measured it by having people consider all the people they had ever worked with and describe the person that they least liked to work with (their least preferred coworker). Those who indicated that they only somewhat disliked their least preferred coworker are relationship-oriented types of people, who are motivated to have close personal relationships with others. However, those who indicated that they did not like this coworker very much were classified as task-oriented types, who are motivated primarily by getting the job done.
In addition to classifying individuals according to their leadership styles, Fiedler also classified the situations in which groups had to perform their tasks, both on the basis of the task itself and on the basis of the leader’s relationship to the group members. Specifically, as shown in Figure \(2\), Fiedler thought that three aspects of the group situation were important:
1. The degree to which the leader already has a good relationship with the group and the support of the group members (leader-member relations)
2. The extent to which the task is structured and unambiguous (task structure)
3. The leader’s level of power or support in the organization (position power)
Furthermore, Fiedler believed that these factors were ordered in terms of their importance, such that leader-member relationships were more important than task structure, which was in turn more important than position power. As a result, he was able to create eight levels of the “situational favorableness” of the group situation, which roughly range from most favorable to least favorable for the leader. The most favorable relationship involves good relationships, a structured task, and strong power for the leader, whereas the least favorable relationship involves poor relationships, an unstructured task, and weak leader power.
The contingency model is interactionist because it pro- poses that individuals with different leadership styles will differ in effectiveness in different group situations. Task-oriented leaders are expected to be most effective in situations in which the group situation is very favorable because this gives the leader the ability to move the group forward, or in situations in which the group situation is very unfavorable and in which the extreme problems of the situation require the leader to engage in decisive action. However, in the situations of moderate favorableness, which occur when there is a lack of support for the leader or when the problem to be solved is very difficult or unclear, the more relationship-oriented leader is expected to be more effective. In short, the contingency model predicts that task-oriented leaders will be most effective either when the group climate is very favorable and thus there is no need to be concerned about the group members’ feelings, or when the group climate is very unfavorable and the task-oriented leader needs to take firm control.
Still another approach to understanding leadership is based on the extent to which a group member embodies the norms of the group. The idea is that people who accept group norms and behave in accordance with them are likely to be seen as particularly good group members and there- fore become leaders (Hogg, 2001; Hogg & van Knippenberg, 2003). Group members who follow group norms are seen as more trustworthy (Dirks & Ferrin, 2002) and are likely to engage in group-oriented behaviors to strengthen their leadership credentials (Platow & van Knippenberg, 2001).
REFERENCES
Anderson, C., & Berdahl, J. L. (2002). The experience of power: Examining the effects of power on approach and inhibition tendencies. Journal of Personality and Social Psychology, 83(6), 1362–1377. doi.org/10.1037/0022-3514.83.6.1362
Anderson, C., & Galinsky, A. D. (2006). Power, optimism, and risk-taking. European Journal of Social Psychology, 36(4), 511–536. doi. org/10.1002/ejsp.324
Avolio, B. J., & Yammarino, F. J. (2003). Transformational and charismatic leadership: The road ahead. Elsevier Press.
Ayman, R., Chemers, M. M., & Fiedler, F. (1995). The contingency model of leadership effectiveness: Its level of analysis. Leadership Quarterly, 6(2), 147–167. https://doi.org/10.1016/1048-9843(95)90032-2
Beyer, J. M. (1999). Taming and promoting charisma to change organizations. Leadership Quarterly, 10(2), 307–330. doi. org/10.1016/S1048-9843(99)00019-3
Blass, T. (1999). The Milgram paradigm after 35 years: Some things we now know about obedience to authority. Journal of Applied Social Psychology, 29(5), 955–978. doi.org/10.1111/j.1559-1816.1999. tb00134.x
Borge, C. (2007, April 26). Basic instincts: The science of evil. ABC News. https://abcnews.go.com/Primetime/sto...2765416&page=1
Chemers, M. M. (2001). Leadership effectiveness: An integrative review. In M. A. Hogg & R. S. Tindale (Eds.), Blackwell handbook of social psychology: Group processes (pp. 376–399). Blackwell.
Chen, S., Lee-Chai, A. Y., & Bargh, J. A. (2001). Relationship orientation as a moderator of the effects of social power. Journal of Personality and Social Psychology, 80(2), 173–187. doi.org/10.1037/ 0022-3514.80.2.173
Conger, J. A., & Kanungo, R. N. (1998). Charismatic leadership in organizations. Sage.
Cronshaw, S. F., & Lord, R. G. (1987). Effects of categorization, attribution, and encoding processes on leadership perceptions. Journal of Applied Psychology, 72, 97–106. doi.org/10.1037/ 0021-9010.72.1.97
Dépret, E., & Fiske, S. T. (1999). Perceiving the powerful: Intriguing individuals versus threatening groups. Journal of Experimental Social Psychology, 35(5), 461–480. https://doi.org/10.1006/jesp.1999.1380
Dirks, K. T., & Ferrin, D. L. (2002). Trust in leadership: Meta-analytic findings and implications for research and practice. Journal of Applied Psychology, 87(4), 611–628. doi.org/10.1037/ 0021-9010.87.4.611
Driefus, C. (2007, April 3). Finding hope in knowing the universal capacity for evil. New York Times. www.nytimes.com/2007/04/03/ science/03conv.html
Fiske, S. T. (1993). Controlling other people: The impact of power on stereotyping. American Psychologist, 48(6), 621–628. doi. org/10.1037/0003-066X.48.6.621
French, J. R. P., & Raven, B. H. (1959). The bases of social power. In D. Cartwright (Ed.), Studies in social power (pp. 150–167). Institute for Social Research.
Galinsky A. D., Gruenfeld, D. H, & Magee, J. C. (2003). From power to action. Journal of Personality and Social Psychology, 85(3), 453–466. doi.org/10.1037/0022-3514.85.3.453
Heath, T. B., McCarthy, M. S., and Mothersbaugh, D. L. (1994). Spokesperson fame and vividness effects in the context of issue- relevant thinking: The moderating role of competitive setting. Journal of Consumer Research, 20(4), 520–534. doi.org/10.1086/ 209367
Henrich, J., & Gil-White, F. (2001). The evolution of prestige: Freely conferred status as a mechanism for enhancing the benefits of cultural transmission. Evolution and Human Behavior, 22(3), 1–32. https://doi.org/10.1016/S1090-5138(00)00071-4
Hogg, M. A. (2001). A social identity theory of leadership. Personality and Social Psychology Review, 5(3), 184–200. doi.org/10.1207/ S15327957PSPR0503_1
Hogg, M. A. (2010). Influence and leadership. In S. F. Fiske, D. T. Gilbert, & G. Lindzey (Eds.), Handbook of social psychology (Vol. 2, pp. 1166– 1207). Wiley.
Hogg, M. A., & van Knippenberg, D. (2003). Social identity and leadership processes in groups. Advances in Experimental Social Psychology, 35, 1–52. https://doi.org/10.1016/S0065-2601(03)01001-3
Judge, T., Bono, J., Ilies, R., & Gerhardt, M. (2002). Personality and leadership: A qualitative and quantitative review. Journal of Applied Psychology, 87(4), 765–780. doi.org/10.1037/ 0021-9010.87.4.765
Kamins, A. M. (1989). Celebrity and noncelebrity advertising in a two-sided context. Journal of Advertising Research, 29(3), 34–42.
Kelman, H. (1961). Processes of opinion change. Public Opinion Quarterly, 25, 57–78. doi.org/10.1086/266996
Keltner, D., Gruenfeld, D. H., & Anderson, C. (2003). Power, approach, and inhibition. Psychological Review, 110(2), 265–284. doi. org/10.1037/0033-295X.110.2.265
Kenny, D. A., & Zaccaro, S. J. (1983). An estimate of variance due to traits in leadership. Journal of Applied Psychology, 68(4), 678–685. https:// doi.org/10.1037/0021-9010.68.4.678
Kilham, W., & Mann, L. (1974). Level of destructive obedience as a function of transmitter and executant roles in the Milgram obedience paradigm. Journal of Personality and Social Psychology, 29(5), 692–702. doi.org/10.1037/h0036636
Kipnis, D. (1972). Does power corrupt? Journal of Personality and Social
Psychology, 24, 33–41. doi.org/10.1037/h0033390
Meeus, W. H., & Raaijmakers, Q. A. (1986). Administrative obedience:
Carrying out orders to use psychological-administrative violence. European Journal of Social Psychology, 16(4), 311–324. doi. org/10.1002/ejsp.2420160402
Milgram, S. (1974). Obedience to authority: An experimental view. Harper & Row.
Molm, L. D. (1997). Coercive power in social exchange. Cambridge University Press.
Platow, M. J., & van Knippenberg, D. (2001). A social identity analysis of leadership endorsement: The effects of leader in-group prototypicality and distributive intergroup fairness. Personality and Social Psychology Bulletin, 27(11), 1508–1519. doi. org/10.1177/01461672012711011
Podsakoff, P. M., MacKenzie, S. B., Moorman, R. H., & Fetter, R. (1990). Transformational leader behaviors and their effects on followers’ trust in leader, satisfaction, and organizational citizenship behaviors. Leadership Quarterly, 1(2), 107–142. doi. org/10.1016/1048-9843(90)90009-7
Raven, B. H. (1992). A power/interaction model of interpersonal influence: French and Raven thirty years later. Journal of Social Behavior and Personality, 7(2), 217–244.
Reicher, S., & Haslam, S. A. (2006). Rethinking the psychology of tyranny: The BBC prison study. British Journal of Social Psychology, 45(1), 1–40. doi.org/10.1348/014466605X48998
Reicher, S. D., & Hopkins, N. (2003). On the science of the art of leadership. In D. van Knippenberg and M. A. Hogg (Eds.), Leadership and power: Identity processes in groups and organizations
(pp. 197–209). Sage.
Simonton, D. K. (1988). Presidential style: Personality, biography and performance. Journal of Personality and Social Psychology, 55(6), 928–936. doi.org/10.1037/0022-3514.55.6.928
Simonton, D. K. (1994). Greatness: Who makes history and why. Guilford Press.
Simonton, D. K. (1995). Personality and intellectual predictors of leadership. In D. H. Saklofske & M. Zeidner (Eds.), International handbook of personality and intelligence (pp. 739–757). Plenum.
Sorrentino, R. M., & Boutillier, R. G. (1975). The effect of quantity and quality of verbal interaction on ratings of leadership ability. Journal of Experimental Social Psychology, 11(5), 403–411. doi. org/10.1016/0022-1031(75)90044-X
Tepper, B. J., Carr, J. C., Breaux, D. M., Geider, S., Hu, C., & Hua, W. (2009). Abusive supervision, intentions to quit, and employees’ workplace deviance: A power/dependence analysis. Organizational Behavior and Human Decision Processes, 109(2), 156–167. https:// doi.org/10.1016/j.obhdp.2009.03.004
van Kleef, G. A., Oveis, C., van der Löwe, I., LuoKogan, A., Goetz, J., & Keltner, D. (2008). Power, distress, and compassion: Turning a blind eye to the suffering of others. Psychological Science, 19(12), 1315–1322. doi.org/10.1111/j.1467-9280.2008.02241.x
Wilson, E. J., &. Sherrell, D. L. (1993). Source effects in communication and persuasion: A meta-analysis of effect size. Journal of the Academy of Marketing Science, 21(2), 101–112. https://doi.org/ 10.1007/BF02894421
Yukl, G. A. (2002). Leadership in organizations. Prentice Hall.
Zaccaro, S. J. (2007). Trait-based perspectives of leadership. American Psychologist, 62(1), 6–16. https://doi.org/10.1037/0003-066X.62.1.6
Zimbardo, P. G. (n.d.). The story: An overview of the experiment. Stanford Prison Experiment. https://www.prisonexp.org/the-story
22.05: Key Takeways
• Social power can be defined as the ability of a person to create conformity, even when the people being influenced may attempt to resist those changes.
• Milgram’s studies on obedience demonstrated the remarkable extent to which the social situation and people with authority have the power to create obedience.
• One of the most influential theories of power was developed by French and Raven, who identified five different types of power— reward power, coercive power, legitimate power, referent power, and expert power. The types vary in terms of whether their use is more likely to create public conformity or private acceptance.
• Although power can be abused by those who have it, having power also creates some positive outcomes for individuals.
• Leadership is determined by person variables, situational variables, and by the person-situation interaction. The contingency model of leadership effectiveness is an example of the latter.
EXERCISES AND CRITICAL THINKING
1. Write a paragraph that expresses your opinions about the Holocaust or about another example of obedience to authority. Consider how social psychological research on obedience informs your interpretation of the event.
2. Write a report on Dr. Philip Zimbardo’s book, The Lucifer Effect: Understanding How Good People Turn Evil, http://lucifereffect.com.
3. Provide an example of someone who has each of the types of power discussed in this section.
4. Consideraleaderwhomyouhaveworkedwithinthepast.What types of leadership did that person use? Were they effective?
5. Choose a recent event that involved a very effective leader or one that involved a very poor one. Analyze the leadership in terms of the topics discussed in this chapter. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/22%3A_Obedience_Power_and_Leadership/22.04%3A_Leaders_And_Leadership.txt |
Although emotional responses such as guilt, personal distress, and empathy are important determinants of altruism, it is the social situation itself—the people around us when we are deciding whether or not to help—that has perhaps the most important influence on whether and when we help.
Consider the unusual case of the killing of 28-year- old Katherine “Kitty” Genovese in New York City at about 3:00 a.m. on March 13, 1964. Her attacker, Winston Moseley, stabbed and sexually assaulted her within a few yards of her apartment building in the borough of Queens. During the struggle with her assailant, Kitty screamed, “Oh my God! He stabbed me! Please help me!” But no one responded. The struggle continued; Kitty broke free from Moseley, but he caught her again, stabbed her several more times, and eventually killed her.
The murder of Kitty Genovese shocked the nation, in large part because of the (often inaccurate) reporting of it. Stories about the killing, in the New York Times and other papers, indicated that as many as 38 people had overheard the struggle and killing, that none of them had bothered to intervene, and that only one person had even called the police, long after Genovese was dead.
Although these stories about the lack of concern by people in New York City proved to be false (Manning et al., 2007), they nevertheless led many people to think about the variables that might lead people to help or, alternatively, to be insensitive to the needs of others. Was this an instance of the uncaring and selfish nature of human beings? Or was there something about this particular social situation that was critical? It turns out, contrary to your expectations I would imagine, that having many people around during an emergency can in fact be the opposite of helpful—it can reduce the likelihood that anyone at all will help.
23.02: Latane And Darley's Model Of Helping
Two social psychologists, Bibb Latané and John Darley, found themselves particularly interested in, and concerned about, the Kitty Genovese case. As they thought about the stories that they had read about it, they considered the nature of emergency situations such as this one. They realized that emergencies are unusual and that people frequently do not really know what to do when they encounter one. Furthermore, emergencies are potentially dangerous to the helper, and it is therefore probably pretty amazing that anyone helps at all.
To better understand the processes of helping in an emergency, Latané and Darley developed a model of helping that took into consideration the important role of the social situation. Their model, which is shown in Figure \(1\), has been extensively tested in many studies, and there is substantial support for it.
Noticing
Latané and Darley thought the first thing that had to happen in order for people to help is that they had to notice the emergency. This seems pretty obvious, but it turns out that the social situation has a big impact on noticing an emergency. Consider, for instance, people who live in a large city such as New York City, Bangkok, or Beijing. These cities are big, noisy, and crowded—it seems like there are a million things going at once. How could people living in such a city even notice, let alone respond to, the needs of all the people around them? They are simply too overloaded by the stimuli in the city (Milgram, 1970).
Many studies have found that people who live in smaller and less dense rural towns are more likely to help than those who live in large, crowded, urban cities (Amato, 1983; Levine et al., 1994). Although there are a lot of reasons for such differences, just noticing the emergency is critical. When there are more people around, it is less likely that people notice the needs of others.
You may have had an experience that demonstrates the influence of the social situation on noticing. Imagine that you have lived with a family or a roommate for a while, but one night you find yourself alone in your house or apartment because your housemates are staying somewhere else that night. If you are like me, I bet you find yourself hearing sounds that you never heard before—and they might make you pretty nervous. Of course the sounds have always been there, but when other people are around you, you are simply less alert to them. The presence of others can divert our attention from the environment—it’s as if we are unconsciously, and probably quite mistakenly, counting on the others to take care of things for us.
Latané and Darley (1968) wondered if they could examine this phenomenon experimentally. To do so, they simply asked their research participants to complete a questionnaire in a small room. Some of the participants completed the questionnaire alone, while others completed the questionnaire in small groups in which two other participants were also working on questionnaires.
A few minutes after the participants had begun the questionnaires, the experimenters started to release some white smoke into the room through a vent in the wall while they watched through a one-way mirror. The smoke got thicker as time went on, until it filled the room. The experimenters timed how long it took before the first person in the room looked up and noticed the smoke. The people who were working alone noticed the smoke in about 5 seconds, and within 4 minutes most of the participants who were working alone had taken some action. But what about the participants working in groups of three? Although we would certainly expect that having more people around would increase the likelihood that someone would notice the smoke, on aver- age, the first person in the group conditions did not notice the smoke until over 20 seconds had elapsed. And although 75% of the participants who were working alone reported the smoke within 4 minutes, the smoke was reported in only 12% of the three-person groups by that time. In fact, in only three of the eight three-person groups did anyone report the smoke at all, even after it had entirely filled the room!
Interpreting
Even if we notice an emergency, we might not interpret it as one. The problem is that events are frequently ambiguous, and we must interpret them to understand what they really mean. Furthermore, we often don’t see the whole event unfolding, so it is difficult to get a good handle on it. Is a man holding an iPod and running away from a group of pursuers a criminal who needs to be apprehended, or is this just a harmless prank? Were the cries of Kitty Genovese really calls for help, or were they simply an argument with a boyfriend? It’s hard for us to tell when we haven’t seen the whole event (Piliavin et al., 1976). Moreover, because emergencies are rare and because we generally tend to assume that events are benign, we may be likely to treat ambiguous cases as not being emergencies.
The problem is compounded when others are present because when we are unsure how to interpret events we normally look to others to help us understand them (this is informational social influence). However, the people we are looking toward for understanding are themselves unsure how to interpret the situation, and they are looking to us for information at the same time we are looking to them.
When we look to others for information we may assume that they know something we do not know. This is often a mistake, because all the people in the situation are doing the same thing. None of us really know what to think, but at the same time we assume that the others do know. Pluralistic ignorance occurs when people think others in their environment have information they do not have and when they base their judgments on what they think the others are thinking.
Pluralistic ignorance seems to have been occurring in Latané and Darley’s studies, because even when the smoke became really heavy in the room, many people in the group conditions did not react to it. Rather, they looked at each other, and because nobody else in the room seemed very concerned, they each assumed that the others thought that every- thing was all right. You can see the problem—each bystander thinks that other people aren’t acting because they don’t see an emergency. Of course, everyone is confused, but believing that the others know something that they don’t, each observer concludes that help is not required.
Pluralistic ignorance is not restricted to emergency situations (Miller et al., 1988; Suls & Green, 2003). Maybe you have had the following experience: You are in one of your classes and the instructor has just finished a complicated explanation. He is unsure whether the students are up to speed and asks, “Are there any questions?” All the class members are of course completely confused, but when they look at each other, nobody raises a hand in response. So everybody in the class (including the instructor) assumes that everyone under- stands the topic perfectly. This is pluralistic ignorance at its worst—we are all assuming that others know something that we don’t, and so we don’t act. The moral to instructors in this situation is clear: Wait until at least one student asks a question. The moral for students is also clear: Ask your question! Don’t think that you will look stupid for doing so—the other students will probably thank you.
Taking Responsibility
Even if we have noticed the emergency and interpret it as being one, this does not necessarily mean that we will come to the rescue of the other person. We still need to decide that it is our responsibility to do something. The problem is that when we see others around, it is easy to assume that they are going to do something and that we don’t need to do anything. Diffusion of responsibility occurs when we assume that others will take action and therefore we do not take action ourselves. The irony of course is that people are more likely to help when they are the only ones in the situation than they are when there are others around.
Darley and Latané (1968) had study participants work on a communication task in which they were sharing ideas about how to best adjust to college life with other people in different rooms using an intercom. According to random assignment to conditions, each participant believed that he or she was communicating with either one, two, or five other people, who were in either one, two, or five other rooms. Each participant had an initial chance to give his opinions over the inter- com, and on the first round one of the other people (actually a confederate of the experimenter) indicated that he had an “epileptic-like” condition that had made the adjustment process very difficult for him. After a few minutes, the subject heard the experimental confederate say,
I-er-um-I think I-I need-er-if-if could-er-er-somebody er-er-er- er-er-er-er give me a liltle-er-give me a little help here because-er- I-er-I’m-er-er having a-a-a real problcm-er-right now and I-er-if somebody could help me out it would-it would-er-er s-s-sure be-sure be good . . . because there-er-er-a cause I-er-I-uh-I’ve got a-a one of the-er-sei er-er-things coming on and-and-and I could really-er-use some help so if somebody would-er-give me a little h-help-uh-er-er-er-er-er c-could somebody-er-er-help- er-uh-uh-uh (choking sounds). . . . I’m gonna die-er-er-I’m . . . gonna die-er-help-er-er-seizure-er- (chokes, then quiet). (Dar- ley & Latané, 1968, p. 379)
As you can see in Table \(1\), the participants who thought that they were the only ones who knew about the emergency (because they were only working with one other person) left the room quickly to try to get help. In the larger groups, how- ever, participants were less likely to intervene and slower to respond when they did. Only 31% of the participants in the largest groups responded by the end of the 6-minute session.
Table \(1\): Effects of Group Size on likelihood and Speed of Helping
Source: Darley and Latané (1968).
Group Size
verage Helping (%)
Average Time to Help (in seconds)
2 (participant and victim)
85 52
3 (participant, victim, and 1 other)
62 93
6 (participant, victim, and 4 others)
31 166
You can see that the social situation has a powerful influence on helping. We simply don’t help as much when other people are with us.
Perhaps you have noticed diffusion of responsibility if you have participated in an Internet users group where people asked questions of the other users. Did you find that it was easier to get help if you directed your request to a smaller set of users than when you directed it to a larger number of people? Consider the following: In 1998, Larry Froistad, a 29-year-old computer programmer, sent the following message to the members of an Internet self-help group that had about 200 members. “Amanda I murdered because her mother stood between us . . . when she was asleep, I got wickedly drunk, set the house on fire, went to bed, listened to her scream twice, climbed out the window and set about putting on a show of shock and surprise.” Despite this clear online confession to a murder, only three of the 200 newsgroup members reported the confession to the authorities (Markey, 2000).
To study the possibility that this lack of response was due to the presence of others, the researchers (Markey, 2000) con- ducted a field study in which they observed about 5,000 participants in about 400 different chat groups. The experimenters sent a message to the group, from either a male (JakeHarmen) or female (SuzyHarmen) screen name. Help was sought by either asking all the participants in the chat group, “Can any- one tell me how to look at someone’s profile?” or by randomly selecting one participant and asking “[name of selected participant], can you tell me how to look at someone’s profile?” The experimenters recorded the number of people present in the chat room, which ranged from 2 to 19, and then waited to see how long it took before a response was given.
It turned out that the gender of the person requesting help made no difference, but that addressing to a single per- son did. Assistance was received more quickly when help was asked for by specifying a participant’s name (in only about 37 seconds) than when no name was specified (51 seconds). Furthermore, a correlational analysis found that when help was requested without specifying a participant’s name, there was a significant negative correlation between the number of people currently logged on in the group and the time it took to respond to the request.
Garcia et al. (2002) found that the presence of others can promote diffusion of responsibility even if those other people are only imagined. In these studies, the researchers had participants read one of three possible scenarios that manipulated whether participants thought about dining out with 10 friends at a restaurant (group condition) or whether they thought about dining at a restaurant with only one other friend (one-person condition). Participants in the group condition were asked to “Imagine you won a dinner for yourself and 10 of your friends at your favorite restaurant.” Participants in the one-person condition were asked to “Imagine you won a dinner for yourself and a friend at your favorite restaurant.”
After reading one of the scenarios, the participants were then asked to help with another experiment supposedly being conducted in another room. Specifically, they were asked: “How much time are you willing to spend on this other experiment?” At this point, participants checked off one of the following minute intervals: 0 minutes, 2 minutes, 5 minutes, 10 minutes, 15 minutes, 20 minutes, 25 minutes, and 30 minutes.
As you can see in Figure \(2\), simply imagining that they were in a group or alone had a significant effect on helping, such that those who imagined being with only one other per- son volunteered to help for more minutes than did those who imagined being in a larger group.
Implementing Action
The fourth step in the helping model is knowing how to help. Of course, for many of us the ways to best help another person in an emergency are not that clear; we are not professionals and we have little training in how to help in emergencies. People who do have training in how to act in emergencies are more likely to help, whereas the rest of us just don’t know what to do and therefore may simply walk by. On the other hand, today most people have cell phones, and we can do a lot with a quick call. In fact, a phone call made in time might have saved Kitty Genovese’s life. The moral: You might not know exactly what to do, but you may well be able to contact someone else who does.
Latané and Darley’s decision model of bystander intervention has represented an important theoretical framework for helping us understand the role of situational variables on helping. Whether or not we help depends on the outcomes of a series of decisions that involve noticing the event, interpreting the situation as one requiring assistance, deciding to take personal responsibility, and deciding how to help.
Fischer et al. (2011) analyzed data from over 105 studies using over 7,500 participants who had been observed helping (or not helping) in situations in which they were alone or with others. The researchers found significant support for the idea that people helped more when fewer others were present.
And, supporting the important role of interpretation, they also found that the differences were smaller when the need for helping was clear and dangerous and thus required little interpretation. They also found that there were at least some situations (such as when bystanders were able to help provide needed physical assistance) in which having other people around increased helping.
Although the Latané and Darley model was initially developed to understand how people respond in emergencies requiring immediate assistance, aspects of the model have been successfully applied to many other situations, ranging from preventing someone from driving drunk to making a decision about whether to donate a kidney to a relative (Schroeder et al., 1995).
REFERENCES
Amato, P. R. (1983). The helpfulness of urbanites and small town dwellers: A test between two broad theoretical positions. Australian Journal of Psychology, 35(2), 233–243. doi.org/10.1080/ 00049538308255068
Darley, J. M., & Latané, B. (1968). Bystander intervention in emergencies: Diffusion of responsibility. Journal of Personality and Social Psychology, 8(4, Pt. 1), 377–383. https://doi.org/10.1037/h0025589
Fischer, P., Krueger, J. I., Greitemeyer, T., Vogrincic, C., Kastenmüller, A., Frey, D., Heene, M., Wicher, M., & Kainbacher, M. (2011). The bystander-effect: A meta-analytic review on bystander intervention in dangerous and non-dangerous emergencies. Psychological Bulletin, 137(4), 517–537. https://doi.org/10.1037/a0023304
Garcia, S. M., Weaver, K., Moskowitz, G. B., & Darley, J. M. (2002). Crowded minds: The implicit bystander effect. Journal of Personality and Social Psychology, 83(4), 843–853. doi.org/10.1037/ 0022-3514.83.4.843
Latané, B., & Darley, J. M. (1968). Group inhibition of bystander intervention in emergencies. Journal of Personality and Social Psychology, 10(3), 215–221. https://doi.org/10.1037/h0026570
Levine, R. V., Martinez, T. S., Brase, G., & Sorenson, K. (1994). Helping in 36 U.S. cities. Journal of Personality and Social Psychology, 67(1), 69–82. https://doi.org/10.1037/0022-3514.67.1.69
Amato, P. R. (1983). The helpfulness of urbanites and small town dwellers: A test between two broad theoretical positions. Australian Journal of Psychology, 35(2), 233–243. doi.org/10.1080/ 00049538308255068
Darley, J. M., & Latané, B. (1968). Bystander intervention in emergencies: Diffusion of responsibility. Journal of Personality and Social Psychology, 8(4, Pt. 1), 377–383. https://doi.org/10.1037/h0025589
Fischer, P., Krueger, J. I., Greitemeyer, T., Vogrincic, C., Kastenmüller, A., Frey, D., Heene, M., Wicher, M., & Kainbacher, M. (2011). The bystander-effect: A meta-analytic review on bystander intervention in dangerous and non-dangerous emergencies. Psychological Bulletin, 137(4), 517–537. https://doi.org/10.1037/a0023304
Garcia, S. M., Weaver, K., Moskowitz, G. B., & Darley, J. M. (2002). Crowded minds: The implicit bystander effect. Journal of Personality and Social Psychology, 83(4), 843–853. doi.org/10.1037/ 0022-3514.83.4.843
Latané, B., & Darley, J. M. (1968). Group inhibition of bystander intervention in emergencies. Journal of Personality and Social Psychology, 10(3), 215–221. https://doi.org/10.1037/h0026570
Levine, R. V., Martinez, T. S., Brase, G., & Sorenson, K. (1994). Helping in 36 U.S. cities. Journal of Personality and Social Psychology, 67(1), 69–82. https://doi.org/10.1037/0022-3514.67.1.69
Manning, R., Levine, M., & Collins, A. (2007). The Kitty Genovese murder and the social psychology of helping: The parable of the 38 witnesses. American Psychologist, 62(6), 555–562. https://doi.org/ 10.1037/0003-066X.62.6.555
Markey, P. M. (2000). Bystander intervention in computer-mediated communication. Computers in Human Behavior, 16(2), 183–188. https://doi.org/10.1016/S0747-5632(99)00056-4
Milgram, S. (1970). The experience of living in cities. Science, 167(3924), 1461–1468. https://doi.org/10.1126/science.167.3924.1461
Miller, D. T., Turnbull, W., & McFarland, C. (1988). Particularistic and universalistic evaluation in the social comparison process. Journal of Personality and Social Psychology, 55, 908–917. doi. org/10.1037/0022-3514.55.6.908
Piliavin, J. A., Piliavin, I. M., & Broll, L. (1976). Time of arrival at an emergency and likelihood of helping. Personality and Social Psychology Bulletin, 2(3), 273–276. doi.org/10.1177/ 014616727600200314
Schroeder, D. A., Penner, L. A., Dovidio, J. F., & Piliavin, J. A. (1995). The psychology of helping and altruism: Problems and puzzles. McGraw-Hill.
Suls, J., & Green, P. (2003). Pluralistic ignorance and college student perceptions of gender-specific alcohol norms. Health Psychology, 22(5), 479–486. https://doi.org/10.1037/0278-6133.22.5.479
23.03: Key Takeways
• The social situation has an important influence on whether or not we help.
• Latané and Darley’s decision model of bystander intervention has represented an important theoretical framework for helping us understand the role of situational variables on helping. According to the model, whether or not we help depends on the outcomes of a series of decisions that involve noticing the event, interpreting the situation as one requiring assistance, deciding to take personal responsibility, and implementing action.
• Latané and Darley’s model has received substantial empirical support and has been applied not only to helping in emergencies but to other helping situations as well.
EXERCISES AND CRITICAL THINKING
1. Analyze the Kitty Genovese incident in terms of the Latané and Darley model of helping. Which factors do you think were most important in preventing helping?
2. Recount a situation in which you did or did not help, and consider how that decision might have been influenced by the variables specified in Latané and Darley’s model. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/23%3A_How_the_Social_Context_Influences_Helping/23.01%3A_Introduction.txt |
Although we have discussed many of the most important factors, there are still other variables that determine our willingness to help others. These include characteristics of the people who are potentially providing help as well as the ways that others respond to the help they may receive. Let us con- sider them now.
24.02: Some People Are More Helpful Than Others- The Altruistic Personality
We have seen that the social situation is a very strong determinant of whether or not we help. But although the effects of personality may not generally be as strong as those of the social context, person variables do matter. Some people are indeed more helpful than others across a variety of situations, and we say that these people have an altruistic or prosocial personality (Penner et al., 1995). Try evaluating yourself according to the statements posed in Table \(1\) to see how you stand on this variable.
Table \(1\): Measuring the Altruistic Personality
Social Responsibility
No matter what a person has done to us, there is no excuse for taking advantage of them.
It makes sense to be very concerned about how we act when we are sick and feeling miserable.
Empathy
I sometimes try to understand my friends better by imagining how things look from their perspective.
When I’m upset at someone, I usually try to “put myself in their shoes” for a while.
Moral Reasoning
My decisions are usually based on my concern for other people. My decisions are usually based on what is the most fair and just
way to act.
I have helped carry a stranger’s belongings (e.g., books, parcels,
etc.).
I have allowed someone to go ahead of me in a line (e.g.,
supermarket, copying machine, etc.)
The altruistic personality involves both the cognitive and the emotional responses that we experience around others. People with altruistic personalities tend to show empathy and sympathy for others and feel that it is appropriate and right to follow the norm of social responsibility. These people help more people in a wider variety of areas, including providing help to coworkers, donating organs, and volunteering, and also have been found to help more quickly than do people who score lower on these measures (Borman et al., 2001; Penner, 2002). A longitudinal study conducted by Nancy Eisenberg and her colleagues (1999) found that children who were the most helpful when they were measured in their preschool classes also were the most helpful later in childhood and in early adulthood, suggesting that they really were helpful people. People with altruistic personalities seem to be people who have a strong other-concern—they like to be with, to relate to, and to help others.
The altruistic personality is in part heritable. Research has found that identical twins are more similar to each other in both their helping-related emotions (such as empathy) and their actual helping than are fraternal twins, who share only a portion of their genetic makeup (Davis et al., 1994).
24.03: Gender Differences In Helping
You may have already asked yourself an important question about helping: Do men or women help more? And perhaps you have answered this question. For instance, you might have decided that women would be more helpful because they are by and large more attuned to the needs of others. Or perhaps you decided that men would be more helpful because helping involves demonstrating bravery and heroicism and men are more likely to desire to be heroes, or at least to look heroic in the eyes of other people.
In fact, on average there are no big differences between men and women in terms of their helping. For instance, in a survey of altruism sponsored by Independent Sector, an established coalition that studies and encourages volunteer- ing, the percentage of women volunteering (46%) was not significantly different than the percentage of men (42%). Rather, there appears to be a person-by-situation interaction, such that gender differences show up more strongly in some situations than in others. The differences depend not only upon the opportunity to help but also on the type of helping that is required (Becker & Eagly, 2004). In general, men are more likely to help in situations that involve physical strength. If you remember photos and videos taken immediately after the World Trade Center attack in 2001, you’ll probably recall the many images of firefighters and police officers, who were primarily men, engaged in heroic acts of helping.
This does not mean that women are any less helpful—in fact, thousands of women helped during and after the World Trade Center attack by tending to the wounded in hospitals, donating blood, raising money for the families of the victims, and helping with cleanup of the disaster sites. Because women are, on average, more focused on other-concern, they are more likely than men to help in situations that involve long-term nurturance and caring, particularly within close relationships. Women are also more likely than men to engage in community behaviors, such as volunteering in the community or helping families (Becker & Eagly, 2004; Eagly & Becker, 2005). Helping within the family is done in large part by mothers, sisters, wives, and female friends. (You might ask yourself when you last received a thank-you note from a man!)
Although this type of helping might be less likely to be rewarded with newspaper stories and medals, providing social support and helping connect people serves to help us meet the important goal of relating to others and thus helps improve the quality of our lives. And women are not afraid to help in situations that are dangerous. In fact, women have been found to be as likely as men are to engage in dangerous behaviors such as donating a kidney to others (Becker & Eagly, 2004).
SOCIAL PSYCHOLOGY IN THE PUBLIC INTEREST
Are the Religious More Altruistic?
Do you think that religious people are more helpful than are people who are less religious? There are plenty of reasons to think this might be so. After all, every major religion preaches the importance of compassion and helpfulness, and many faith-based organizations help the poor and disadvantaged every year. Religious organizations help pro- vide education, food, clothes, financial support, and other essentials to the needy across the globe.
There is support, based on surveys and questionnaires, that religious people do indeed report being more helpful than the less religious (Penner, 2002). For instance, Morgan (1983) found that people who reported that they prayed more often also said that they were more good, friendly, and cooperative toward others. Furrow et al. (2004) found a significant positive relationship between religiousness and prosocial concerns such as empathy, moral reasoning, and responsibility in urban high school students. And Benson et al. (1989) found that adolescents who said they were more religious were also more likely to have been involved in a volunteer service project in the last year.
Batson and his colleagues (1989) wondered if religious people were actually more likely to help or if they simply indicated that they would be on questionnaires. To test this question, they recruited college students and first asked them to report on their religious beliefs. On the basis of these responses, Batson categorized the students into one of four groups:
• The nonreligious students were those who did not indicate much interest in religion.
• The externally religious students were those who primarily indicated that they used religion for self-concern, such as for feeling more comfortable and being comforted by others, for gaining social status, and for finding support for one’s chosen way of life. The externally religious tended to agree with such statements as “The church is most important as a place to formulate good social relationships” and “What religion offers me most is comfort when sorrows and misfortune strike.”
• The internally religious were those who indicated that they had accepted religion and that it was part of their inner experiences. The internally religious agreed with statements such as “I try hard to carry my religion over into all my other dealings in life” and “Quite often I have been keenly aware of the presence of God or the Divine Being.”
• Finally, people who agreed with such statements as “It might be said that I value my religious doubts and uncertainties” and “Questions are far more central to my religious experience than are answers” were considered to be quest-oriented. These students see religion as a lifelong commitment to getting answers to important moral and religious questions.
Then Batson and his colleagues asked the participants whether or not they would be willing to volunteer their time by helping a woman in need or by walking in a walkathon for a charity. However, in each case Batson also gave one half of the participants a possible excuse for not helping, by informing them that a number of other students had already volunteered to help the woman or that they would have to complete a difficult physical exam before they could be in the walkathon.
The researchers found that the externally religious were not more likely to help overall and were actually less likely to help when there was an easy excuse not to. It seems that the externally religious were not really altruistic at all. The internally religious participants seemed somewhat more altruistic— they helped more when the helping was easy, but they did not continue to help when the task got difficult. However, Batson and his team found that the quest-oriented students were the true altruists—they volunteered to help even when doing so required engaging in some difficult exercise and continued to help even when there was an easy excuse not to.
Although most studies investigating the role of religion on altruism have been correlational, there is also some experimental research showing that activating symbols relating to religion causes increased altruism. Shariff and Norenzayan (2007) showed their research participants religious words such as divine, God, sacred, and prophet and then later asked them to contribute some money to a charity. The participants who had seen the religious words were more likely to donate money to an anonymous recipient than were a control group of people who had been exposed to non- religious control words. However, religion was not the only concept that increased helping. Similar increases in altruism were found when people were shown words related to civil duty, such as civic, jury, court, police, and contract.
In summary, when surveyed, religious people say that they are more helpful than are the nonreligious, but whether they really help when helping conflicts with self-interest seems to depend on what type of religious person they are. People who are religious for personal reasons related to self-concern generally are not more helpful. On the other hand, those who are more quest-oriented—those who really believe that helping is an important part of religious experience—are likely to help even when doing so requires effort. Furthermore, religion is not the only thing that makes us helpful. Being reminded of other social norms, such as our civil responsibility to others, also makes us more helpful. ■ | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/24%3A_Other_Determinants_of_Helping/24.01%3A_Introduction.txt |
We do not help everyone equally—some people just seem to be more worthy of help than others. Our cognitions about people in need matter as do our emotions toward them. For one, our perception of the amount of the need is important. Bickman and Kamzan (1973) found that people were considerably more reluctant to help someone requesting money in a grocery store to buy some cookie dough (a relative luxury item) than they were to help someone requesting money to buy milk (which seems more necessary).
In addition to attempting to determine whether the help is really needed, we also tend to determine whether people are deserving of the help. We tend to provide less help to people who seem to have brought on their problems themselves or who don’t seem to be working very hard to solve them on their own than we do to people who need help as a result of events that seem to be out of their control. Imagine, for instance, that a student in your class asks to borrow your class notes to prepare for an exam. And then imagine if the student said, “I just can’t take good notes—I attend every class, and I really try, but I just can’t do it.” I’m guessing that you might be willing to help this student. On the other hand, imagine that the student said, “Well, I miss class a lot because I don’t feel like coming, and even when I’m here I don’t bother to take notes every day.” I bet you’d be less likely to help this person, who doesn’t seem to be trying very hard.
Supporting this idea, Dooley (1995) had students read scenarios about a person who had been diagnosed with AIDS. Participants who learned that the person had contracted the disease through a blood transfusion felt more empathy and pity for the person, and also expressed a greater desire to help them, than did participants who believed that the disease was caused by unprotected sex or by illicit drug use. One reason we may be particularly likely to help victims of hurricanes and other natural disasters, then, is that we see that these people did not cause their own problems. Those who do argue against helping these victims may well take the opposite position because they believe that the individuals deserved what they got (“they should have known better than to live there.”)
It has been argued that a fundamental difference between individuals who hold politically conservative views and those who hold politically liberal views is how they perceive the necessity or moral responsibility of helping others, and that this relates to how they perceive the causes of people’s out- comes. Consider people who appear to need help because they have inadequate food, shelter, or health care, for example. Liberals tend to attribute these outcomes more externally, blaming them on unjust social practices and societal structures that create inequalities. Because they are likely to believe that the people do not deserve their unfortunate situation, they are likely to favor spending on social programs designed to help these people. Conservatives, on the other hand, are more likely to hold just world beliefs—beliefs that people get what they deserve in life (Lerner, 1980). Conservatives make more internal attributions for negative outcomes, believing that the needs are caused by the lack of effort or ability on the part of the individual. They are therefore less likely than liberals to favor government spending on welfare and other social programs designed to help people (Kluegel & Smith, 1986; Skitka, 1999). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/24%3A_Other_Determinants_of_Helping/24.04%3A_Who_Do_We_Help_Attributions_And_Helping.txt |
To this point in the chapter we have proceeded as if helping is always a good thing—that people need to receive help and that they are appreciative of and thankful to the people who help them. But perhaps this is not always true. We haven’t yet considered the cognitive and affective reactions of the people who are receiving the help. Can you remember a time when somebody tried to help you make a decision or perform a task, but you didn’t really want the help? How did that make you think and feel about yourself? Maybe there are costs involved in receiving help, just as there are in giving it.
Although people who receive help often really need the help and may indeed feel appreciative and grateful to those who help them, receiving help may also have some negative consequences. When we help another person, it indicates that we have enough resources that we can afford to give some of them to the recipient; it also indicates that the recipient is dependent on our goodwill. Thus, helping creates a status disparity in the sense that the helper is seen as having higher status than the person being helped. This inequality makes giving help an indication of high status and power, and receiving help a potentially self-threatening experience for the recipient (Nadler, 2002; Nadler & Halabi, 2006). There are a variety of emotions that help recipients might feel in these cases, including embarrassment and worry that they are, or are seen as, incompetent or dependent (DePaulo et al., 1981; Nadler et al., 1983). Research has found that people frequently respond negatively when they receive help and may in some cases even prefer to endure hardships rather than to seek out help (Nadler, 1991). Receiving help, then, can be a potential blow to our self-esteem.
The negative feelings that we experience when receiving help are likely to be particularly strong when the recipient feels that the implication of the helping is that they are unable to care for themselves. In these cases, the help is perceived as being dependency oriented (Nadler et al., 1983). When the helper takes control of the situation and solves the problem facing the individual, leaving little left for the individual to accomplish on his or her own, the behavior may be seen as indicating that the individual cannot help herself. The potential recipients of help are likely to reject offers of dependency oriented help, refrain from seeking it, and react negatively when it is offered.
Another situation in which people may not appreciate the help they are receiving is when that help comes on the basis of one’s presumed need. For instance Blaine et al. (1995) found that people who imagined that they had been hired for a job because they were disabled experienced lower self-esteem and felt that they were less likely to work hard on the job than those who imagined that they were hired on the basis of their job qualifications. You can see that government programs, such as those based on affirmative action, although likely to be helpful for the people who receive them, may also lead those people to feel dependent on others.
In contrast to dependency-oriented help, autonomy-oriented help is partial and temporary and provides information to the other, for instance, by giving instructions or guidance or providing ideas about how to help oneself. Autonomy-oriented help reflects the helper’s view that, given the appropriate tools, recipients can help themselves (Brickman et al., 1982). Autonomy-oriented help allows help recipients to retain their independence despite their reliance on the more resourceful helper. This type of help is less likely to clash with help recipients’ view of themselves as capable people who can help themselves.
There are also observed gender differences in the willingness to seek help. Boys and men are less likely to ask for help overall, perhaps in part because they feel that asking for help indicates to others that they are less capable of handling their own affairs or that they have low status (Addis & Mahalik, 2003; Mansfield et al., 2003).
In short, when we help others we must be careful that we do it in a way that allows them to maintain their independence and that reminds them that they are still able to help themselves. This type of help will be more easily accepted and more beneficial in the long run. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/24%3A_Other_Determinants_of_Helping/24.05%3A_Reactions_To_Receiving_Help.txt |
Although almost every culture has a social responsibility norm, the strength of those norms varies across cultures. And these differences relate well to what we know about individualism and collectivism. In one study, Miller et al. (1990) found that children and adults in the United States (a Western and therefore individualistic culture) were less likely than children and adults in India (an Eastern and therefore collectivistic culture) to believe that people have an obligation to provide assistance to others. The Indian respondents believed that there was an absolute requirement to help, whereas the Americans offered their help more selectively, even to their friends. Similarly, Baron and Miller (2000) found that Indian students were more likely than U.S. students to view donating bone marrow to save someone’s life as morally required, whereas U.S. students were more likely than Indian students to say that donating was a decision that the potential donor had to make himself or herself.
Perlow and Weeks (2002) found that there were substantial cultural differences in the behavior of software engineers working at similar companies and doing the same type of work in the United States and in India. Engineers at the American site were more focused on exchange and reciprocity—they tended to provide help to others only if they thought those people could be helpful to them in the future. The engineers at the Indian company, on the other hand, were more willing to help anyone who seemed to need help, regardless of the potential for a return. Perlow and Weeks interpreted these differences in terms of different ways of meeting the goal of self-interest. Among the Americans, helping was seen as an unwanted interruption on the time of the individual, and thus helping was not personally beneficial. At the Indian company, however, helping was seen more as an opportunity for improving one’s skills by helping. These results suggest that helping, at least in Western contexts such as the United States, can be increased if it is framed to be perceived as important toward achieving one’s goals.
One important difference between Eastern and Western cultures is that the importance of self-concern (versus other-concern) is higher in the latter. In fact, the strong individualistic norms in cultures such as the United States make it sometimes inappropriate to try to help in cases where we do not have a personal interest. Rebecca Ratner and Dale Miller (2001) had participants read a scenario in which a governmental funding agency was planning to reduce funding for research regarding a disease. The disease was said to affect only women or only men. Then the participants were asked to indicate both whether they were opposed to the reduction in funding and how comfortable they would be in attending a meeting to protest the funding changes.
In terms of their attitudes toward the reduction in funding, there were no significant gender differences. Men thought that the funding should be maintained even when the disease only affected women, and vice versa. However, as you can see in Figure \(1\), when asked how comfortable they would feel attending a meeting protesting the funding decreases, significant differences occurred. The men predicted that they would feel less comfortable attending a meeting to protest the funding reductions when the disease only affected women, and the women predicted that they would feel less comfortable attending a meeting to protest the funding reductions when the disease only affected men.
Ratner and Miller argued that in Western cultures there is a norm of self-interest that influences whether or not we feel that we can be involved in actions designed to help others. In short, people are not expected to volunteer for, or to be involved in, causes that do not affect them personally. It is simply inappropriate to lend help to others unless the person is personally involved in the issue and thus stands to benefit. Indeed, participants in another study by Ratner and Miller reacted more negatively to an individual’s altruistic behaviors when they did not appear consistent with his or her self-interest.
There is still another example of the subtle role of self-interest in helping. Did you ever notice that many people who are looking for contributions to a cause do not ask directly but rather ask that you purchase something from them, allowing them to keep the profit from the sale? Bake sales, car washes, and address sticker and magazine subscription charity campaigns are all examples of this. Of course, it would be more profitable for the charity if people simply gave the same amount of money rather than taking the gift—and perhaps the people who are making the purchases would prefer not to have to buy the product anyway.
Is it possible that people are simply more comfortable making donations in exchange for a product than they are simply giving money to a charity? Research by John Holmes and his colleagues (2002) has supported this idea, finding that people are more likely to help when they can pretend that they are acting in their own self-interest. In one study, Holmes and his team found that students were more likely to donate money to a needy charity when they were offered a small candle in return for their donation than when they were not offered the candle. However, and suggesting that they didn’t really care about the candle that much, when the request was to contribute to a charity that did not seem that needy, contributions were smaller overall but were not greater when the candle was offered than when it was not. Again, it seems that people feel more comfortable being altruistic when they can pretend that they are really helping themselves—not violating the norm of self-interest. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/24%3A_Other_Determinants_of_Helping/24.06%3A_Cultural_Issues_In_Helping.txt |
Now that we have a fundamental understanding of the variables that influence the likelihood that we will help others, let’s spend some time considering how we might use this information in our everyday life to try to become more helpful ourselves and to encourage those around us to do the same. In doing so we will make use of many of the principles of altruism that we have discussed in this chapter.
First, we need to remember that not all helping is based on other-concern—self-concern is important. People help in part because it makes them feel good, and therefore any- thing that we can do to increase the benefits of helping and to decrease the costs of helping would be useful. Consider, for instance, the research of Mark Snyder, who has extensively studied the people who volunteer to help other people who are suffering from AIDS (Snyder & Omoto, 2004; Snyder et al., 2004). To help understand which volunteers were most likely to continue to volunteer over time, Snyder and his colleagues (Omoto & Snyder, 1995) asked the AIDS volunteers to indicate why they volunteered. As you can see in Table \(1\), the researchers found that the people indicated that they volunteered for many different reasons, and these reasons fit well with our assumptions about human nature—they involve both self-concern and other-concern.
Table \(1\): Reasons for volunteering to Help AIDS victims
Values
• Because of my humanitarian obligation to help others
• Because I enjoy helping other people
• Because I consider myself to be a loving and caring person
• Because people should do something about issues that are important to them
• Because of my personal values, convictions, and beliefs
Understanding
• To learn more about how to prevent AIDS
• To learn how to help people with AIDS
• To learn about how people cope with AIDS
• To understand AIDS and what it does to people
Personal Development
• To get to know people who are similar to myself
• To meet new people and make new friends
• To gain experience dealing with emotionally difficult topics
• To challenge myself and test my skills
• To learn about myself and my strengths and weaknesses
Community Concern
• Because of my sense of obligation to the gay community
• Because I consider myself an advocate for gay-related issues
• Because of my concern and worry about the gay community
• To get to know people in the gay community
• To help members of the gay community
Esteem Enhancement
• To make my life more stable
• To escape other pressures and stress in my life (e.g., from work, from home)
• To feel less lonely
• To feel needed
Omoto and Snyder (1995) found that the volunteers were more likely to continue their volunteer work if their reasons for volunteering involved self-related activities, such as under- standing, personal development, or esteem enhancement. The volunteers who felt that they were getting something back from their work were likely to stay involved. In addition, Snyder and his colleagues found that people were more likely to continue volunteering when their existing social support networks were weak. This result suggests that some volunteers were using the volunteer opportunity to help them create better social con- nections (Omoto & Snyder, 1995). On the other hand, the volunteers who reported experiencing negative reactions about their helping from their friends and family members, which made them feel embarrassed, uncomfortable, and stigmatized for helping, were also less likely to continue working as volunteers (Snyder et al., 1999).
These results again show that people will help more if they see it as rewarding. So if you want to get people to help, try to increase the rewards of doing so, for instance by enhancing their mood or by offering incentives. Simple things, such as noticing, praising, and even labeling helpful behavior can be enough. When children are told that they are “kind and helpful children,” they contribute more of their prizes to other children (Grusec et al., 1978). Rewards work for adults too: People were more likely to donate to charity several weeks after they were described by another person as being “gener- ous” and “charitable” people (Kraut, 1973). In short, once we start to think of ourselves as helpful people, self-perception takes over and we continue to help.
The nations and states that have passed Good Samaritan laws realize the importance of self-interest: If people must pay fines or face jail sentences if they don’t help, then they are naturally more likely to help. And the programs in many schools, businesses, and other institutions that encourage students and workers to volunteer by rewarding them for doing so are also effective in increasing volunteering (Clary, Snyder, Ridge, et al., 1998; Clary, Snyder, & Stoukas, 1998).
Helping also occurs in part because of other-concern. We are more likely to help people we like and care about, feel similar to, and experience positive emotions with. Therefore, anything we can do to increase our connections with others will likely increase helping. We must work to encourage ourselves, our friends, and our children to interact with others—to help them meet and accept new people and to instill a sense of community and caring in them. These social connections will make us feel closer to others and increase the likelihood we will help them. We must also work to install the appropriate norms in our children. Kids must be taught not to be selfish and to value the norms of sharing and altruism.
One way to increase our connection with others is to make those people highly salient and personal. Charities and other organizations that seek to promote helping understand this and do the best they can to individualize the people they are asking us to help. When we see a single person suffering, we naturally feel strong emotional responses to that person. And, as we have seen, the emotions that we feel when others are in need are powerful determinants of helping. In fact, Paul Slovic (2007) found that people are simply unable to identify with statistical and abstract descriptions of need because they do not feel emotions for these victims in the same way they do for individuals. They argued that when people seem completely oblivious or numb to the needs of millions of people who are victims of genocide, hurricanes, and other atrocities, it is because the victims are presented as statistics rather than as individual cases. As Joseph Stalin, the Russian dictator who executed millions of Russians, put it, “A single death is a tragedy, a million deaths is a statistic.”
We can also use what we have learned about helping in emergency situations to increase the likelihood of responding. Most importantly, we must remember how strongly pluralistic ignorance can influence the interpretation of events and how quickly responsibility can be diffused among the people present at an emergency. Therefore, in emergency situations we must attempt to counteract pluralistic ignorance and diffusion of responsibility by remembering that others do not necessarily know more than we do. Depend on your own interpretation— don’t simply rely on your assumptions about what others are thinking and don’t just assume that others will do the helping.
We must be sure to follow the steps in Latané and Darley’s model, attempting to increase helping at each stage. We must make the emergency noticeable and clearly an emergency, for instance, by yelling out: “This is an emergency! Please call the police! I need help!” And we must attempt to avoid the diffusion of responsibility, for instance, by designating one individual to help: “You over there in the red shirt, please call 911 now!”
REFERENCES
Addis, M. E., & Mahalik, J. R. (2003). Men, masculinity, and the contexts of help seeking. American Psychologist, 58(1), 5–14. https://doi.org/ 10.1037/0003-066X.58.1.5
Baron, J., & Miller, J. G. (2000). Limiting the scope of moral obligations to help: A cross-cultural investigation. Journal of Cross-Cultural Psychology, 31(6), 703–725. doi.org/10.1177/ 0022022100031006003
Batson, C. D., Oleson, K. C., Weeks, J. L., Healy, S. P., Reeves, P. J., Jennings, P., & Brown, T. (1989). Religious prosocial motivation: Is it altruistic or egoistic? Journal of Personality and Social Psychology, 57(5), 873–884. https://doi.org/10.1037/0022-3514.57.5.873
Becker, S. W., & Eagly, A. H. (2004). The heroism of women and men. American Psychologist, 59(3), 163–178. doi.org/10.1037/ 0003-066X.59.3.163
Benson, P. L., Donahue, M. J., & Erickson, J. A. (Eds.). (1989). Adolescence and religion: A review of the literature from 1970 to 1986. Research in the Social Scientific Study of Religion, 1, 153–181.
Bickman, L., & Kamzan, M. (1973). The effect of race and need on helping behavior. Journal of Social Psychology, 89(1), 73–77. https://doi.org/ 10.1080/00224545.1973.9922569
Blaine, B., Crocker, J., & Major, B. (1995). The unintended negative consequences of sympathy for the stigmatized. Journal of Applied Social Psychology, 25(10), 889–905. doi.org/10.1111/ j.1559-1816.1995.tb02651.x
Borman, W. C., Penner, L. A., Allen, T. D., & Motowidlo, S. J. (2001). Personality predictors of citizenship performance. International Journal of Selection and Assessment, 9(1–2), 52–69. https://doi.org/ 10.1111/1468-2389.00163
Brickman, P., Rabinowitz, V. C., Karuza, J., Jr., Coates, D., Cohn, E., & Kidder, L. (1982). Models of helping and coping. American Psychologist, 37(4), 368–384. doi.org/10.1037/ 0003-066X.37.4.368
Clary, E. G., Snyder, M., Ridge, R. D., Copeland, J., Stukas, A. A., Haugen, J., & Miene, P. (1998). Understanding and assessing the motivations of volunteers: A functional approach. Journal of Personality and Social Psychology, 74(6), 1516–1530. doi.org/10.1037/ 0022-3514.74.6.1516
Clary, E. G., Snyder, M., & Stukas, A. (1998). Service-learning and psychology: Lessons from the psychology of volunteers’ motivations. American Psychological Association.
Davis, M. H., Luce, C., & Kraus, S. J. (1994). The heritability of characteristics associated with dispositional empathy. Journal of Personality, 62(3), 369–391. doi.org/10.1111/j.1467-6494.1994. tb00302.x
DePaulo, B. M., Brown, P. L., Ishii, S., & Fisher, J. D. (1981). Help that works: The effects of aid on subsequent task performance. Journal of Personality and Social Psychology, 41(3), 478–487. https://doi.org/ 10.1037/0022-3514.41.3.478
Dooley, P. A. (1995). Perceptions of the onset controllability of AIDS and helping judgments: An attributional analysis. Journal of Applied Social Psychology, 25(10), 858–869. doi.org/10.1111/ j.1559-1816.1995.tb02649.x
Eagly, A. H., & Becker, S. W. (2005). Comparing the heroism of women and men. American Psychologist, 60(4), 343–344. https://doi.org/ 10.1037/0003-066X.60.4.343
Eisenberg, N., Guthrie, I. K., Murphy, B. C., Shepard, S. A., Cumberland, A., & Carlo, G. (1999). Consistency and development of prosocial dispositions: A longitudinal study. Child Development, 70(6), 1360–1372. doi.org/10.1111/1467-8624.00100
Furrow, J. L., King, P. E., & White, K. (2004). Religion and positive youth development: Identity, meaning, and prosocial concerns. Applied Developmental Science, 8(1), 17–26. doi.org/10.1207/ S1532480XADS0801_3
Grusec, J. E., Kuczynski, L., Rushton, J. P., & Simutis, Z. M. (1978). Modeling, direct instruction, and attributions: Effects on altruism. Developmental Psychology, 14(1), 51–57. doi.org/10.1037/ 0012-1649.14.1.51
Holmes, J. G., Miller, D. T., & Lerner, M. J. (2002). Committing altruism under the cloak of self-interest: The exchange fiction. Journal of Experimental Social Psychology, 38(2), 144–151. https://doi.org/ 10.1006/jesp.2001.1494
Kluegel, J. R., & Smith, E. R. (1986). Beliefs about inequality: Americans’ views of what is and what ought to be. Aldine de Gruyter.
Kraut, R. E. (1973). Effects of social labeling on giving to charity. Journal of Experimental Social Psychology, 9(6), 551–562. https://doi.org/ 10.1016/0022-1031(73)90037-1
Lerner, M. (1980). The belief in a just world: A fundamental delusion. Plenum.
Mansfield, A. K., Addis, M. E., & Mahalik, J. R. (2003). “Why won’t he go to the doctor?”: The psychology of men’s help seeking. International Journal of Men’s Health, 2(2), 93–109. doi.org/10.3149/ jmh.0202.93
Miller, J. G., Bersoff, D. M., & Harwood, R. L. (1990). Perceptions of social responsibilities in India and in the United States: Moral imperatives or personal decisions? Journal of Personality and Social Psychology, 58(1), 33–47. https://doi.org/10.1037/0022-3514.58.1.33
Morgan, S. P. (1983). A research note on religion and morality: Are religious people nice people? Social Forces, 61(3), 683–692. https:// doi.org/10.2307/2578129
Nadler, A. (2002). Inter-group helping relations as power relations: Maintaining or challenging social dominance between groups through helping. Journal of Social Issues, 58(3), 487–502. doi. org/10.1111/1540-4560.00272
Nadler, A. (Ed.). (1991). Help-seeking behavior: Psychological costs and instrumental benefits. Sage.
Nadler, A., & Halabi, S. (2006). Intergroup helping as status relations: Effects of status stability, identification, and type of help on receptivity to high-status group’s help. Journal of Personality and Social Psychology, 91(1), 97–110. doi.org/10.1037/ 0022-3514.91.1.97
Nadler, A., Fisher, J. D., & Itzhak, S. B. (1983). With a little help from my friend: Effect of single or multiple act aid as a function of donor and task characteristics. Journal of Personality and Social Psychology, 44(2), 310–321. https://doi.org/10.1037/0022-3514.44.2.310
Omoto, A. M., & Snyder, M. (1995). Sustained helping without obligation: Motivation, longevity of service, and perceived attitude change among AIDS volunteers. Journal of Personality and Social Psychology, 68(4), 671–686. doi.org/10.1037/ 0022-3514.68.4.671
Penner, L. A. (2002). Dispositional and organizational influences on sustained volunteerism: An interactionist perspective. Journal of Social Issues, 58(3), 447–467. doi.org/10.1111/ 1540-4560.00270
Penner, L. A., Fritzsche, B. A., Craiger, J. P., & Freifeld, T. S. (1995). Measuring the prosocial personality. In J. Butcher & C. Speigelberger (Eds.), Advances in personality assessment (Vol. 10, pp. 147–163). Lawrence Erlbaum.
Perlow, L., & Weeks, J. (2002). Who’s helping whom? Layers of culture and workplace behavior. Journal of Organizational Behavior, 23(4), 345–361. doi.org/10.1002/job.150
Ratner, R. K., & Miller, D. T. (2001). The norm of self-interest and its effects on social action. Journal of Personality and Social Psychology, 81(1), 5–16. https://doi.org/10.1037/0022-3514.81.1.5
Shariff, A. F., & Norenzayan, A. (2007). God is watching you: Priming God concepts increases prosocial behavior in an anonymous economic game. Psychological Science, 18(9), 803–809. doi. org/10.1111/j.1467-9280.2007.01983.x
Skitka, L. J. (1999). Ideological and attributional boundaries on public compassion: Reactions to individuals and communities affected by a natural disaster. Personality and Social Psychology Bulletin, 25(7), 793–808. https://doi.org/10.1177/0146167299025007003
Slovic, P. (2007). “If I look at the mass I will never act”: Psychic numbing and genocide. Judgment and Decision Making, 2(2), 79–95.
Snyder, M., & Omoto, A. M. (Eds.). (2004). Volunteers and volunteer organizations: Theoretical perspectives and practical concerns. Jossey-Bass.
Snyder, M., Omoto, A. M., & Crain, A. L. (1999). Punished for their good deeds: Stigmatization of AIDS volunteers. American Behavioral Scientist, 42(7), 1175–1192. doi.org/10.1177/ 0002764299042007009
Snyder, M., Omoto, A. M., & Lindsay, J. J. (Eds.). (2004). Sacrificing time and effort for the good of others: The benefits and costs of volunteerism. Guilford Press.
24.08: Key Takeways
• Some people—for instance, those with altruistic personalities—are more helpful than others.
• Gender differences in helping depend on the type of helping that is required. Men are more likely to help in situations that involve physical strength, whereas women are more likely to help in situations that involve long-term nurturance and caring, particularly within close relationships.
• Our perception of the amount of the need is important. We tend to provide less help to people who seem to have brought on their own problems or who don’t seem to be working very hard to solve them on their own.
• In some cases, helping can create negative consequences. Dependency-oriented help may make the helped feel negative emotions, such as embarrassment and worry that they are seen as incompetent or dependent. Autonomy-oriented help is more easily accepted and will be more beneficial in the long run.
• Norms about helping vary across cultures, for instance, between Eastern and Western cultures.
• We can increase helping by using our theoretical knowledge about the factors that produce it. Our strategies can be based on using both self-concern and other-concern.
EXERCISES AND CRITICAL THINKING
1. Consider your own personality and compare it to that of some other people you know. Do you have an altruistic personality? Do you know people who seem to have one?
2. Imagine that you knew someone who was ill and needed help. How would you frame your help to make him or her willing to accept it?
3. Assume for a moment that you were in charge of creating an advertising campaign designed to increase people’s altruism. On the basis of your reading, what approaches might you take? | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/24%3A_Other_Determinants_of_Helping/24.07%3A_Increasing_Helping.txt |
This module discusses gender and its related concepts, including sex, gender roles, gender identity, sexual orientation, and sexism. In addition, this module includes a discussion of differences that exist between males and females and how these real gender differences compare to the stereotypes society holds about gender differences. In fact, there are significantly fewer real gender differences than one would expect relative to the large number of stereotypes about gender differences. This module then discusses theories of how gender roles develop and how they contribute to strong expectations for gender differences. Finally, the module concludes with a discussion of some of the consequences of relying on and expecting gender differences, such as gender discrimination, sexual harassment, and ambivalent sexism.
Before we discuss gender in detail, it is important to understand what gender actually is. The terms sex and gender are frequently used interchangeably, though they have different meanings. In this context, sex refers to the biological category of male or female, as defined by physical differences in genetic composition and in reproductive anatomy and function. On the other hand, gender refers to the cultural, social, and psychological meanings that are associated with masculinity and femininity (Wood & Eagly, 2002). You can think of “male” and “female” as distinct categories of sex (a person is typically born a male or a female), but “masculine” and “feminine” as continuums associated with gender (every- one has a certain degree of masculine and feminine traits and qualities).
Beyond sex and gender, there are a number of related terms that are also often misunderstood. Gender roles are the behaviors, attitudes, and personality traits that are designated as either masculine or feminine in a given culture. It is common to think of gender roles in terms of gender stereotypes, or the beliefs and expectations people hold about the typical characteristics, preferences, and behaviors of men and women. A person’s gender identity refers to their psychological sense of being male or female. In contrast, a person’s sexual orientation is the direction of their emotional and erotic attraction toward members of the opposite sex, the same sex, or both sexes. These are important distinctions, and although we will not discuss each of these terms in detail, it is import- ant to recognize that sex, gender, gender identity, and sexual orientation do not always correspond with one another. A person can be biologically male but have a female gender identity while being attracted to women, or any other combination of identities and orientations.
THE (TRANS)GENDER-BATHROOM DEBATE
In recent years, there has been much cultural and political debate over the right of transgender individuals to use the public bathroom of their choosing. This debate made major headlines in 2016 when North Carolina passed the Public Facilities Privacy & Security Act (commonly called House Bill 2 or HB2). This law required individuals to use the restroom that corresponded with their birth sex, meaning that transgender individuals could not use the bathroom that matched their gender identity. This law and the similar “bathroom bills” proposed by other states were met with widespread controversy, with opponents arguing that they were discriminatory and perpetuated inequality (Barnett et al., 2018). HB2 has since been repealed, but many states still do not protect the rights of transgender individuals to use their restrooms of choice. ■ | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/25%3A_Gender/25.01%3A_Introduction.txt |
Historically, the terms gender and sex have been used inter- changeably. Because of this, gender is often viewed as a binary—a person is either male or female—and it is assumed that a person’s gender matches their biological sex. This is not always the case, however, and more recent research has separated these two terms. While the majority of people do identify with the gender that matches their biological sex (cisgender), an estimated 0.6% of the population identify with a gender that does not match their biological sex (trans- gender) (Flores et al., 2016). For example, an individual who is biologically male may identify as female, or vice versa.
In addition to separating gender and sex, recent research has also begun to conceptualize gender in ways beyond the gender binary. Genderqueer and gender nonbinary are umbrella terms used to describe a wide range of individuals who do not identify with and/or conform to the gender binary. These terms encompass a variety of more specific labels individuals may use to describe themselves. Some common labels are genderfluid, agender, and bigender. An individual who is genderfluid may identify as male, female, both, or neither at different times and in different circumstances. An individual who is agender may have no gender or describe themselves as having a neutral gender, while bigender individuals identify as two genders.
It is important to remember that sex and gender do not always match and that gender is not always binary; however, a large majority of prior research examining gender has not made these distinctions. Thus, the following sections will discuss gender as a binary.
25.03: Gender Differences
Differences between males and females can be based on (a) actual gender differences (i.e., men and women are actually different in some abilities), (b) gender roles (i.e., differences in how men and women are supposed to act), or (c) gender stereotypes (i.e., differences in how we think men and women are). Sometimes gender stereotypes and gender roles reflect actual gender differences, but sometimes they do not.
What are actual gender differences? In terms of language and language skills, girls develop language skills earlier and know more words than boys; this does not, however, trans- late into long-term differences. Girls are also more likely than boys to offer praise, to agree with the person they’re talking to, and to elaborate on the other person’s comments; boys, in contrast, are more likely than girls to assert their opinion and offer criticisms (Leaper & Smith, 2004). In terms of temperament, boys are slightly less able to suppress inappropriate responses and slightly more likely to blurt things out than girls (Else-Quest et al., 2006).
With respect to aggression, boys exhibit higher rates of unprovoked physical aggression than girls, but no difference in provoked aggression (Hyde, 2005). Some of the biggest differences involve the play styles of children. Boys frequently play organized rough-and-tumble games in large groups, while girls often play less physical activities in much smaller groups (Maccoby, 1998). There are also differences in the rates of depression, with girls much more likely than boys to be depressed after puberty. After puberty, girls are also more likely to be unhappy with their bodies than boys.
However, there is considerable variability between individual males and individual females. Also, even when there are mean level differences, the actual size of most of these differences is quite small. This means knowing someone’s gender does not help much in predicting his or her actual traits. For example, in terms of activity level, boys are considered more active than girls. However, 42% of girls are more active than the average boy (but so are 50% of boys; see Figure \(2\) for a depiction of this phenomenon in a comparison of male and female self-esteem). Furthermore, many gender differences do not reflect innate differences, but instead reflect differences in specific experiences and socialization. For example, one presumed gender difference is that boys show better spatial abilities than girls. However, Tzuriel and Egozi (2010) gave girls the chance to practice their spatial skills (by imagining a line drawing was different shapes) and discovered that, with practice, this gender difference completely disappeared.
Many domains we assume differ across genders are really based on gender stereotypes and not actual differences. Based on large metanalyses, the analyses of thousands of studies across more than one million people, research has shown: Girls are not more fearful, shy, or scared of new things than boys; boys are not more angry than girls; girls are not more emotional than boys; boys do not perform better at math than girls; and girls are not more talkative than boys (Hyde, 2005).
In the following sections, we’ll investigate gender roles, the part they play in creating these stereotypes, and how they can affect the development of real gender differences. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/25%3A_Gender/25.02%3A_Defining_Gender.txt |
As mentioned earlier, gender roles are well-established social constructions that may change from culture to culture and over time. In American culture, we commonly think of gender roles in terms of gender stereotypes, or the beliefs and expectations people hold about the typical characteristics, preferences, and behaviors of men and women.
By the time we are adults, our gender roles are a stable part of our personalities, and we usually hold many gender stereotypes. When do children start to learn about gender? Very early (see Figure \(1\)). By their first birthday, children can distinguish faces by gender. By their second birthday, they can label others’ gender and even sort objects into gendetyped categories. By the third birthday, children can consistently identify their own gender (see Martin et al., 2002, for a review). At this age, children believe sex is determined by external attributes, not biological attributes. Between 3 and 6 years of age, children learn that gender is constant and can’t change simply by changing external attributes, having developed gender constancy. During this period, children also develop strong and rigid gender stereotypes. Stereotypes can refer to play (e.g., boys play with trucks, and girls play with dolls), traits (e.g., boys are strong, and girls like to cry), and occupations (e.g., men are doctors and women are nurses). These stereotypes stay rigid until children reach about age 8 or 9. Then they develop cognitive abilities that allow them to be more flexible in their thinking about others.
How do our gender roles and gender stereotypes develop and become so strong? Many of our gender stereotypes are so strong because we emphasize gender so much in culture (Bigler & Liben, 2007). For example, males and females are treated differently before they are even born. When someone learns of a new pregnancy, the first question asked is “Is it a boy or a girl?” Immediately upon hearing the answer, judgments are made about the child: Boys will be rough and like blue, while girls will be delicate and like pink. Developmental intergroup theory postulates that adults’ heavy focus on gender leads children to pay attention to gender as a key source of information about themselves and others, to seek out any possible gender differences, and to form rigid stereotypes based on gender that are subsequently difficult to change.
There are also psychological theories that partially explain how children form their own gender roles after they learn to differentiate based on gender. The first of these theories is gender schema theory. Gender schema theory argues that children are active learners who essentially socialize them- selves. In this case, children actively organize others’ behavior, activities, and attributes into gender categories, which are known as schemas. These schemas then affect what children notice and remember later. People of all ages are more likely to remember schema-consistent behaviors and attributes than schema-inconsistent behaviors and attributes. So, people are more likely to remember men, and forget women, who are firefighters. They also misremember schema-inconsistent information. If research participants are shown pictures of someone standing at the stove, they are more likely to remember the person to be cooking if depicted as a woman, and the person to be repairing the stove if depicted as a man. By only remembering schema-consistent information, gender schemas strengthen more and more over time.
A second theory that attempts to explain the formation of gender roles in children is social learning theory. Social learning theory argues that gender roles are learned through reinforcement, punishment, and modeling. Children are rewarded and reinforced for behaving in concordance with gender roles and punished for breaking gender roles. In addition, social learning theory argues that children learn many of their gender roles by modeling the behavior of adults and older children and, in doing so, develop ideas about what behaviors are appropriate for each gender. Social learning theory has less support than gender schema theory—research shows that parents do reinforce gender-appropriate play but for the most part treat their male and female children similarly (Lytton & Romney, 1991). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/25%3A_Gender/25.04%3A_Gender_Roles.txt |
Treating boys and girls, and men and women, differently is both a consequence of gender differences and a cause of gender differences. Differential treatment on the basis of gender is also referred to as gender discrimination and is an inevitable consequence of gender stereotypes. When it is based on unwanted treatment related to sexual behaviors or appearance, it is called sexual harassment. By the time boys and girls reach the end of high school, most have experienced some form of sexual harassment, most commonly in the form of unwanted touching or comments, being the target of jokes, having their body parts rated, or being called names related to sexual orientation.
Different treatment by gender begins with parents. A meta-analysis of research from the United States and Canada found that parents most frequently treated sons and daughters differently by encouraging gender-stereotypical activities (Lytton & Romney, 1991). Fathers, more than mothers, are particularly likely to encourage gender-stereotypical play, especially in sons. Parents also talk to their children differently based on stereotypes. For example, parents talk about numbers and counting twice as often with sons than daughters (Chang et al., 2011) and talk to sons in more detail about science than with daughters. Parents are also much more likely to discuss emotions with their daughters than their sons.
Children do a large degree of socializing themselves. By age 3, children play in gender-segregated play groups and expect a high degree of conformity. Children who are perceived as gender atypical (i.e., do not conform to gender stereotypes) are more likely to be bullied and rejected than their more gender-conforming peers.
Gender stereotypes typically maintain gender inequalities in society. The concept of ambivalent sexism recognizes the complex nature of gender attitudes, in which women are often associated with positive and negative qualities (Glick & Fiske, 2001). It has two components. First, hostile sexism refers to the negative attitudes of women as inferior and incompetent relative to men. Second, benevolent sexism refers to the perception that women need to be protected, supported, and adored by men. There has been considerable empirical support for benevolent sexism, possibly because it is seen as more socially acceptable than hostile sexism. Gender stereo- types are found not just in American culture. Across cultures, males tend to be associated with stronger and more active characteristics than females (Best, 2001).
In recent years, gender and related concepts have become a common focus of social change and social debate. Many societies, including American society, have seen a rapid change in perceptions of gender roles, media portrayals of gender, and legal trends relating to gender. For example, there has been an increase in children’s toys attempting to cater to both genders (such as Legos marketed to girls), rather than catering to traditional stereotypes. Nationwide, the drastic surge in acceptance of homosexuality and gender questioning has resulted in a rapid push for legal change to keep up with social change. Laws such as “Don’t Ask, Don’t Tell” and the Defense of Marriage Act (DOMA), both of which were enacted in the 1990s, have met severe resistance on the grounds of being discriminatory toward sexual minority groups and have been accused of unconstitutionality less than 20 years after their implementation. Change in perceptions of gender is also evident in social issues such as sexual harassment, a term that only entered the mainstream mindset in the 1991 Clarence Thomas/Anita Hill scandal. As society’s gender roles and gender restrictions continue to fluctuate, the legal system and the structure of American society will continue to change and adjust.
important Gender-Related Events in the United States
• 1920:19th Amendment (women’s suffrage ratified)
• 1941–1945: World War II forces millions of women to enter the workforce
• 1948: Universal Declaration of Human Rights
• 1963: Congress passes Equal Pay Act
• 1964: Congress passes Civil Rights Act, which outlaws sex discrimination
• 1969: Stonewall riots in New York City, forcing gay rights into the American spotlight
• 1972: Congress passes Equal Rights Amendment; Title IX prohibits sex discrimination in schools and sports
• 1973: American Psychiatric Association removes homosexuality from the DSM
• 1981: First woman appointed to the U.S. Supreme Court • 1987: Average woman earned \$0.68 for every \$1.00 earned by a man
• 1992: World Health Organization no longer considers homosexuality an illness
• 1993: Supreme Court rules that sexual harassment in the workplace is illegal
• 2011: Don’t Ask, Don’t Tell is repealed, allowing people who identify as gay to serve openly in the U.S. military • 2012: President Barack Obama becomes the first American president to openly support LGBT rights and marriage equality
• 2013: Ban against women serving in military combat positions is lifted
REFERENCES
Barnett, B. S., Nesbit, A. E., & Sorrentino, R. M. (2018). The transgender bathroom debate at the intersection of politics, law, ethics, and science. Journal of the American Academy of Psychiatry and the Law, 46(2), 232–241.
Best, D. L. (2001). Gender concepts: Convergence in cross-cultural research and methodologies. Cross-Cultural Research, 35(1), 23–43. https://doi.org/10.1177/106939710103500102
Bigler, R. S., & Liben, L. S. (2007). Developmental intergroup theory: Explaining and reducing children’s social stereotyping and prejudice. Current Directions in Psychological Science, 16(3), 162–166. https:// doi.org/10.1111/j.1467-8721.2007.00496.x
Chang, A. Sandhofer, C., & Brown, C. S. (2011). Gender biases in early number exposure to preschool-aged children. Journal of Language and Social Psychology, 30(4), 440–450. doi.org/10.1177/ 0261927X11416207
Else-Quest, N. M., Hyde, J. S., Goldsmith, H. H., & Van Hulle, C. A. (2006). Gender differences in temperament: A meta-analysis. Psychological Bulletin, 132(1), 33–72. https://doi.org/10.1037/0033-2909.132.1.33
Flores, A. R., Herman, J. L., Gates, G. J., & Brown, T. N. T. (2016). How many adults identify as transgender in the United States? The Williams Institute. https://williamsinstitute.law.ucla.edu/publications/ trans-adults-united-states/
Glick, P., & Fiske, S. T. (2001). An ambivalent alliance: Hostile and benevolent sexism as complementary justifications for gender inequality. American Psychologist, 56(2), 109–118. https://doi.org/ 10.1037/0003-066X.56.2.109
Hyde, J. S. (2005). The gender similarities hypothesis. American Psychologist, 60(6), 581–592. doi.org/10.1037/ 0003-066X.60.6.581
Leaper, C., & Smith, T. E. (2004). A meta-analytic review of gender variations in children’s language use: Talkativeness, affiliative speech, and assertive speech. Developmental Psychology, 40(6), 993–1027. https://doi.org/10.1037/0012-1649.40.6.993
Lytton, H., & Romney, D. M. (1991). Parents’ differential socialization of boys and girls: A meta-analysis. Psychological Bulletin, 109(2), 267–296. https://doi.org/10.1037/0033-2909.109.2.267
Maccoby, E. E. (1998). The two sexes: Growing up apart, coming together. Belknap Press/Harvard University Press.
Martin, C. L., Ruble, D. N., & Szkrybalo, J. (2002). Cognitive theories of early gender development. Psychological Bulletin, 128(6), 903–933. https://doi.org/10.1037/0033-2909.128.6.903
Tzuriel, D., & Egozi, G. (2010). Gender differences in spatial ability of young children: The effects of training and processing strategies. Child Development, 81(5), 1417–1430. doi.org/10.1111/ j.1467-8624.2010.01482.x
Wood, W., & Eagly, A. H. (2002). A cross-cultural analysis of the behavior of women and men: Implications for the origins of sex differences. Psychological Bulletin, 128(5), 699–727. doi.org/10.1037/ 0033-2909.128.5.699 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/25%3A_Gender/25.05%3A_Gender_Sexism_And_Socialization.txt |
Human conflict can result in crime, war, and mass murder, such as genocide. Prejudice and discrimination often are root causes of human conflict, which explains how strangers come to hate one another to the extreme of causing others harm. Prejudice and discrimination affect everyone. In this section we will examine the definitions of prejudice and discrimination, examples of these concepts, and causes of these biases.
26.02: Understanding Prejudice And Discrimination
Humans are very diverse, and although we share many similarities, we also have many differences. The social groups we belong to help form our identities (Tajfel, 1974). These differences may be difficult for some people to reconcile, which may lead to prejudice toward people who are different. Prejudice is a negative attitude and feeling toward an individual based solely on one’s membership in a particular social group (Allport, 1954; Brown, 2010). Prejudice is common against people who are members of an unfamiliar cultural group. Thus, certain types of education, contact, interactions, and building relationships with members of different cultural groups can reduce the tendency toward prejudice. In fact, simply imagining interacting with members of different cultural groups might affect prejudice. Indeed, when experimental participants were asked to imagine themselves positively interacting with someone from a different group, this led to an increased positive attitude toward the other group and an increase in positive traits associated with the other group. Furthermore, imagined social interaction can reduce anxiety associated with inter-group interactions (Crisp & Turner, 2009).
What are some examples of social groups that you belong to that contribute to your identity? Social groups can include gender, race, ethnicity, nationality, social class, religion, sexual orientation, profession, and many more. And, as is true for social roles, you can simultaneously be a member of more than one social group. An example of prejudice is having a negative attitude toward people who are not born in the United States. Although people holding this prejudiced attitude do not know all people who were not born in the United States, they dislike them due to their status as foreigners.
Can you think of a prejudiced attitude you have held toward a group of people? How did your prejudice develop? Prejudice often begins in the form of a stereotype—that is, a specific belief or assumption about individuals based solely on their membership in a group, regardless of their individual characteristics. Stereotypes become overgeneralized and applied to all members of a group. For example, someone holding prejudiced attitudes toward older adults may believe that older adults are slow and incompetent (Cuddy et al., 2005; Nelson, 2004). We cannot possibly know each individual person of advanced age to know that all older adults are slow and incompetent. Therefore, this negative belief is overgeneralized to all members of the group, even though many of the individual group members may in fact be spry and intelligent.
Another example of a well-known stereotype involves beliefs about racial differences among athletes. As Hodge et al. (2008) point out, Black male athletes are often believed to be more athletic, yet less intelligent, than their White male counterparts. These beliefs persist despite a number of high-profile examples to the contrary. Sadly, such beliefs often influence how these athletes are treated by others and how they view themselves and their own capabilities. Whether or not you agree with a stereotype, stereotypes are generally well-known within a given culture (Devine, 1989).
Sometimes people will act on their prejudiced attitudes toward a group of people, and this behavior is known as discrimination. Discrimination is negative action toward an individual as a result of the individual’s membership in a particular group (Allport, 1954; Dovidio & Gaertner, 2004). As a result of holding negative beliefs (stereotypes) and negative attitudes (prejudice) about a particular group, people often treat the target of prejudice poorly, such as excluding older adults from their circle of friends. Table \(1\) summarizes the characteristics of stereotypes, prejudice, and discrimination. Have you ever been the target of discrimination? If so, how did this negative treatment make you feel?
Table \(1\): Connecting Stereotypes, Prejudice, and Discrimination
Item
Function
Connection
Example
Stereotype
Cognitive; thoughts about people
Overgeneralized beliefs about people may lead to prejudice.
“Yankees fans are arrogant and obnoxious.”
Prejudice
Affective; feelings about people, both positive and negative
Feelings may influence treatment of others, leading to discrimination.
“I hate Yankees fans; they make me angry.”
Discrimination
Behavioral; positive or negative treatment of others
Holding stereotypes and harboring prejudice may lead to excluding, avoiding, and biased treatment of group members.
“I would never hire or become friends with a person if I knew he or she were a Yankees fan.”
So far, we’ve discussed stereotypes, prejudice, and discrimination as negative thoughts, feelings, and behaviors because these are typically the most problematic. However, it is important to also point out that people can hold positive thoughts, feelings, and behaviors toward individuals based on group membership; for example, they would show preferential treatment for people who are like themselves—that is, who share the same gender, race, or favorite sports team.
LINK TO LEARNING
This video demonstrates the concepts of prejudice, stereotypes, and discrimination. In the video, a social experiment is conducted in a park where three people try to steal a bike out in the open. The race and gender of the thief is varied: a White male teenager, a Black male teenager, and a White female. Does anyone try to stop them? The treatment of the teenagers in the video demonstrates the concept of racism. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/26%3A_Prejudice_and_Discrimination/26.01%3A_Introduction.txt |
When we meet strangers we automatically process three pieces of information about them: race, gender, and age (Ito & Urland, 2003). Why are these aspects of an unfamiliar person so important? Why don’t we instead notice whether their eyes are friendly, whether they are smiling, their height, the type of clothes they are wearing? Although these secondary characteristics are important in forming a first impression of a stranger, the social categories of race, gender, and age pro- vide a wealth of information about an individual. This information, however, often is based on stereotypes. We may have different expectations of strangers depending on their race, gender, and age. What stereotypes and prejudices do you hold about people who are from a race, gender, and age group different from your own?
Racism
Racism is prejudice and discrimination against an individual based solely on the individual’s membership in a specific racial group (such as toward African Americans, Asian Americans, Latinos, Native Americans, European Americans). What are some stereotypes of various racial or ethnic groups? Research suggests cultural stereotypes for Asian Americans include cold, sly, and intelligent; for Latinos, cold and unintelligent; for European Americans, cold and intelligent; and for African Americans, aggressive, athletic, and more likely to be law breakers (Devine & Elliot, 1995; Dixon & Linz, 2000; Fiske et al., 2002; Sommers & Ellsworth, 2000).
Racism exists for many racial and ethnic groups. For example, Blacks are significantly more likely to have their vehicles searched during traffic stops than Whites, particularly when Blacks are driving in predominately White neighborhoods, a phenomenon often termed “DWB,” or “driving while Black” (Rojek et al., 2012). Mexican Americans and other Latino groups also are targets of racism from the police and other members of the community. For example, when purchasing items with a personal check, Latino shoppers are more likely than White shoppers to be asked to show formal identification (Dovidio et al., 2010).
In one case of alleged harassment by the police, several East Haven, Connecticut, police officers were arrested on federal charges due to reportedly continued harassment and brutalization of Latinos. When the accusations came out, the mayor of East Haven was asked, “What are you doing for the Latino community today?” The Mayor responded, “I might have tacos when I go home, I’m not quite sure yet” (Latino Rebels, 2012). This statement undermines the important issue of racial profiling and police harassment of Latinos, while belittling Latino culture by emphasizing an interest in a food product stereotypically associated with Latinos.
Racism is prevalent toward many other groups in the United States, including Native Americans, Arab Americans, Jewish Americans, and Asian Americans. Have you witnessed racism toward any of these racial or ethnic groups? Are you aware of racism in your community?
One reason modern forms of racism, and prejudice in general, are hard to detect is related to the dual attitudes model (Wilson et al., 2000). Humans have two forms of attitudes: explicit attitudes, which are conscious and controllable, and implicit attitudes, which are unconscious and uncontrollable (Devine, 1989; Olson & Fazio, 2003). Because holding egalitarian views is socially desirable (Plant & Devine, 1998), most people do not show extreme racial bias or other prejudices on measures of their explicit attitudes. However, measures of implicit attitudes often show evidence of mild to strong racial bias or other prejudices (Greenwald et al., 1998; Olson & Fazio, 2003).
Sexism
Sexism is prejudice and discrimination toward individuals based on their sex. Typically, sexism takes the form of men holding biases against women, but either sex can show sexism toward their own or their opposite sex. Like racism, sexism may be subtle and difficult to detect. Common forms of sexism in modern society include gender role expectations, such as expecting women to be the caretakers of the household. Sexism also includes people’s expectations for how members of a gender group should behave. For example, women are expected to be friendly, passive, and nurturing, and when women behave in an unfriendly, assertive, or neglectful manner they often are disliked for violating their gender role (Rudman, 1998). Research by Laurie Rudman (1998) finds that when female job applicants self-promote, they are likely to be viewed as competent, but they may be disliked and are less likely to be hired because they violated gender expectations for modesty. Sexism can exist on a societal level such as in hiring, employment opportunities, and education. Women are less likely to be hired or promoted in male-dominated professions such as engineering, aviation, and construction (Figure \(1\)) (Blau et al., 2010; Ceci & Williams, 2011). Have you ever experienced or witnessed sexism? Think about your family members’ jobs or careers. Why do you think there are differences in the jobs women and men have, such as more women nurses but more male surgeons (Betz, 2008)?
Ageism
People often form judgments and hold expectations about people based on their age. These judgments and expectations can lead to ageism, or prejudice and discrimination toward individuals based solely on their age. Typically, ageism occurs against older adults, but ageism also can occur toward younger adults. Think of expectations you hold for older adults. How could someone’s expectations influence the feelings they hold toward individuals from older age groups? Ageism is wide- spread in U.S. culture (Nosek, 2005), and a common ageist attitude toward older adults is that they are incompetent, physically weak, and slow (Greenberg et al., 2002), and some people consider older adults less attractive. Some cultures, however, including some Asian, Latino, and African American cultures, both outside and within the United States, afford older adults respect and honor.
Ageism can also occur toward younger adults. What expectations do you hold toward younger people? Does society expect younger adults to be immature and irresponsible? How might these two forms of ageism affect a younger and older adult who are applying for a sales clerk position?
Homophobia
Another form of prejudice is homophobia: prejudice and discrimination of individuals based solely on their sexual orientation. Like ageism, homophobia is a widespread prejudice in U.S. society that is tolerated by many people (Herek & McLemore, 2013; Nosek, 2005). Negative feelings often result in discrimination, such as the exclusion of lesbian, gay, bisexual, and transgender (LGBT) people from social groups and the avoidance of LGBT neighbors and co-workers. This discrimination also extends to employers deliberately declining to hire qualified LGBT job applicants. Have you experienced or witnessed homophobia? If so, what stereotypes, prejudiced attitudes, and discrimination were evident? | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/26%3A_Prejudice_and_Discrimination/26.03%3A_Types_Of_Prejudice_And_Discrimination.txt |
Prejudice and discrimination persist in society due to social learning and conformity to social norms. Children learn prejudiced attitudes and beliefs from society: their parents, teachers, friends, the media, and other sources of socialization, such as Facebook (O’Keeffe & Clarke-Pearson, 2011). If certain types of prejudice and discrimination are acceptable in a society, there may be normative pressures to conform and share those prejudiced beliefs, attitudes, and behaviors. For example, public and private schools are still some- what segregated by social class. Historically, only children from wealthy families could afford to attend private schools, whereas children from middle- and low-income families typically attended public schools. If a child from a low-income family received a merit scholarship to attend a private school, how might the child be treated by classmates? Can you recall a time when you held prejudiced attitudes or beliefs or acted in a discriminatory manner because your group of friends expected you to?
DIG DEEPER
Research into Homophobia
Some people are quite passionate in their hatred for non-heterosexuals in our society. In some cases, people have been tortured and/or murdered simply because they were not heterosexual. This passionate response has led some researchers to question what motives might exist for homophobic people. Adams et al. (1996) conducted a study investigating this issue, and their results were quite an eye-opener.
In this experiment, male college students were given a scale that assessed how homophobic they were; those with extreme scores were recruited to participate in the experiment. In the end, 64 men agreed to participate and were split into 2 groups: homophobic men and non-homophobic men. Both groups of men were fitted with a penile plethysmograph, an instrument that measures changes in blood flow to the penis and serves as an objective measurement of sexual arousal.
All men were shown segments of sexually explicit videos. One of these videos involved a sexual interaction between a man and a woman (heterosexual clip). One video displayed two females engaged in a sexual interaction (homosexual female clip), and the final video displayed two men engaged in a sexual interaction (homosexual male clip). Changes in penile tumescence were recorded during all three clips, and a subjective measurement of sexual arousal was also obtained. While both groups of men became sexually aroused to the heterosexual and female homosexual video clips, only those men who were identified as homophobic showed sexual arousal to the homosexual male video clip. While all men reported that their erections indicated arousal for the heterosexual and female homosexual clips, the homophobic men indicated that they were not sexually aroused (despite their erections) to the male homo- sexual clips. Adams et al. (1996) suggest that these findings may indicate that homophobia is related to homosexual arousal that the homophobic individuals either deny or are unaware of. ■
26.05: Stereotypes And Self-fulfilling Prophecy
When we hold a stereotype about a person, we have expectations that he or she will fulfill that stereotype. A self-fulfilling prophecy is an expectation held by a person that alters his or her behavior in a way that tends to make it true. When we hold stereotypes about a person, we tend to treat the person according to our expectations. This treatment can influence the person to act according to our stereotypic expectations, thus confirming our stereotypic beliefs. Research by Rosenthal and Jacobson (1968) found that disadvantaged students whose teachers expected them to perform well had higher grades than disadvantaged students whose teachers expected them to do poorly.
Consider this example of cause and effect in a self-fulfilling prophecy: If an employer expects an openly gay male job applicant to be incompetent, the potential employer might treat the applicant negatively during the interview by engaging in less conversation, making little eye contact, and generally behaving coldly toward the applicant (Hebl et al., 2002). In turn, the job applicant will perceive that the potential employer dislikes him, and he will respond by giving shorter responses to interview questions, making less eye contact, and generally disengaging from the interview. After the inter- view, the employer will reflect on the applicant’s behavior, which seemed cold and distant, and the employer will conclude, based on the applicant’s poor performance during the interview, that the applicant was in fact incompetent. Thus, the employer’s stereotype—gay men are incompetent and do not make good employees—is reinforced. Do you think this job applicant is likely to be hired? Treating individuals according to stereotypic beliefs can lead to prejudice and discrimination.
Another dynamic that can reinforce stereotypes is confirmation bias. When interacting with the target of our prejudice, we tend to pay attention to information that is consistent with our stereotypic expectations and ignore information that is inconsistent with our expectations. In this process, known as confirmation bias, we seek out information that supports our stereotypes and ignore information that is inconsistent with our stereotypes (Wason & Johnson-Laird, 1972). In the job interview example, the employer may not have noticed that the job applicant was friendly and engaging, and that he provided competent responses to the interview questions in the beginning of the interview. Instead, the employer focused on the job applicant’s performance in the later part of the interview, after the applicant changed his demeanor and behavior to match the interviewer’s negative treatment.
Have you ever fallen prey to the self-fulfilling prophecy or confirmation bias, either as the source or target of such bias? How might we stop the cycle of the self-fulfilling prophecy? Social class stereotypes of individuals tend to arise when information about the individual is ambiguous. If information is unambiguous, stereotypes do not tend to arise (Baron et al., 1995). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/26%3A_Prejudice_and_Discrimination/26.04%3A_Why_Do_Prejudice_And_Discrimination_Exist.txt |
As discussed previously in this section, we all belong to a gender, race, age, and socioeconomic group. These groups provide a powerful source of our identity and self-esteem (Tajfel & Turner, 1979). These groups serve as our in-groups. An in-group is a group that we identify with or see ourselves as belonging to. A group that we don’t belong to, or an out- group, is a group that we view as fundamentally different from us. For example, if you are female, your gender in-group includes all females, and your gender out-group includes all males (Figure \(1\)). People often view gender groups as being fundamentally different from each other in personality traits, characteristics, social roles, and interests. Because we often feel a strong sense of belonging and emotional connection to our in-groups, we develop in-group bias: a preference for our own group over other groups. This in-group bias can result in prejudice and discrimination because the out-group is perceived as different and is less preferred than our in-group.
Despite the group dynamics that seem only to push groups toward conflict, there are forces that promote reconciliation between groups: the expression of empathy, the acknowledgment of past suffering on both sides, and the halt of destructive behaviors.
One function of prejudice is to help us feel good about ourselves and maintain a positive self-concept. This need to feel good about ourselves extends to our in-groups: We want to feel good and protect our in-groups. We seek to resolve threats individually and at the in-group level. This often happens by blaming an out-group for the problem. Scapegoating is the act of blaming an out-group when the in-group experiences frustration or is blocked from obtaining a goal (Allport, 1954).
REFERENCES
Adams, H. E., Wright, L. W., Jr., & Lohr, B.A. (1996). Is homophobia associated with homosexual arousal? Journal of Abnormal Psychology, 105(3), 440–445. doi.org/10.1037/ 0021-843X.105.3.440
Allport, G. W. (1954). The nature of human prejudice. Addison-Wesley.
Baron, R. M., Albright, L., & Malloy, T. E. (1995). Effects of behavioral and social class information on social judgment. Personality and Social Psychology Bulletin, 21(4), 308–315. doi.org/10.1177/ 0146167295214001
Betz, N. E. (2008). Women’s career development. In F. Denmark & M. Paludi (Eds.), Psychology of women: Handbook of issues and theories (2nd ed., pp. 717–752). Praeger.
Blau, F. D., Ferber, M. A., & Winkler, A. E. (2010). The economics of women, men, and work (6th ed.). Prentice Hall.
Brown, R. (2010). Prejudice: Its social psychology (2nd ed.). Wiley-Blackwell.
Ceci, S. J., & Williams, W. M. (2011). Understanding current causes of women’s underrepresentation in science. Proceedings of the National Academy of Sciences, 108(8), 3157–3162. https://doi.org/ 10.1073/pnas.1014871108
Crisp, R. J., & Turner, R. N. (2009). Can imagined interactions produce positive perceptions? Reducing prejudice through simulated social contact. American Psychologist, 64(4), 231–240. https://doi.org/ 10.1037/a0014718
Cuddy, A. J., Norton, M. I., & Fiske, S. T. (2005). This old stereotype: The pervasiveness and persistence of the elderly stereotype. Journal of Social Issues, 61(2), 267–285. doi.org/10.1111/ j.1540-4560.2005.00405.x
Devine, P. G. (1989). Stereotypes and prejudice: Their automatic and controlled components. Journal of Personality and Social Psychology, 56(1), 5–18. https://doi.org/10.1037/0022-3514.56.1.5
Devine, P. G., & Elliot, A. J. (1995). Are racial stereotypes really fading? The Princeton trilogy revisited. Personality and Social Psychology Bulletin, 21(11), 1139–1150. https://doi.org/10.1177/01461672952111002
Dixon, T. L., & Linz D. (2000). Overrepresentation and underrepresentation of African Americans and Latinos as lawbreakers on television news. Journal of Communication, 50(2), 131–154. https://doi.org/10.1111/j.1460-2466.2000.tb02845.x
Dovidio, J. F., & Gaertner, S. L. (2004). On the nature of contemporary prejudice. In P. S. Rothenberg, (Ed.), Race, class, and gender in the United States: An integrated study (6th ed., pp. 132–142). Worth.
Dovidio, J. F., Gluszek, A., John, M. S., Ditlmann, R., & Lagunes, P. (2010). Understanding bias toward Latinos: Discrimination, dimensions of difference, and experience of exclusion. Journal of Social Issues, 66(1), 59–78. doi.org/10.1111/j.1540-4560.2009.01633.x
Fiske, S. T., Cuddy, A. J., Glick, P., & Xu, J. (2002). A model of (often mixed) stereotype content: Competence and warmth respectively follow from perceived status and competition. Journal of Personality and Social Psychology, 82(6), 878–902. doi. org/10.1037/0022-3514.82.6.878
Greenberg, J., Schimel, J., & Martens, A. (2002). Ageism: Denying the face of the future. In T. D. Nelson (Ed.), Ageism: Stereotyping and prejudice against older persons (pp. 27–48). MIT Press.
Greenwald, A. G., McGhee, D. E., & Schwartz, J. L. (1998). Measuring individual differences in implicit cognition: The implicit association test. Journal of Personality and Social Psychology, 74(6), 1464–1480. https://doi.org/10.1037/0022-3514.74.6.1464
Hebl, M. R., Foster, J. B., Mannix, L. M., & Dovidio, J. F. (2002). Formal and interpersonal discrimination: A field study of bias toward homosexual applicants. Personality and Social Psychology Bulletin, 28(6), 815–825. https://doi.org/10.1177/0146167202289010
Herek, G. M., & McLemore, K. A. (2013). Sexual prejudice. Annual Review of Psychology, 64(1), 309–333. doi.org/10.1146/ annurev-psych-113011-143826
Hodge, S. R., Burden, J. W., Jr., Robinson, L. E., & Bennett, R. A., III. (2008). Theorizing on the stereotyping of black male student- athletes: Issues and implications. Journal for the Study of Sports and Athletes in Education, 2(2), 203–226. doi.org/10.1179/ ssa.2008.2.2.203
Ito, T. A., & Urland, G. R. (2003). Race and gender on the brain: Electrocortical measures of attention to race and gender of multiply categorizable individuals. Journal of Personality and Social Psychology, 85(4), 616–626. doi.org/10.1037/ 0022-3514.85.4.616
Latino Rebels. (2012, January 25). East Haven mayor suggests “he might have tacos” to repair relations with Latinos [Video]. YouTube. https:// www.youtube.com/watch?v=PCUwtfqF4wU
Nelson, T. (Ed.). (2004). Ageism: Stereotyping and prejudice against older persons. MIT Press.
Nosek, B. A. (2005). Moderators of the relationship between implicit and explicit evaluation. Journal of Experimental Psychology: General, 134(4), 565–584. https://doi.org/10.1037/0096-3445.134.4.565
O’Keeffe, G. S., & Clarke-Pearson, K. (2011). The impact of social media on children, adolescents, and families. Pediatrics, (127)4, 800–4. https://doi.org/10.1542/peds.2011-0054
Olson, M. A., & Fazio, R. H. (2003). Relations between implicit measures of prejudice: What are we measuring? Psychological Science, 14(6), 636–639. https://doi.org/10.1046/j.0956-7976.2003.psci_1477.x
Plant, E. A., & Devine, P. G. (1998). Internal and external motivation to respond without prejudice. Journal of Personality and Social Psychology, 75(3), 811–832. doi.org/10.1037/ 0022-3514.75.3.811
Rojek, J., Rosenfeld, R., & Decker, S. (2012). Policing race: The racial stratification of searches in police traffic stops. Criminology, 50(4), 993–1024. doi.org/10.1111/j.1745-9125.2012.00285.x
Rosenthal, R., & Jacobson, L. F. (1968). Teacher expectations for the disadvantaged. Scientific American, 218(4), 19–23. https://doi.org/ 10.1038/scientificamerican0468-19
Rudman, L. A. (1998). Self-promotion as a risk factor for women: The costs and benefits of counterstereotypical impression management. Journal of Personality and Social Psychology, 74(3), 629–645. https:// doi.org/10.1037/0022-3514.74.3.629
Sommers, S. R., & Ellsworth, P. C. (2000). Race in the courtroom: Perceptions of guilt and dispositional attributions. Personality and Social Psychology Bulletin, 26(11), 1367–1379. doi.org/10.1177/ 0146167200263005
Tajfel, H. (1974). Social identity and intergroup behaviour. Social Science Information, 13(2), 65–93. doi.org/10.1177/ 053901847401300204
Tajfel, H., & Turner, J. C. (1979). An integrative theory of intergroup conflict. In W. G. Austin & S. Worchel (Eds.), The social psychology of intergroup relations (pp. 33–48). Brooks-Cole.
Wason, P. C., & Johnson-Laird, P. N. (1972). The psychology of deduction: Structure and content. Harvard University Press.
Wilson, T. D., Lindsey, S., & Schooler, T. Y. (2000). A model of dual attitudes. Psychological Review, 107(1), 101–126. doi. org/10.1037/0033-295X.107.1.101 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/26%3A_Prejudice_and_Discrimination/26.06%3A_In-groups_And_Out-groups.txt |
This module examines what cognitive development is, major theories about how it occurs, the roles of nature and nurture, whether it is continuous or discontinuous, and how research in the area is being used to improve education.
By the time you reach adulthood, you have learned a few things about how the world works. You know, for instance, that you can’t walk through walls or leap into the tops of trees. You know that although you cannot see your car keys they’ve got to be around here someplace. What’s more, you know that if you want to communicate complex ideas like ordering a triple-shot soy vanilla latte with chocolate sprinkles it’s better to use words with meanings attached to them rather than simply gesturing and grunting. People accumulate all this useful knowledge through the process of cognitive development, which involves a multitude of factors, both inherent and learned.
Cognitive development refers to the development of thinking across the lifespan. Defining thinking can be problematic, because no clear boundaries separate thinking from other mental activities. Thinking obviously involves the higher mental processes: problem solving, reasoning, creating, conceptualizing, categorizing, remembering, planning, and so on. However, thinking also involves other mental processes that seem more basic and at which even toddlers are skilled—such as perceiving objects and events in the environment, acting skillfully on objects to obtain goals, and understanding and producing language. Yet other areas of human development that involve thinking are not usually associated with cognitive development, because thinking isn’t a prominent feature of them—such as personality and temperament.
As the name suggests, cognitive development is about change. Children’s thinking changes in dramatic and surprising ways. Consider DeVries’s (1969) study of whether young children understand the difference between appearance and reality. To find out, she brought an unusually even-tempered cat named Maynard to a psychology laboratory and allowed the 3- to 6-year-old participants in the study to pet and play with him. DeVries then put a mask of a fierce dog on Maynard’s head, and asked the children what Maynard was. Despite all of the children having identified Maynard previously as a cat, now most 3-year-olds said that he was a dog and claimed that he had a dog’s bones and a dog’s stomach. In contrast, the 6-year-olds weren’t fooled; they had no doubt that Maynard remained a cat. Understanding how children’s thinking changes so dramatically in just a few years is one of the fascinating challenges in studying cognitive development.
There are several main types of theories of child development. Stage theories, such as Piaget’s stage theory, focus on whether children progress through qualitatively different stages of development. Sociocultural theories, such as that of Lev Vygotsky, emphasize how other people and the attitudes, values, and beliefs of the surrounding culture, influence children’s development. Information processing theories, such as that of David Klahr, examine the mental processes that produce thinking at any one time and the transition processes that lead to growth in that thinking.
At the heart of all of these theories, and indeed of all research on cognitive development, are two main questions: (1) How do nature and nurture interact to produce cognitive development? (2) Does cognitive development progress through qualitatively distinct stages? In the remainder of this module, we examine the answers that are emerging regarding these questions, as well as ways in which cognitive develop- mental research is being used to improve education. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/27%3A_Cognitive_Development_in_Childhood/27.01%3A_Introduction.txt |
The most basic question about child development is how nature and nurture together shape development. Nature refers to our biological endowment, the genes we receive from our parents. Nurture refers to the environments, social as well as physical, that influence our development, everything from the womb in which we develop before birth to the homes in which we grow up, the schools we attend, and the many people with whom we interact.
The nature-nurture issue is often presented as an either-or question: Is our intelligence (for example) due to our genes or to the environments in which we live? In fact, however, every aspect of development is produced by the interaction of genes and environment. At the most basic level, without genes, there would be no child, and without an environment to provide nurture, there also would be no child.
The way in which nature and nurture work together can be seen in findings on visual development. Many people view vision as something that people either are born with or that is purely a matter of biological maturation, but it also depends on the right kind of experience at the right time. For example, development of depth perception, the ability to actively perceive the distance from oneself to objects in the environment, depends on seeing patterned light and having normal brain activity in response to the patterned light, in infancy (Held, 1993). If no patterned light is received, for example when a baby has severe cataracts or blindness that is not surgically corrected until later in development, depth perception remains abnormal even after the surgery.
Adding to the complexity of the nature-nurture interaction, children’s genes lead to their eliciting different treatment from other people, which influences their cognitive development. For example, infants’ physical attractiveness and temperament are influenced considerably by their genetic inheritance, but it is also the case that parents provide more sensitive and affectionate care to easygoing and attractive infants than to difficult and less attractive ones, which can contribute to the infants’ later cognitive development (Langlois et al., 1995; van den Boom & Hoeksma, 1994).
Also contributing to the complex interplay of nature and nurture is the role of children in shaping their own cognitive development. From the first days out of the womb, children actively choose to attend more to some things and less to others. For example, even 1-month-olds choose to look at their mother’s face more than at the faces of other women of the same age and general level of attractiveness (Bartrip et al., 2001). Children’s contributions to their own cognitive development grow larger as they grow older (Scarr & McCartney, 1983). When children are young, their parents largely determine their experiences: whether they will attend day care, the children with whom they will have play dates, the books to which they have access, and so on. In contrast, older children and adolescents choose their environments to a larger degree. Their parents’ preferences largely determine how 5-year-olds spend time, but 15-year-olds’ own preferences largely determine when, if ever, they set foot in a library. Children’s choices often have large consequences. To cite one example, the more children choose to read, the more their reading improves in future years (Baker et al., 2000). Thus, the issue is not whether cognitive development is a product of nature or nurture; rather, the issue is how nature and nurture work together to produce cognitive development. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/27%3A_Cognitive_Development_in_Childhood/27.02%3A_Nature_And_Nurture.txt |
Some aspects of the development of living organisms, such as the growth of the width of a maple tree, involve quantitative changes, with the tree getting a little wider each year. Other changes, such as the life cycle of a ladybug, involve qualitative changes, with the creature becoming a totally different type of entity after a transition than before (Figure \(1\)). The existence of both gradual, quantitative changes and relatively sudden, qualitative changes in the world has led researchers who study cognitive development to ask whether changes in children’s thinking are gradual and continuous or sudden and discontinuous.
The great Swiss psychologist Jean Piaget proposed that children’s thinking progresses through a series of four discrete stages. By “stages,” he meant periods during which children reasoned similarly about many superficially different problems, with the stages occurring in a fixed order and the thinking within different stages differing in fundamental ways. The four stages that Piaget hypothesized were the sensorimotor stage (birth to 2 years), the preoperational reasoning stage (2 to 6 or 7 years), the concrete operational reasoning stage (6 or 7 to 11 or 12 years), and the formal operational reasoning stage (11 or 12 years and throughout the rest of life).
During the sensorimotor stage (Figure \(2\)), children’s thinking is largely realized through their perceptions of the world and their physical interactions with it. Their mental representations are very limited. Consider Piaget’s object permanence task, which is one of his most famous problems. If an infant younger than 9 months of age is playing with a favorite toy, and another person removes the toy from view, for example by putting it under an opaque cover and not letting the infant immediately reach for it, the infant is very likely to make no effort to retrieve it and to show no emotional distress (Piaget, 1954). This is not due to their being uninterested in the toy or unable to reach for it; if the same toy is put under a clear cover, infants below 9 months readily retrieve it (Munakata et al., 1997). Instead, Piaget claimed that infants less than 9 months do not understand that objects continue to exist even when out of sight.
During the preoperational stage, according to Piaget (Figure \(2\)), children can solve not only this simple problem (which they actually can solve after 9 months) but show a wide variety of other symbolic-representation capabilities, such as those involved in drawing and using language. However, such 2- to 7-year-olds tend to focus on a single dimension, even when solving problems would require them to consider multiple dimensions. This is evident in Piaget’s (1952) conservation problems. For example, if a glass of water is poured into a taller, thinner glass, children below age 7 generally say that there now is more water than before. Similarly, if a clay ball is reshaped into a long, thin sausage, they claim that there is now more clay, and if a row of coins is spread out, they claim that there are now more coins. In all cases, the children are focusing on one dimension, while ignoring the changes in other dimensions (for example, the greater width of the glass and the clay ball).
Children overcome this tendency to focus on a single dimension during the concrete operations stage (Figure \(3\)), and think logically in most situations. However, according to Piaget, they still cannot think in systematic scientific ways, even when such thinking would be useful. Thus, if asked to find out which variables influence the period that a pendulum takes to complete its arc, and given weights that they can attach to strings in order to do experiments with the pendulum to find out, most children younger than age 12 perform biased experiments from which no conclusion can be drawn, and then conclude that whatever they originally believed is correct. For example, if a boy believed that weight was the only variable that mattered, he might put the heaviest weight on the shortest string and push it the hardest, and then conclude that just as he thought, weight is the only variable that matters (Inhelder & Piaget, 1958).
Finally, in the formal operations period (Figure \(3\)), children attain the reasoning power of mature adults, which allows them to solve the pendulum problem and a wide range of other problems. However, this formal operations stage tends not to occur without exposure to formal education in scientific reasoning and appears to be largely or completely absent from some societies that do not provide this type of education.
Although Piaget’s theory has been very influential, it has not gone unchallenged. Many more recent researchers have obtained findings indicating that cognitive development is considerably more continuous than Piaget claimed. For example, Diamond (1985) found that on the object permanence task described above, infants show earlier knowledge if the waiting period is shorter. At age 6 months, they retrieve the hidden object if the wait is no longer than 2 seconds; at 7 months, they retrieve it if the wait is no longer than 4 seconds; and so on. Even earlier, at 3 or 4 months, infants show surprise in the form of longer looking times if objects suddenly appear to vanish with no obvious cause (Baillargeon 1987). Similarly, children’s specific experiences can greatly influence when developmental changes occur. Children of pottery makers in Mexican villages, for example, know that reshaping clay does not change the amount of clay at much younger ages than children who do not have similar experiences (Price-Williams et al., 1969).
So, is cognitive development fundamentally continuous or fundamentally discontinuous? A reasonable answer seems to be, “It depends on how you look at it and how often you look.” For example, under relatively facilitative circumstances, infants show early forms of object permanence by 3 or 4 months, and they gradually extend the range of times for which they can remember hidden objects as they grow older. However, on Piaget’s original object permanence task, infants do quite quickly change toward the end of their first year from not reaching for hidden toys to reaching for them, even after they’ve experienced a substantial delay before being allowed to reach. Thus, the debate between those who emphasize discontinuous, stage-like changes in cognitive development and those who emphasize gradual continuous changes remains a lively one. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/27%3A_Cognitive_Development_in_Childhood/27.03%3A_Does_Cognitive_Development_Progress_Through_Distinct_Stages.txt |
Understanding how children think and learn has proven useful for improving education. One example comes from the area of reading. Cognitive developmental research has shown that phonemic awareness—that is, awareness of the component sounds within words—is a crucial skill in learning to read. To measure awareness of the component sounds within words, researchers ask children to decide whether two words rhyme, to decide whether the words start with the same sound, to identify the component sounds within words, and to indicate what would be left if a given sound were removed from a word. Kindergartners’ performance on these tasks is the strongest predictor of reading achievement in third and fourth grade, even stronger than IQ or social class back- ground (Nation, 2008). Moreover, teaching these skills to randomly chosen 4- and 5-year-olds results in their being better readers years later (National Reading Panel, 2000).
Another educational application of cognitive developmental research involves the area of mathematics. Even before they enter kindergarten, the mathematical knowledge of children from low-income backgrounds lags far behind that of children from more affluent backgrounds. Ramani and Siegler (2008) hypothesized that this difference is due to the children in middle- and upper-income families engaging more frequently in numerical activities, for example playing numerical board games such as Chutes and Ladders. Chutes and Ladders is a game with a number in each square; children start at the number one and spin a spinner or throw a die to determine how far to move their token. Playing this game seemed likely to teach children about numbers, because in it, larger numbers are associated with greater values on a variety of dimensions. In particular, the higher the number that a child’s token reaches, the greater the distance the token will have traveled from the starting point, the greater the number of physical movements the child will have made in moving the token from one square to another, the greater the number of number-words the child will have said and heard, and the more time will have passed since the beginning of the game. These spatial, kinesthetic, verbal, and time-based cues pro- vide a broad-based, multisensory foundation for knowledge of numerical magnitudes (the sizes of numbers), a type of knowledge that is closely related to mathematics achievement test scores (Booth & Siegler, 2006).
Playing this numerical board game for roughly 1 hour, distributed over a 2-week period, improved low-income children’s knowledge of numerical magnitudes, ability to read printed numbers, and skill at learning novel arithmetic problems. The gains lasted for months after the game-playing experience (Ramani & Siegler, 2008; Siegler & Ramani, 2009). An advantage of this type of educational intervention is that it has minimal if any cost—a parent could just draw a game on a piece of paper.
Understanding of cognitive development is advancing on many different fronts. One exciting area is linking changes in brain activity to changes in children’s thinking (Nelson et al., 2006). Although many people believe that brain maturation is something that occurs before birth, the brain actually continues to change in large ways for many years thereafter. For example, a part of the brain called the prefrontal cortex, which is located at the front of the brain and is particularly involved with planning and flexible problem solving, continues to develop throughout adolescence (Blakemore & Choudhury, 2006). Such new research domains, as well as enduring issues such as nature and nurture, continuity and discontinuity, and how to apply cognitive development research to education, insure that cognitive development will continue to be an exciting area of research in the coming years. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/27%3A_Cognitive_Development_in_Childhood/27.04%3A_Applications_To_Education.txt |
Research into cognitive development has shown us that minds don’t just form according to a uniform blueprint or innate intellect, but through a combination of influencing factors. For instance, if we want our kids to have a strong grasp of language, we could concentrate on phonemic awareness early on. If we want them to be good at math and science we could engage them in numerical games and activities early on. Perhaps most importantly, we no longer think of brains as empty vessels waiting to be filled up with knowledge but as adaptable organs that develop all the way through early adulthood.
REFERENCES
Baillargeon, R. (1987). Object permanence in 31⁄2- and 41⁄2-month-old infants. Developmental Psychology, 23(5), 655–664. https://doi.org/ 10.1037/0012-1649.23.5.655
Baker, L., Dreher, M. J., & Guthrie, J. T., (Eds.). (2000). Engaging young readers: Promoting achievement and motivation. New York: Guilford.
Bartrip, J., Morton, J., & de Schonen, S. (2001). Responses to mother’s face in 3-week to 5-month old infants. British Journal of Developmental Psychology, 19(2), 219–232. doi.org/10.1348/ 026151001166047
Blakemore, S.-J., & Choudhury, S. (2006). Development of the adolescent brain: Implications for executive function and social cognition. Journal of Child Psychiatry and Psychology, 47(3–4), 296–312. doi.org/10.1111/j.1469-7610.2006.01611.x
Booth, J. L., & Siegler, R. S. (2006). Developmental and individual differences in pure numerical estimation. Developmental Psychology, 42(1), 189–201. https://doi.org/10.1037/0012-1649.41.6.189
DeVries, R. (1969). Constancy of generic identity in the years three to six. Monographs of the Society for Research in Child Development, 34(3). doi.org/10.2307/1165683
Diamond, A. (1985). Development of the ability to use recall to guide action, as indicated by infants’ performance on AB. Child Development, 56(4), 868–883. doi.org/10.2307/1130099
Held, R. (1993). What can rates of development tell us about underlying mechanisms? In C. E. Granrud (Ed.), Visual perception and cognition in infancy (pp. 75–90). Erlbaum.
Inhelder, B., & Piaget, J. (1958). The growth of logical thinking from childhood to adolescence. Basic Books.
Langlois, J. H., Ritter, J. M., Casey, R. J., & Sawin, D. B. (1995). Infant attractiveness predicts maternal behaviors and attitudes. Developmental Psychology, 31(3), 464–472.
Munakata, Y., McClelland, J. L., Johnson, M. H., & Siegler, R. S. (1997). Rethinking infant knowledge: Toward an adaptive process account of successes and failures in object permanence tasks. Psychological Review, 104(4), 686–713. doi.org/10.1037/ 0033-295X.104.4.686
Nation, K. (2008). Learning to read words. Quarterly Journal of Experimental Psychology, 61(8), 1121–1133. doi.org/10.1080/ 17470210802034603
National Reading Panel. (2000). Teaching children to read: An evidence- based assessment of the scientific research literature on reading and its implications for reading instruction. National Institute of Child Health and Human Development.
Nelson, C. A., Thomas, K. M., & de Haan, M. (2006). Neural bases of cognitive development. In W. Damon & R. M. Lerner (Series Eds.) & D. Kuhn & R. S. Siegler (Vol. Eds.), Handbook of child psychology: Volume 2: Cognition, perception, and language (6th ed., pp. 3–57). Wiley.
Piaget, J. (1952). The child’s concept of number. W. W. Norton. Piaget, J. (1954). The construction of reality in the child. Basic Books.
Price-Williams, D. R., Gordon, W., & Ramirez, M. (1969). Skill and conservation: A study of pottery making children. Developmental Psychology, 1(6, Pt. 1), 769. https://doi.org/10.1037/h0028264
Ramani, G. B., & Siegler, R. S. (2008). Promoting broad and stable improvements in low-income children’s numerical knowledge through playing number board games. Child Development, 79(2), 375–394. doi.org/10.1111/j.1467-8624.2007.01131.x
Scarr, S., & McCartney, K. (1983). How people make their own environments: A theory of genotype-environment effects. Child Development, 54(2), 424–435. doi.org/10.2307/1129703
Siegler, R. S., & Ramani, G. B. (2009). Playing linear number board games—but not circular ones—improves low-income preschoolers’ numerical understanding. Journal of Educational Psychology, 101(3), 545–560. https://doi.org/10.1037/a0014239
van den Boom, D. C., & Hoeksma, J. B. (1994). The effect of infant irritability on mother-infant interaction: A growth curve analysis. Developmental Psychology, 30(4), 581–590. doi.org/10.1037/ 0012-1649.30.4.581 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/27%3A_Cognitive_Development_in_Childhood/27.05%3A_Conclusion.txt |
Children are completely egoistic; they feel their needs intensely and strive ruthlessly to satisfy them.
It is impossible to overlook the extent to which civilization is built upon a renunciation of instinct.
The first requisite of civilization is that of justice.
—Sigmund Freud
Frankenstein’s monster understood the difference between “good” (e.g., food) and “bad” (e.g., fire). Unfortunately, he didn’t have “right” and “wrong” down pat! Piaget’s interests extended beyond the development of knowledge and skills related to nature (i.e., scientific thinking). He was also deeply interested in the individual’s development of a moral code (Piaget, 1932). Not surprisingly, Piaget believed that the cognitive changes occurring as the child and teenager advanced through the developmental stages influenced their moral thinking as well as their understanding of nature. During the pre-verbal sensorimotor stage, direct learning principles account for changes in behavior. The child increases the frequency of behaviors resulting in appetitive (i.e., “feel good”) or reducing aversive (i.e., “feel bad”) outcomes and sup- presses behaviors resulting in aversive or the loss of appetitive outcomes. As the child initially acquires language during the preoperational stage, rules are imposed by adults (primarily parents and caregivers) and understood in a literal, inflexible way. Later, the child gradually interacts with other children, makes friends, and goes to school. The parents’ influence is diluted by the direct and indirect (i.e., observational and verbal) contingencies experienced with different adults (e.g., teachers, members of the clergy, etc.) and their peers. As the child becomes less egocentric during the stage of concrete operations, he/she is able to appreciate the perspectives of others and recognize the possibility and need to cooperate by negotiating rules of conduct. Once attaining the stage of formal operations, teenagers and adults are able to appreciate and consider more subtle and abstract aspects of interpersonal and moral issues (e.g., the benefits and need for fairness, justice, responsibility, etc.).
28.02: Kohlberg's State Model Of Moral Development
Lawrence Kohlberg (1976) developed a very influential stage model of moral development based on Piaget’s stage model of cognitive development (see Figure \(1\)). He distinguished between three different levels (“styles”) of reasoning: pre-conventional, conventional, and post-conventional, each subdivided for a total of six stages. Pre-conventional morality is based upon extrinsic rewards and punishes. At first, during Piaget’s sensorimotor period, the child is only sensitive to extrinsic rewards and punishers. Once the child acquires speech during Piaget’s pre-operational stage, distinctions between right and wrong are taught by parents and other authority figures. The child learns the value of cooperation (e.g., “I’ll scratch your back and you scratch mine”) once making friends and interacting with others. Conventional morality is based on reference to an authority figure (e.g., parent, teacher, clergy member, etc.) at first and then advances to written sources (e.g., the Bible, the Koran, the Constitution, etc.). The child acquires a more abstract and flexible understanding of morality once progressing to the stages of concrete and formal operations. The highest (and rarest) post-conventional morality level is based on the application of universal principles such as the Golden Rule (Do unto others as you would have others do unto you).
In attempting to teach codes of moral conduct, much parenting consists of the intentional or non-intentional administration of appetitive and aversive events. We may consider how the different parental styles implement learning procedures and how they may relate to Kohlberg’s levels of moral development (see Table \(1\)).
Table \(1\): Parental Styles and Stages of Morality
Adapted from Levy (2013).
Indifferent
Indulgent
Authoritarian
Authoritative
Unavailable to monitor behavior, administer consequences consistently, or provide explanations
Available to administer non-contingent presentation of appetitive events and provide praise
Available to administer contingent presentation of mostly aversive events without explanation
Available to administer contingent presentation of appetitive and aversive events with explanation
Pre-conventional morality
Sense of entitlement
Conventional morality
Post-conventional morality
Indifferent parents (low demandingness, low responsiveness) do not specify codes of conduct or respond to their children’s needs. If other people (siblings, relatives, peers) do not provide rules and/or consequences, the children will most likely base right and wrong on the outcomes of their actions (if it feels good it is right; if it feels bad it is wrong). Indifferent parenting would appear to be most likely to produce pre-conventional reasoning in children.
Indulgent parents (low demandingness, high responsiveness) do not specify codes of conduct but are affectionate and responsive. They provide “unconditional positive regard” (Rogers, 1957), the type of non-contingent appetitive consequence likely to result in “spoiling” and a sense of entitlement. This could create problems for the children in other contexts (e.g., school, playgrounds) when others react differently to their behavior.
Authoritarian parents (high demandingness, low responsiveness) specify strict codes of conduct in a non- responsive manner. If the children ask for reasons, they may reply “because I say so!” In their parents’ absence, the children would seek other sources of authority. The authoritarian parenting style would appear likely to result in conventional reasoning.
Authoritative parents (high demandingness, high responsiveness) specify strict codes of conduct within a con- text of warmth and sensitivity to the children’s needs. They are likely to provide reasons for their codes of conduct, listen to their children’s perspective, and in some instances, negotiate alternative codes. The authoritative style would likely result in post-conventional reasoning. Ideally, by providing reasons and explanations in age-appropriate language, our children would internalize principles of moral conduct and apply them appropriately throughout their lives.
It is very difficult to administer punishment immediately and on a consistent basis in the natural, free-living environment. Therefore, it is not likely that punishment will work as intended, to suppress undesired behavior. Often, instead, the child will learn to become deceptive or lie in order to avoid being punished. Indifferent parents are not likely to be present to appropriately administer punishment and will probably be inconsistent. Indulgent parents are less likely to administer punishment than other parents, if at all. Authoritarian parents (“my way or the highway”) might effectively suppress the undesired behavior when they are present. However, the behavior may occur when they are not present or when the child is in different situations. Authoritative parents, taking advantage of their children’s verbal and reasoning skills, probably have the greatest likelihood of attaining the desired result. For example, an older sibling picking on a younger one might be told the following scenario which includes stipulation of rules of conduct:
There is a difference between a jungle and a society. In the jungle, strong animals often attack weaker animals who receive no protection. Human beings have families and societies in which the strong protect the weak and help them grow stronger. You have to decide whether you want to live in our family and be a member of society. If you keep picking on your little brother/ sister, we will need to treat you like an animal from the jungle. We put dangerous animals in a zoo so they cannot hurt anyone, so we will keep you in your room. If you take care of your little brother/sister, mommy and daddy will let you play together and have fun.
By relying upon language to stipulate and enforce rules in this manner, a parent is most likely to achieve the short- term objective of encouraging appropriate and discouraging inappropriate behavior. In addition, by providing thoughtful explanations and justifications of rules, the parent increases the likelihood that the child will internalize a moral code of conduct as he/she matures. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/28%3A_Theories_of_Development/28.01%3A_Moral_Development.txt |
Erik Erikson (1950; 1959) proposed a “cradle to grave” sequence of development that complements the stage theories of Piaget and Kohlberg. Erikson described eight “con- flicts” associated with different periods of one’s life (see Table \(1\)). It was assumed that successful resolution of the conflict associated with a particular stage resulted in acquisition of the related “virtue” (e.g., trust, autonomy, initiative, etc.) for the rest of one’s life. Unsuccessful resolution would result in developmental problems during subsequent stages.
Table \(1\): Erikson’s Eight Stages of Human Development
Erikson’s Psychosocial Stages
Stages Crisis Favorable Outcome Unfavorable Outcome
Childhood
1st year of life
Trust vs. mistrust
Faith in the environment and future events
Suspicion, fear of future events
2nd year
Autonomy vs. doubt
A sense of self-control and adequacy
Feelings of shame and self-doubt
3rd through 5th years
Initiative vs. guilt
Ability to be a “self-starter,” to initiate one’s own activities
A sense of guilt and inadequacy to be on one’s own
6th year to puberty
Industry vs. inferiority
Ability to learn how things work, to understand and organize
A sense of inferiority at understanding and organizing
Transition Years
Adolescence
Identity vs. confusion
Seeing oneself as a unique and integrated person
Confusion over who and what one really is
Adulthood
Early adulthood
Intimacy vs. isolation
Ability to make commitments to others, to love
Inability to form affectionate relationship
Middle age
Generativity vs. self-absorption
Concern for family and society in general
Concern only for self—one’s own well- being and prosperity
Aging years
Integrity vs. despair
A sense of integrity and fulfillment; willingness to face death
Dissatisfaction with life; despair over prospect of death
28.04: Childhood
During Erikson’s first (infancy) stage, taking place during Piaget’s pre-verbal sensorimotor stage, the attachment style of the caregiver will influence whether or not the infant experiences a nurturing and responsive social environment. If the caregiver is consistent in satisfying the basic needs for food, comfort, and relief from pain, the infant learns to trust them. If negligent, inconsistent, or abusive, the child will mistrust and perhaps fear the caregiver.
In the second (early childhood) stage, starting toward the end of Piaget’s sensorimotor stage and extending into the beginning of the verbal preoperational stage, the young child is exploring and learning to control the environment on its own. A patient caregiver waits until the zone of proximal development is reached and applies encouraging, supportive scaffolding techniques during toilet training and other learn- ing experiences. Such a parent is likely to insure the child’s success, resulting in the feeling of independence and autonomy. If the child is hurried, scolded, or punished for failures, she/he may feel shame and doubt her/his capabilities.
In the third (preschool) stage, occurring in the middle of Piaget’s preoperational stage, the child must learn to dress and groom in a manner consistent with social norms and standards. If the child is encouraged to explore options, satisfy its curiosity, and express its own preferences and interests, it is likely to develop initiative. If discouraged, the child may become passive and doubtful of its own capabilities and experience guilt regarding its choices.
Erikson’s lengthy fourth (school age) stage starts toward the end of Piaget’s preoperational stage and extends through concrete operations into the beginning of the final formal operations stage. If at home and school, the child is appropriately challenged and succeeds at progressively more difficult tasks, it becomes competent, confident, and industrious.
The child must experience and learn to cope with frustration and inevitable failure. It is during this stage that the child becomes concerned about its own performance in compar- ison to others in and out of school. Feelings of inferiority can result from perceived inadequacies and negative social comparisons.
REFERENCES
Erikson, E. H. (1950). Childhood and society. W. W. Norton & Company. Erikson, E. H. (1959). Identity and the life cycle. International Universities
Press.
Kohlberg, L. (1976). Moral stages and moralization: The cognitive-
development approach. In T. Lickona (Ed.), Moral development and behavior. Holt, Rinehart and Winston.
Levy, J. C. (2013). Adaptive learning and the human condition. Pearson. Piaget, J. (1932). The moral judgment of the child. Routledge & Kegan
Paul.
Rogers, C. (1957). The necessary and sufficient conditions of therapeutic
personality change. Journal of Consulting Psychology, 21(2), 95–103. https://doi.org/10.1037/h0045357 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/28%3A_Theories_of_Development/28.03%3A_Erik_Erikson%27s_Stage_Theory_Of_Lifespan_Development.txt |
The purpose of this module is to provide a brief review of attachment theory—a theory designed to explain the significance of the close, emotional bonds that children develop with their caregivers and the implications of those bonds for understanding personality development. The module discusses the origins of the theory, research on individual differences in attachment security in infancy and childhood, and the role of attachment in adult relationships.
Some of the most rewarding experiences in people’s lives involve the development and maintenance of close relation- ships. For example, some of the greatest sources of joy involve falling in love, starting a family, being reunited with distant loved ones, and sharing experiences with close others. And, not surprisingly, some of the most painful experiences in people’s lives involve the disruption of important social bonds, such as separation from a spouse, losing a parent, or being abandoned by a loved one.
Why do close relationships play such a profound role in human experience? Attachment theory is one approach to understanding the nature of close relationships. In this module, we review the origins of the theory, the core theoretical principles, and some ways in which attachment influences human behavior, thoughts, and feelings across the life course.
29.02: Attachment Theory- A Brief History And Core Concepts
Attachment theory was originally developed in the 1940s by John Bowlby, a British psychoanalyst who was attempting to understand the intense distress experienced by infants who had been separated from their parents. Bowlby (1969) observed that infants would go to extraordinary lengths to prevent separation from their parents or to reestablish proximity to a missing parent. For example, he noted that children who had been separated from their parents would often cry, call for their parents, refuse to eat or play, and stand at the door in desperate anticipation of their parents’ return. At the time of Bowlby’s initial writings, psychoanalytic writers held that these expressions were manifestations of immature defense mechanisms that were operating to repress emotional pain. However, Bowlby observed that such expressions are common to a wide variety of mammalian species and speculated that these responses to separation may serve an evolutionary function (see Focus Topic 1).
FOCUS TOPIC 1
Harlow’s Research on Contact Comfort
When Bowlby was originally developing his theory of attachment, there were alternative theoretical perspectives on why infants were emotionally attached to their primary caregivers (most often, their biological mothers). Bowlby and other theorists, for example, believed that there was something important about the responsiveness and contact provided by mothers. Other theorists, in contrast, argued that young infants feel emotionally connected to their mothers because mothers satisfy more basic needs, such as the need for food. That is, the child comes to feel emotionally connected to the mother because she is associated with the reduction of primary drives, such as hunger, rather than the reduction of drives that might be relational in nature.
In a classic set of studies, psychologist Harry Harlow placed young monkeys in cages that contained two artificial, surrogate “mothers” (Harlow, 1958). One of those surrogates was a simple wire contraption; the other was a wire contraption covered in cloth. Both of the surrogate mothers were equipped with a feeding tube so that Harrow and his colleagues had the option to allow the surrogate to deliver or not deliver milk. Harlow found that the young macaques spent a disproportionate amount of time with the cloth surrogate as opposed to the wire surrogate. Moreover, this was true even when the infants were fed by the wire surrogate rather than the cloth surrogate. This suggests that the strong emotional bond that infants form with their primary caregivers is rooted in something more than whether the caregiver provides food per se. Harlow’s research is now regarded as one of the first experimental demonstrations of the importance of “contact comfort” in the establishment of infant– caregiver bonds. ■
Drawing on evolutionary theory, Bowlby (1969) argued that these behaviors are adaptive responses to separation from a primary attachment figure—a caregiver who provides support, protection, and care. Because human infants, like other mammalian infants, cannot feed or protect them- selves, they are dependent upon the care and protection of “older and wiser” adults for survival. Bowlby argued that, over the course of evolutionary history, infants who were able to maintain proximity to an attachment figure would be more likely to survive to a reproductive age.
According to Bowlby, a motivational system, what he called the attachment behavioral system, was gradually “designed” by natural selection to regulate proximity to an attachment figure. The attachment system functions much like a thermostat that continuously monitors the ambient temperature of a room, comparing that temperature against a desired state and adjusting behavior (e.g., activating the furnace) accordingly. In the case of the attachment system, Bowlby argued that the system continuously monitors the accessibility of the primary attachment figure. If the child perceives the attachment figure to be nearby, accessible, and attentive, then the child feels loved, secure, and confident and, behaviorally, is likely to explore his or her environment, play with others, and be sociable. If, however, the child perceives the attachment figure to be inaccessible, the child experiences anxiety and, behaviorally, is likely to exhibit attachment behaviors ranging from simple visual searching on the low extreme to active searching, following, and vocal signaling on the other. These attachment behaviors continue either until the child is able to reestablish a desirable level of physical or psychological proximity to the attachment figure or until the child exhausts himself or herself or gives up, as may happen in the context of a prolonged separation or loss. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/29%3A_Attachment_Through_the_Life_Course/29.01%3A_Introduction.txt |
Although Bowlby believed that these basic dynamics captured the way the attachment system works in most children, he recognized that there are individual differences in the way children appraise the accessibility of the attachment figure and how they regulate their attachment behavior in response to threats. However, it was not until his colleague, Mary Ainsworth, began to systematically study infant–parent separations that a formal understanding of these individual differences emerged. Ainsworth and her students developed a technique called the strange situation—a laboratory task for studying infant–parent attachment (Ainsworth et al., 1978). In the strange situation, 12-month-old infants and their parents are brought to the laboratory and, over a period of approximately 20 minutes, are systematically separated from and reunited with one another. In the strange situation, most children (about 60%) behave in the way implied by Bowlby’s normative theory. Specifically, they become upset when the parent leaves the room, but, when he or she returns, they actively seek the parent and are easily comforted by him or her. Children who exhibit this pattern of behavior are often called secure. Other children (about 20% or less) are ill at ease initially and, upon separation, become extremely distressed. Importantly, when reunited with their parents, these children have a difficult time being soothed and often exhibit conflicting behaviors that suggest they want to be comforted, but that they also want to “punish” the parent for leaving. These children are often called anxious-resistant. The third pattern of attachment that Ainsworth and her colleagues documented is often labeled avoidant. Avoidant children (about 20%) do not consistently behave as if they are stressed by the separation but, upon reunion, actively avoid seeking contact with their parent, sometimes turning their attention to play objects on the laboratory floor.
Ainsworth’s work was important for at least three reasons. First, she provided one of the first empirical demonstrations of how attachment behavior is organized in unfamiliar con- texts. Second, she provided the first empirical taxonomy of individual differences in infant attachment patterns. According to her research, at least three types of children exist: those who are secure in their relationship with their parents, those who are anxious-resistant, and those who are avoidant. Finally, she demonstrated that these individual differences were correlated with infant–parent interactions in the home during the first year of life. Children who appear secure in the strange situation, for example, tend to have parents who are responsive to their needs. Children who appear insecure in the strange situation (i.e., anxious-resistant or avoidant) often have parents who are insensitive to their needs or inconsistent or rejecting in the care they provide.
29.04: Antecedents Of Attachment Patterns
In the years that have followed Ainsworth’s ground-breaking research, researchers have investigated a variety of factors that may help determine whether children develop secure or insecure relationships with their primary attachment figures. As mentioned above, one of the key determinants of attachment patterns is the history of sensitive and responsive interactions between the caregiver and the child. In short, when the child is uncertain or stressed, the ability of the caregiver to provide support to the child is critical for his or her psychological development. It is assumed that such supportive interactions help the child learn to regulate his or her emotions, give the child the confidence to explore the environment, and provide the child with a safe haven during stressful circumstances.
Evidence for the role of sensitive caregiving in shaping attachment patterns comes from longitudinal and experimental studies. For example, Grossmann et al. (1985) studied parent–child interactions in the homes of 54 families up to three times during the first year of the child’s life. At 12 months of age, infants and their mothers participated in the strange situation. Grossmann and her colleagues found that children who were classified as secure in the strange situation at 12 months of age were more likely than children classified as insecure to have mothers who provided responsive care to their children in the home environment.
Dymphna van den Boom (1994) developed an intervention that was designed to enhance maternal sensitive responsiveness. When their infants were 9 months of age, the mothers in the intervention group were rated as more responsive and attentive in their interaction with the infants compared to mothers in the control group. In addition, their infants were rated as more sociable, more self-soothing, and more likely to explore the environment. At 12 months of age, children in the intervention group were more likely to be classified as secure than insecure in the strange situation.
29.05: Attachment Patterns And Child Outcomes
Attachment researchers have studied the association between children’s attachment patterns and their adaptation over time. Researchers have learned, for example, that children who are classified as secure in the strange situation are more likely to have high functioning relationships with peers, to be evaluated favorably by teachers, and to persist with more diligence in challenging tasks. In contrast, insecure-avoidant children are more likely to be construed as “bullies” or to have a difficult time building and maintaining friendships (Weinfield et al., 2008). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/29%3A_Attachment_Through_the_Life_Course/29.03%3A_Individual_Differences_In_Infant_Attachment.txt |
Although Bowlby was primarily focused on understanding the nature of the infant–caregiver relationship, he believed that attachment characterized human experience across the life course. It was not until the mid-1980s, however, that researchers began to take seriously the possibility that attachment processes may be relevant to adulthood. Hazan and Shaver (1987) were two of the first researchers to explore Bowlby’s ideas in the context of romantic relationships. According to Hazan and Shaver, the emotional bond that develops between adult romantic partners is partly a function of the same motivational system—the attachment behavioral system—that gives rise to the emotional bond between infants and their caregivers. Hazan and Shaver noted that in both kinds of relationship, people (a) feel safe and secure when the other person is present; (b) turn to the other person during times of sickness, distress, or fear; (c) use the other person as a “secure base” from which to explore the world; and (d) speak to one another in a unique language, often called “motherese” or “baby talk.” (See Focus Topic 2.)
On the basis of these parallels, Hazan and Shaver (1987) argued that adult romantic relationships, similar to infant– caregiver relationships, are attachments. According to Hazan and Shaver, individuals gradually transfer attachment-related functions from parents to peers as they develop. Thus, although young children tend to use their parents as their primary attachment figures, as they reach adolescence and young adulthood, they come to rely more upon close friends and/or romantic partners for basic attachment-related functions. Thus, although a young child may turn to his or her mother for comfort, support, and guidance when distressed, scared, or ill, young adults may be more likely to turn to their romantic partners for these purposes under similar situations.
Hazan and Shaver (1987) asked a diverse sample of adults to read the three paragraphs below and indicate which paragraph best characterized the way they think, feel, and behave in close relationships:
1. I am somewhat uncomfortable being close to others; I find it difficult to trust them completely, difficult to allow myself to depend on them. I am nervous when anyone gets too close, and often, others want me to be more intimate than I feel comfortable being.
2. I find it relatively easy to get close to others and am comfortable depending on them and having them depend on me. I don’t worry about being abandoned or about some- one getting too close to me.
3. I find that others are reluctant to get as close as I would like. I often worry that my partner doesn’t really love me or won’t want to stay with me. I want to get very close to my partner, and this sometimes scares people away.
Conceptually, these descriptions were designed to represent what Hazan and Shaver considered to be adult analogues of the kinds of attachment patterns Ainsworth described in the strange situation (avoidant, secure, and anxious, respectively). Hazan and Shaver (1987) found that the distribution of the three patterns was similar to that observed in infancy. In other words, about 60% of adults classified themselves as secure (paragraph B), about 20% described themselves as avoidant (paragraph A), and about 20% described themselves as anxious-resistant (paragraph C). Moreover, they found that people who described themselves as secure, for example, were more likely to report having had warm and trusting relation- ships with their parents when they were growing up. In addition, they were more likely to have positive views of romantic relationships. Based on these findings, Hazan and Shaver (1987) concluded that the same kinds of individual differences that exist in infant attachment also exist in adulthood.
FOCUS TOPIC 2
Attachment and Social Media
Social media websites and mobile communication services play an increasing role in people’s lives. Many people use Facebook, for example, to keep in touch with family and friends, to update their loved ones regarding things going on in their lives, and to meet people who share similar interests. Moreover, modern cellular technology allows people to get in touch with their loved ones much easier than was possible a mere 20 years ago.
From an attachment perspective, these innovations in communications technology are important because they allow people to stay connected virtually to their attachment figures—regardless of the physical distance that might exist between them. Recent research has begun to examine how attachment processes play out in the use of social media. Oldmeadow et al. (2013), for example, studied a diverse sample of individuals and assessed their attachment security and their use of Facebook. Oldmeadow and his colleagues found that the use of Facebook may serve attachment functions. For example, people were more likely to report using Facebook to connect with others when they were experiencing negative emotions. In addition, the researchers found that people who were more anxious in their attachment orientation were more likely to use Face- book frequently, but people who were more avoidant used Facebook less and were less open on the site. ■
Research on Attachment in Adulthood
Attachment theory has inspired a large amount of literature in social, personality, and clinical psychology. In the sections below, I provide a brief overview of some of the major research questions and what researchers have learned about attachment in adulthood.
Who Ends Up with Whom?
When people are asked what kinds of psychological or behavioral qualities they are seeking in a romantic partner, a large majority of people indicate that they are seeking someone who is kind, caring, trustworthy, and understanding—the kinds of attributes that characterize a “secure” caregiver (Chappell & Davis, 1998). But we know that people do not always end up with others who meet their ideals. Are secure people more likely to end up with secure partners—and, vice versa, are insecure people more likely to end up with insecure partners? The majority of the research that has been con- ducted to date suggests that the answer is “yes.” Frazier et al. (1996), for example, studied the attachment patterns of more than 83 heterosexual couples and found that, if the man was relatively secure, the woman was also likely to be secure.
One important question is whether these findings exist because (a) secure people are more likely to be attracted to other secure people, (b) secure people are likely to create security in their partners over time, or (c) some combination of these possibilities. Existing empirical research strongly supports the first alternative. For example, when people have the opportunity to interact with individuals who vary in security in a speed-dating context, they express a greater interest in those who are higher in security than those who are more insecure (McClure et al., 2010). However, there is also some evidence that people’s attachment styles mutually shape one another in close relationships. For example, in a longitudinal study, Hudson et al. (2014) found that, if one person in a relationship experienced a change in security, his or her partner was likely to experience a change in the same direction.
Relationship Functioning
Research has consistently demonstrated that individuals who are relatively secure are more likely than insecure individu- als to have high functioning relationships—relationships that are more satisfying, more enduring, and less characterized by conflict. For example, Feeney and Noller (1992) found that insecure individuals were more likely than secure individu- als to experience a breakup of their relationship. In addition, secure individuals are more likely to report satisfying rela- tionships (e.g., Collins & Read, 1990) and are more likely to provide support to their partners when their partners are feel- ing distressed (Simpson et al., 1992).
Do Early Experiences Shape Adult Attachment?
The majority of research on this issue is retrospective—that is, it relies on adults’ reports of what they recall about their childhood experiences. This kind of work suggests that secure adults are more likely to describe their early childhood experiences with their parents as being supportive, loving, and kind (Hazan & Shaver, 1987). A number of longitudinal studies are emerging that demonstrate prospective associations between early attachment experiences and adult attachment styles and/or interpersonal functioning in adulthood. For example, Fraley et al. (2013) found in a sample of more than 700 indi- viduals studied from infancy to adulthood that maternal sen- sitivity across development prospectively predicted security at age 18. Simpson et al. (2007) found that attachment security, assessed in infancy in the strange situation, predicted peer competence in grades 1 to 3, which, in turn, predicted the quality of friendship relationships at age 16, which, in turn, predicted the expression of positive and negative emotions in their adult romantic relationships at ages 20 to 23.
It is easy to come away from such findings with the mis- taken assumption that early experiences “determine” later outcomes. To be clear: Attachment theorists assume that the relationship between early experiences and subsequent out- comes is probabilistic, not deterministic. Having support- ive and responsive experiences with caregivers early in life is assumed to set the stage for positive social development.
But that does not mean that attachment patterns are set in stone. In short, even if an individual has far from optimal experiences in early life, attachment theory suggests that it is possible for that individual to develop well-functioning adult relationships through a number of corrective experi- ences—including relationships with siblings, other family members, teachers, and close friends. Security is best viewed as a culmination of a person’s attachment history rather than a reflection of his or her early experiences alone. Those early experiences are considered important not because they deter- mine a person’s fate, but because they provide the foundation for subsequent experiences.
REFERENCES
Ainsworth, M. D. S., Blehar, M. C., Waters, E., & Wall, S. (1978). Patterns of attachment. Erlbaum.
Bowlby, J. (1969). Attachment and loss: Vol. 1. Attachment. Basic Books. Chappell, K. D., & Davis, K. E. (1998). Attachment, partner choice, and
perception of romantic partners: An experimental test of the attachment-security hypothesis. Personal Relationships, 5(3), 327–342. doi.org/10.1111/j.1475-6811.1998.tb00175.x
Collins, N., & Read, S. (1990). Adult attachment, working models and relationship quality in dating couples. Journal of Personality and Social Psychology, 58(4), 644–663. doi.org/10.1037/ 0022-3514.58.4.644
Feeney, J. A., & Noller, P. (1992). Attachment style and romantic love: Relationship dissolution. Australian Journal of Psychology, 44(2), 69–74. doi.org/10.1080/00049539208260145
Fraley, R. C., Roisman, G. I., Booth-LaForce, C., Owen, M. T., & Holland, A. S. (2013). Interpersonal and genetic origins of adult attachment styles: A longitudinal study from infancy to early adulthood. Journal of Personality and Social Psychology, 104(5), 817–838. doi. org/10.1037/a0031435
Frazier, P. A., Byer, A. L., Fischer, A. R., Wright, D. M., & DeBord, K. A. (1996). Adult attachment style and partner choice: Correlational and experimental findings. Personal Relationships, 3(2), 117–136. https:// doi.org/10.1111/j.1475-6811.1996.tb00107.x
Grossmann, K., Grossmann, K. E., Spangler, G., Suess, G., & Unzner, L. (1985). Maternal sensitivity and newborns orientation responses as related to quality of attachment in northern Germany. Monographs of the Society for Research in Child Development, 50(1–2), 233–256. doi.org/10.2307/3333836
Harlow, H. F. (1958). The nature of love. American Psychologist, 13(12), 673–685. https://doi.org/10.1037/h0047884
Hazan, C., & Shaver, P. R. (1987). Romantic love conceptualized as an attachment process. Journal of Personality and Social Psychology, 52(3), 511–524. https://doi.org/10.1037/0022-3514.52.3.511
Hudson, N. W., Fraley, R. C., Brumbaugh, C. C., & Vicary, A. M. (2014). Coregulation in romantic partners’ attachment styles: A longitudinal investigation. Personality and Social Psychology Bulletin, 40(7), 845–857. https://doi.org/10.1177/0146167214528989
McClure, M. J., Lydon., J. E., Baccus, J. R., & Baldwin, M. W. (2010). A signal detection analysis of chronic attachment anxiety at speed- dating: Being unpopular is only the first part of the problem. Personality and Social Psychology Bulletin, 36(8), 1024–1036. https:// doi.org/10.1177/0146167210374238
Oldmeadow, J. A., Quinn, S., & Kowert, R. (2013). Attachment style, social skills, and Facebook use amongst adults. Computers in Human Behavior, 29(3), 1142–1149. https://doi.org/10.1016/j.chb.2012.10.006
Simpson, J. A., Collins, W. A., Tran, S., & Haydon, K. C. (2007). Attachment and the experience and expression of emotions in romantic relationships: A developmental perspective. Journal of Personality and Social Psychology, 92(2), 355–367. doi.org/10.1037/ 0022-3514.92.2.355
Simpson, J. A., Rholes, W. S., & Nelligan, J. S. (1992). Support seeking and support giving within couples in an anxiety-provoking situation: The role of attachment styles. Journal of Personality and Social Psychology, 62(3), 434–446. doi.org/10.1037/ 0022-3514.62.3.434
van den Boom, D. C. (1994). The influence of temperament and mothering on attachment and exploration: An experimental manipulation of sensitive responsiveness among lower-class mothers with irritable infants. Child Development, 65(5), 1457–1477. https:// doi.org/10.2307/1131511
Weinfield, N. S., Sroufe, L. A., Egeland, B., & Carlson, E. A. (2008). Individual differences in infant-caregiver attachment: Conceptual and empirical aspects of security. In J. Cassidy & P. R. Shaver (Eds.), Handbook of attachment: Theory, research, and clinical applications (2nd ed., pp. 78–101). Guilford Press. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/29%3A_Attachment_Through_the_Life_Course/29.06%3A_Attachment_In_Adulthood.txt |
What do infants know about the world in which they live—and how do they grow and change with age? These are the kinds of questions answered by developmental scientists. This module describes different research techniques that are used to study psychological phenomena in infants and children, research designs that are used to examine age-related changes in development, and unique challenges and special issues associated with conducting research with infants and children. Child development is a fascinating field of study, and many interesting questions remain to be examined by future generations of developmental scientists— maybe you will be among them!
A group of children were playing hide-and-seek in the yard. Pilar raced to her hiding spot as her 6-year-old cousin, Lucas, loudly counted, “. . . six, seven, eight, nine, ten! Ready or not, here I come!” Pilar let out a small giggle as Lucas ran over to find her—in the exact location where he had found his sister a short time before. At first glance, this behavior is puzzling: why would Pilar hide in exactly the same location where someone else was just found? Whereas older children and adults realize that it is likely best to hide in locations that have not been searched previously, young children do not have the same cognitive sophistication. But why not . . . and when do these abilities first develop?
Developmental psychologists investigate questions like these using research methods that are tailored to the particular capabilities of the infants and children being studied. Importantly, research in developmental psychology is more than simply examining how children behave during games of hide-and-seek—the results obtained from developmental research have been used to inform best practices in parenting, education, and policy.
This module describes different research techniques that are used to study psychological phenomena in infants and children, research designs that are used to examine age-related changes in developmental processes and changes over time, and unique challenges and special issues associated with con- ducting research with infants and children.
30.02: Research Methods
Infants and children—especially younger children—cannot be studied using the same research methods used in studies with adults. Researchers, therefore, have developed many creative ways to collect information about infant and child development. In this section, we highlight some of the methods that have been used by researchers who study infants and older children, separating them into three distinct categories: involuntary or obligatory responses, voluntary responses, and psychophysiological responses. We will also discuss other methods such as the use of surveys and questionnaires. At the end of this section, we give an example of how interview techniques can be used to study the beliefs and perceptions of older children and adults—a method that cannot be used with infants or very young children.
involuntary or Obligatory Responses
One of the primary challenges in studying very young infants is that they have limited motor control—they cannot hold their heads up for short amounts of time, much less grab an interesting toy, play the piano, or turn a door knob. As a result, infants cannot actively engage with the environment in the same way as older children and adults. For this reason, developmental scientists have designed research methods that assess involuntary or obligatory responses. These are behaviors in which people engage without much conscious thought or effort. For example, think about the last time you heard your name at a party—you likely turned your head to see who was talking without even thinking about it. Infants and young children also demonstrate involuntary responses to stimuli in the environment. When infants hear the voice of their mother, for instance, their heart rate increases—whereas if they hear the voice of a stranger, their heart rate decreases (Kisilevsky et al., 2003). Researchers study involuntary behaviors to better understand what infants know about the world around them.
One research method that capitalizes on involuntary or obligatory responses is a procedure known as habituation. In habituation studies, infants are presented with a stimulus such as a photograph of a face over and over again until they become bored with it. When infants become bored, they look away from the picture. If infants are then shown a new picture—such as a photograph of a different face—their interest returns and they look at the new picture. This is a phenomenon known as dishabituation. Habituation procedures work because infants generally look longer at novel stimuli relative to items that are familiar to them. This research technique takes advantage of involuntary or obligatory responses because infants are constantly looking around and observing their environments; they do not have to be taught to engage with the world in this way.
One classic habituation study was conducted by Baillargeon and colleagues (1985). These researchers were interested in the concept of object permanence, or the understanding that objects exist even when they cannot be seen or heard. For example, you know your toothbrush exists even though you are probably not able to see it right this second. To investigate object permanence in 5-month-old infants, the researchers used a violation of expectation paradigm. The researchers first habituated infants to an opaque screen that moved back and forth like a drawbridge (using the same procedure you just learned about in the previous paragraph). Once the infants were bored with the moving screen, they were shown two different scenarios to test their understanding of physical events. In both of these test scenarios, an opaque box was placed behind the moving screen. What differed between these two scenarios, however, was whether they confirmed or violated the solidity principle—the idea that two solid objects cannot occupy the same space at the same time. In the possible scenario, infants watched as the moving drawbridge stopped when it hit the opaque box (as would be expected based on the solidity principle). In the impossible scenario, the drawbridge appeared to move right through the space that was occupied by the opaque box! This impossible scenario violates the solidity principle in the same way as if you got out of your chair and walked through a wall, reappearing on the other side.
The results of this study revealed that infants looked longer at the impossible test event than at the possible test event. The authors suggested that the infants reacted in this way because they were surprised—the demonstration went against their expectation that two solids cannot move through one another. The findings indicated that 5-month-old infants understood that the box continued to exist even when they could not see it. Subsequent studies indicated that 31⁄2- and 41⁄2-month-old infants also demonstrate object permanence under similar test conditions (Baillargeon, 1987). These findings are notable because they suggest that infants understand object permanence much earlier than had been reported pre- viously in research examining voluntary responses (although see more recent research by Cashon & Cohen, 2000).
Voluntary Responses
As infants and children age, researchers are increasingly able to study their understanding of the world through their voluntary responses. Voluntary responses are behaviors that a person completes by choice. For example, think about how you act when you go to the grocery store: you select whether to use a shopping cart or a basket, you decide which sections of the store to walk through, and you choose whether to stick to your grocery list or splurge on a treat. Importantly, these behaviors are completely up to you (and are under your control). Although they do not do a lot of grocery shopping, infants and children also have voluntary control over their actions. Children, for instance, choose which toys to play with.
Researchers study the voluntary responses of infants and young children in many ways. For example, developmental scientists study recall memory in infants and young children by looking at voluntary responses. Recall memory is memory of past events or episodes, such as what you did yesterday afternoon or on your last birthday. Whereas older children and adults are simply asked to talk about their past experiences, recall memory has to be studied in a different way in infants and very young children who cannot discuss the past using language. To study memory in these subjects, researchers use a behavioral method known as elicited imitation (Lukowski & Milojevich, 2016).
In the elicited imitation procedure, infants play with toys that are designed in the lab to be unlike the kinds of things infants usually have at home. These toys (or event sequences, as researchers call them) can be put together in a certain way to produce an outcome that infants commonly enjoy. One of these events is called Find the Surprise. As shown in Figure \(4\), this toy has a door on the front that is held in place by a latch—and a small plastic figure is hidden on the inside. During the first part of the study, infants play with the toy in whichever way they want for a few minutes. The researcher then shows the infant how make the toy work by (1) flipping the latch out of the way and (2) opening the door, revealing the plastic toy inside. The infant is allowed to play with the toy again either immediately after the demonstration or after a longer delay. As the infant plays, the researcher records whether the infant finds the surprise using the same procedure that was demonstrated.
Use of the elicited imitation procedure has taught developmental scientists a lot about how recall memory develops. For example, we now know that 6-month-old infants remember one step of a 3-step sequence for 24 hours (Barr et al., 1996; Collie & Hayne, 1999). Nine-month-olds remember the individual steps that make up a 2-step event sequence for 1 month, but only 50% of infants remember to do the first step of the sequence before the second (Bauer et al., 2001, 2003; Carver & Bauer, 1999). When children are 20 months old, they remember the individual steps and temporal order of 4-step events for at least 12 months—the longest delay that has been tested to date (Bauer et al., 2000).
Psychophysiology
Behavioral studies have taught us important information about what infants and children know about the world. Research on behavior alone, however, cannot tell scientists how brain development or biological changes impact (or are impacted by) behavior. For this reason, researchers may also record data about psychophysiological responses, such as heart rate, hormone levels, or brain activity. These measures may be recorded by themselves or in combination with behavioral data to better understand the bidirectional relations between biology and behavior.
One manner of understanding associations between brain development and behavioral advances is through the recording of event-related potentials, or ERPs. ERPs are recorded by fitting a research participant with a stretchy cap that contains many small sensors or electrodes. These electrodes record tiny electrical currents on the scalp of the participant in response to the presentation of particular stimuli, such as a picture or a sound (for additional information on recording ERPs from infants and children, see DeBoer et al., 2005). The recorded responses are then amplified thousands of times using specialized equipment so that they look like squiggly lines with peaks and valleys. Some of these brain responses have been linked to psychological phenomena. For example, researchers have identified a negative peak in the recorded waveform that they have called the N170 (Bentin et al., 2010). The peak is named in this way because it is negative (hence the N) and because it occurs about 140 ms to 170 ms after a stimulus is presented (hence the 170). This peak is particularly sensitive to the presentation of faces, as it is commonly more negative when participants are presented with photo- graphs of faces rather than with photographs of objects. In this way, researchers are able to identify brain activity associated with real-world thinking and behavior.
The use of ERPs has provided important insight as to how infants and children understand the world around them. In one study (Webb et al., 2006), researchers examined face and object processing in children with autism spectrum disorders, those with developmental delays, and those who were typically developing. The children wore electrode caps and had their brain activity recorded as they watched still photographs of faces (of their mother or of a stranger) and objects (including those that were familiar or unfamiliar to them). The researchers examined differences in face and object processing by group by observing a component of the brainwave they called the prN170 (because it was believed to be a precursor to the adult N170). Their results showed that the height of the prN170 peak (commonly called the amplitude) did not differ when faces or objects were presented to typically developing children. When considering children with autism, however, the peaks were higher when objects were presented relative to when faces were shown. Differences were also found in how long it took the brain to reach the negative peak (commonly called the latency of the response). Whereas the peak was reached more quickly when typically developing children were presented with faces relative to objects, the opposite was true for children with autism. These findings suggest that children with autism are in some way processing faces differently than typically developing children (and, as reported in the manuscript, children with more general developmental delays).
Parent-Report Questionnaires
Developmental science has come a long way in assessing various aspects of infant and child development through behavior and psychophysiology—and new advances are happening every day. In many ways, however, the very youngest of research participants are still quite limited in the information they can provide about their own development. As such, researchers often ask the people who know infants and children best—commonly, their parents or guardians—to complete surveys or questionnaires about various aspects of their lives. These parent-report data can be analyzed by themselves or in combination with any collected behavioral or psycho- physiological data.
One commonly used parent-report questionnaire is the Child Behavior Checklist (CBCL) (Achenbach & Rescorla, 2000). Parents complete the preschooler version of this questionnaire by answering questions about child strengths, behavior problems, and disabilities, among other things. The responses provided by parents are used to identify whether the child has any behavioral issues, such as sleep difficulties, aggressive behaviors, depression, or attention deficit/hyper- activity problems.
A recent study used the CBCL-Preschool questionnaire (Achenbach & Rescorla, 2000) to examine preschooler functioning in relation to levels of stress experienced by their mothers while they were pregnant (Ronald et al., 2011). Almost 3,000 pregnant women were recruited into the study during their pregnancy and were interviewed about their stressful life experiences. Later, when their children were 2 years old, mothers completed the CBCL-Preschool questionnaire. The results of the study showed that higher levels of maternal stress during pregnancy (such as a divorce or moving to a new house) were associated with increased attention deficit/ hyperactivity problems in children over 2 years later. These findings suggest that stressful events experienced during prenatal development may be associated with problematic child behavioral functioning years later—although additional research is needed.
Interview Techniques
Whereas infants and very young children are unable to talk about their own thoughts and behaviors, older children and adults are commonly asked to use language to discuss their thoughts and knowledge about the world. In fact, these verbal report paradigms are among the most widely used in psychological research. For instance, a researcher might present a child with a vignette or short story describing a moral dilemma, and the child would be asked to give their own thoughts and beliefs (Walrath, 2011). For example, children might react to the following:
Mr. Kohut’s wife is sick and only one medication can save her life. The medicine is extremely expensive and Mr. Kohut cannot afford it. The druggist will not lower the price. What should Mr. Kohut do, and why?
Children can provide written or verbal answers to these types of scenarios. They can also offer their perspectives on issues ranging from attitudes toward drug use to the experience of fear while falling asleep to their memories of get- ting lost in public places—the possibilities are endless. Verbal reports such as interviews and surveys allow children to describe their own experience of the world. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/30%3A_Research_Methods_in_Developmental_Psychology/30.01%3A_Introduction.txt |
Now you know about some tools used to conduct research with infants and young children. Remember, research methods are the tools that are used to collect information. But it is easy to confuse research methods and research design. Research design is the strategy or blueprint for deciding how to collect and analyze information. Research design dictates which methods are used and how.
Researchers typically focus on two distinct types of comparisons when conducting research with infants and children. The first kind of comparison examines change within individuals. As the name suggests, this type of analysis measures the ways in which a specific person changes (or remains the same) over time. For example, a developmental scientist might be interested in studying the same group of infants at 12 months, 18 months, and 24 months to examine how vocabulary and grammar change over time. This kind of question would be best answered using a longitudinal research design. Another sort of comparison focuses on changes between groups. In this type of analysis, researchers study average changes in behavior between groups of different ages. Returning to the language example, a scientist might study the vocabulary and grammar used by 12-month-olds, 18-month-olds, and 24-month-olds to examine how language abilities change with age. This kind of question would be best answered using a cross-sectional research design.
Longitudinal Research Designs
Longitudinal research designs are used to examine behavior in the same infants and children over time. For example, when considering our example of hide-and-seek behaviors in preschoolers, a researcher might conduct a longitudinal study to examine whether 2-year-olds develop into better hiders over time. To this end, a researcher might observe a group of 2-year-old children playing hide-and-seek with plans to observe them again when they are 4 years old—and again when they are 6 years old. This study is longitudinal in nature because the researcher plans to study the same children as they age. Based on her data, the researcher might conclude that 2-year-olds develop more mature hiding abilities with age. Remember, researchers examine games such as hide- and-seek not because they are interested in the games them- selves, but because they offer clues to how children think, feel, and behave at various ages.
Longitudinal studies may be conducted over the short term (over a span of months, as in Wiebe et al., 2010) or over much longer durations (years or decades, as in Lukowski et al., 2010). For these reasons, longitudinal research designs are optimal for studying stability and change over time. Longitudinal research also has limitations, however. For one, longitudinal studies are expensive: they require that researchers maintain continued contact with participants over time, and they necessitate that scientists have funding to conduct their work over extended durations (from infancy to when participants were 19 years old in Lukowski et al., 2010). An additional risk is attrition. Attrition occurs when participants fail to complete all portions of a study. Participants may move, change their phone numbers, or simply become disinterested in participating over time. Researchers should account for the possibility of attrition by enrolling a larger sample into their study initially, as some participants will likely drop out over time.
The results from longitudinal studies may also be impacted by repeated assessments. Consider how well you would do on a math test if you were given the exact same exam every day for a week. Your performance would likely improve over time—not necessarily because you developed better math abilities, but because you were continuously practicing the same math problems. This phenomenon is known as a practice effect. Practice effects occur when participants become better at a task over time because they have done it again and again; not due to natural psychological development. A final limitation of longitudinal research is that the results may be impacted by cohort effects. Cohort effects occur when the results of the study are affected by the particular point in historical time during which participants are tested. As an example, think about how peer relationships in childhood have likely changed since February 2004—the month and year Facebook was founded. Cohort effects can be problematic in longitudinal research because only one group of participants are tested at one point in time—different findings might be expected if participants of the same ages were tested at different points in historical time.
Cross-Sectional Designs
Cross-sectional research designs are used to examine behavior in participants of different ages who are tested at the same point in time. When considering our example of hide-and- seek behaviors in children, for example, a researcher might want to examine whether older children more often hide in novel locations (those in which another child in the same game has never hidden before) when compared to younger children. In this case, the researcher might observe 2-, 4-, and 6-year-old children as they play the game (the various age groups represent the “cross sections”). This research is cross-sectional in nature because the researcher plans to examine the behavior of children of different ages within the same study at the same time. Based on her data, the researcher might conclude that 2-year-olds more commonly hide in previously searched locations relative to 6-year-olds.
Cross-sectional designs are useful for many reasons. Because participants of different ages are tested at the same point in time, data collection can proceed at a rapid pace. In addition, because participants are only tested at one point in time, practice effects are not an issue—children do not have the opportunity to become better at the task over time. Cross-sectional designs are also more cost-effective than longitudinal research designs because there is no need to maintain contact with and follow-up on participants over time.
One of the primary limitations of cross-sectional research, however, is that the results yield information on age-related change, not development per se. That is, although the study described above can show that 6-year-olds are more advanced in their hiding behavior than 2-year-olds, the data used to come up with this conclusion were collected from different children. It could be, for instance, that this specific sample of 6-year-olds just happened to be particularly clever at hide-and-seek. As such, the researcher cannot conclude that 2-year-olds develop into better hiders with age; she can only state that 6-year-olds, on average, are more sophisticated hiders relative to children 4 years younger.
Sequential Research Designs
Sequential research designs include elements of both longitudinal and cross-sectional research designs. Similar to longitudinal designs, sequential research features participants who are followed over time; similar to cross-sectional designs, sequential work includes participants of different ages. This research design is also distinct from those that have been discussed previously in that children of different ages are enrolled into a study at various points in time to examine age-related changes, assess development within the same individuals as they age, and account for the possibility of cohort effects.
Consider, once again, our example of hide-and-seek behaviors. In a study with a sequential design, a researcher might enroll three separate groups of children (Groups A, B, and C). Children in Group A would be enrolled when they are 2 years old and would be tested again when they are 4 and 6 years old (similar in design to the longitudinal study described previously). Children in Group B would be enrolled when they are 4 years old and would be tested again when they are 6 and 8 years old. Finally, children in Group. C would be enrolled when they are 6 years old and would be tested again when they are 8 and 10 years old.
Studies with sequential designs are powerful because they allow for both longitudinal and cross-sectional comparisons. This research design also allows for the examination of cohort effects. For example, the researcher could examine the hide- and-seek behavior of 6-year-olds in Groups A, B, and C to determine whether performance differed by group when participants were the same age. If performance differences were found, there would be evidence for a cohort effect. In the hide-and-seek example, this might mean that children from different time periods varied in the amount they giggled or how patient they are when waiting to be found. Sequential designs are also appealing because they allow researchers to learn a lot about development in a relatively short amount of time. In the previous example, a 4-year research study would provide information about 8 years of developmental time by enrolling children ranging in age from 2 to 10 years old.
Because they include elements of longitudinal and cross-sectional designs, sequential research has many of the same strengths and limitations as these other approaches. For example, sequential work may require less time and effort than longitudinal research, but more time and effort than cross-sectional research. Although practice effects may be an issue if participants are asked to complete the same tasks or assessments over time, attrition may be less problematic than what is commonly experienced in longitudinal research since participants may not have to remain involved in the study for such a long period of time.
When considering the best research design to use in their research, scientists think about their main research question and the best way to come up with an answer. A table of advantages and disadvantages for each of the described research designs is provided here to help you as you consider what sorts of studies would be best conducted using each of these different approaches (Table \(1\)).
Table \(1\):Advantages and Disadvantages of Different Research Designs
Research Design
Disadvantages
Disadvantages
Longitudinal
• Examines changes within individuals over time
• Provides a developmental analysis
• Expensive
• Takes a long time
• Participant attrition
• Possibility of practice effects
• Cannot examine cohort effects
Cross-sectional
• Examines changes between participants of different ages at the same point in time
• Provides information on age-related change
• Cannot examine change over time
• Cannot examine cohort effects
Sequential
• Examines changes within individuals over time
• Examines changes between participants of different ages at the
• same point in time
• Can be used to examine cohort effects
• May be expensive
• Possibility of practice effects | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/30%3A_Research_Methods_in_Developmental_Psychology/30.03%3A_Research_Design.txt |
The previous sections describe research tools to assess development in infancy and early childhood, as well as the ways that research designs can be used to track age-related changes and development over time. Before you begin con- ducting developmental research, however, you must also be aware that testing infants and children comes with its own unique set of challenges. In the final section of this module, we review some of the main issues that are encountered when conducting research with the youngest of human participants. In particular, we focus our discussion on ethical concerns, recruitment issues, and participant attrition.
Ethical Concerns
As a student of psychological science, you may already know that Institutional Review Boards (IRBs) review and approve of all research projects that are conducted at universities, hospitals, and other institutions. An IRB is typically a panel of experts who read and evaluate proposals for research. IRB members want to ensure that the proposed research will be carried out ethically and that the potential benefits of the research outweigh the risks and harm for participants. What you may not know though, is that the IRB considers some groups of participants to be more vulnerable or at-risk than others. Whereas university students are generally not viewed as vulnerable or at-risk, infants and young children commonly fall into this category. What makes infants and young children more vulnerable during research than young adults? One reason infants and young children are perceived as being at increased risk is due to their limited cognitive capabilities, which makes them unable to state their willingness to participate in research or tell researchers when they would like to drop out of a study. For these reasons, infants and young children require special accommodations as they participate in the research process.
When thinking about special accommodations in developmental research, consider the informed consent process. If you have ever participated in psychological research, you may know through your own experience that adults commonly sign an informed consent statement (a contract stating that they agree to participate in research) after learning about a study. As part of this process, participants are informed of the procedures to be used in the research, along with any expected risks or benefits. Infants and young children can- not verbally indicate their willingness to participate, much less understand the balance of potential risks and benefits. As such, researchers are oftentimes required to obtain writ- ten informed consent from the parent or legal guardian of the child participant, an adult who is almost always present as the study is conducted. In fact, children are not asked to indicate whether they would like to be involved in a study at all (a process known as assent) until they are approximately seven years old. Because infants and young children also cannot easily indicate if they would like to discontinue their participation in a study, researchers must be sensitive to changes in the state of the participant (determining whether a child is too tired or upset to continue) as well as to parent desires (in some cases, parents might want to discontinue their involvement in the research). As in adult studies, researchers must always strive to protect the rights and well-being of the minor participants and their parents when conducting developmental science.
Recruitment
An additional challenge in developmental science is participant recruitment. Recruiting university students to participate in adult studies is typically easy. Many colleges and universities offer extra credit for participation in research and have locations such as bulletin boards and school newspapers where research can be advertised. Unfortunately, young children cannot be recruited by making announcements in Introduction to Psychology courses, by posting ads on campuses, or through online platforms such as Amazon Mechanical Turk. Given these limitations, how do researchers go about finding infants and young children to be in their studies?
The answer to this question varies along multiple dimensions. Researchers must consider the number of participants they need and the financial resources available to them, among other things. Location may also be an important consideration. Researchers who need large numbers of infants and children may attempt to do so by obtaining infant birth records from the state, county, or province in which they reside. Some areas make this information publicly available for free, whereas birth records must be purchased in other areas (and in some locations birth records may be entirely unavailable as a recruitment tool). If birth records are available, researchers can use the obtained information to call families by phone or mail them letters describing possible research opportunities. All is not lost if this recruitment strategy is unavailable, however. Researchers can choose to pay a recruitment agency to contact and recruit families for them. Although these methods tend to be quick and effective, they can also be quite expensive. More economical recruitment options include posting advertisements and fliers in locations frequented by families, such as mommy-and-me classes, local malls, and preschools or day care centers. Researchers can also utilize online social media outlets like Facebook, which allows users to post recruitment advertisements for a small fee. Of course, each of these different recruitment techniques requires IRB approval.
Another important consideration when conducting research with infants and young children is attrition. Although attrition is quite common in longitudinal research in particular, it is also problematic in developmental science more generally, as studies with infants and young children tend to have higher attrition rates than studies with adults. For example, high attrition rates in ERP studies oftentimes result from the demands of the task: infants are required to sit still and have a tight, wet cap placed on their heads before watching still photographs on a computer screen in a dark, quiet room. In other cases, attrition may be due to motivation (or a lack thereof). Whereas adults may be motivated to participate in research in order to receive money or extra course credit, infants and young children are not as easily enticed. In addition, infants and young children are more likely to tire easily, become fussy, and lose interest in the study procedures than are adults. For these reasons, research studies should be designed to be as short as possible—it is likely better to break up a large study into multiple short sessions rather than cram all of the tasks into one long visit to the lab. Researchers should also allow time for breaks in their study protocols so that infants can rest or have snacks as needed. Happy, comfortable participants provide the best data. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/30%3A_Research_Methods_in_Developmental_Psychology/30.04%3A__Challenges_Associated_With_Conducting_Developmental_Research.txt |
Child development is a fascinating field of study—but care must be taken to ensure that researchers use appropriate methods to examine infant and child behavior, use the correct experimental design to answer their questions, and be aware of the special challenges that are part-and-parcel of develop- mental research. After reading this module, you should have a solid understanding of these various issues and be ready to think more critically about research questions that interest you. For example, when considering our initial example of hide-and-seek behaviors in preschoolers, you might ask questions about what other factors might contribute to hiding behaviors in children. Do children with older siblings hide in locations that were previously searched less often than children without siblings? What other abilities are associated with the development of hiding skills? Do children who use more sophisticated hiding strategies as preschoolers do better on other tests of cognitive functioning in high school? Many interesting questions remain to be examined by future generations of developmental scientists—maybe you will make one of the next big discoveries!
REFERENCES
Achenbach, T. M., & Rescorla, L. A. (2000). Manual for the ASEBA preschool forms and profiles: An integrated system of multi-informant assessment. University of Vermont Department of Psychiatry.
Baillargeon, R. (1987). Object permanence in 31⁄2- and 41⁄2-month-old infants. Developmental Psychology, 23(5), 655–664. doi. org/10.1037/0012-1649.23.5.655
Baillargeon, R., Spelke, E., & Wasserman, S. (1985). Object permanence in five-month-old infants. Cognition, 20(3), 191–208. doi. org/10.1016/0010-0277(85)90008-3
Barr, R., Dowden, A., & Hayne, H. (1996). Developmental changes in deferred imitation by 6- to 24-month-old infants. Infant Behavior and Development, 19(2), 159–170. doi.org/10.1016/ S0163-6383(96)90015-6
Bauer, P. J., Wenner, J. A., Dropik, P. L., Wewerka, S. S., & Howe, M. L. (2000). Parameters of remembering and forgetting in the transition from infancy to early childhood. Monographs of the Society for Research in Child Development, 65(4), 1–204. https://doi.org/ 10.1016/j.imlet.2014.04.001
Bauer, P. J., Wiebe, S. A., Carver, L. J., Waters, J. M., & Nelson, C. A. (2003). Developments in long-term explicit memory late in the first year of life: Behavioral and electrophysiological indices. Psychological Science, 14(6), 629–635. doi.org/10.1046/ j.0956-7976.2003.psci_1476.x
Bauer, P. J., Wiebe, S. A., Waters, J. M., & Bangston, S. K. (2001). Reexposure breeds recall: Effects of experience on 9-month-olds’ ordered recall. Journal of Experimental Child Psychology, 80(2), 174–200. doi.org/10.1006/jecp.2000.2628
Bentin, S., Allison, T., Puce, A., Perez, E., & McCarthy, G. (2010). Electrophysiological studies of face perception in humans. Journal of Cognitive Neuroscience, 8(6), 551–565. doi.org/10.1162/ jocn.1996.8.6.551
Carver, L. J., & Bauer, P. J. (1999). When the event is more than the sum of its parts: 9-month-olds’ long-term ordered recall. Memory, 7(2), 147–174. https://doi.org/10.1080/741944070
Cashon, C. H., & Cohen, L. B. (2000). Eight-month-old infants’ perception of possible and impossible events. Infancy, 1(4), 429–446. doi. org/10.1207/S15327078IN0104_4
Collie, R., & Hayne, H. (1999). Deferred imitation by 6- and 9-month-old infants: More evidence for declarative memory. Developmental Psychobiology, 35(2), 83–90. doi.org/10.1002/ (SICI)1098-2302(199909)35:2%3C83::AID-DEV1%3E3.0.CO;2-S
DeBoer, T., Scott, L. S., & Nelson, C. A. (2005). ERPs in developmental populations. In T. C. Handy (Ed.), Event-related potentials: A methods handbook (pp. 263–297). MIT Press.
Kisilevsky, B. S., Hains, S. M. J., Lee, K., Xie, X., Huang, H., Ye, H. H., Zhang, K., & Wang, Z. (2003). Effects of experience on fetal voice recognition. Psychological Science, 14(3), 220–224.
Lukowski, A. F., & Milojevich, H. M. (2013). Sleeping like a baby: Examining relations between habitual infant sleep, recall memory, and generalization across cues at 10 months. Infant Behavior and Development, 36(3), 369–376. doi.org/10.1016/j. infbeh.2013.02.001
Lukowski, A. F., & Milojevich, H. M. (2016). Examining recall memory in infancy and early childhood using the elicited imitation paradigm. Journal of Visualized Experiments, 110, e53347. https://doi.org/ 10.3791/53347
Lukowski, A. F., Koss, M., Burden, M. J., Jonides, J., Nelson, C. A., Kaciroti, N., Jimenez, E., & Lozoff, B. (2010). Iron deficiency in infancy and neurocognitive functioning at 19 years: Evidence of long-term deficits in executive function and recognition memory. Nutritional Neuroscience, 13(2), 54–70. doi.org/10.1179/1476830 10X12611460763689
Ronald, A., Pennell, C. E., & Whitehouse, A. J. O. (2011). Prenatal maternal stress associated with ADHD and autistic traits in early childhood. Frontiers in Psychology, 1, 1–8. doi.org/10.3389/ fpsyg.2010.00223
Walrath, R. (2011). Kohlberg’s theory of moral development. In S. Goldstein & J. A. Naglieri (Eds.), Encyclopedia of child behavior and development. Springer. https://doi.org/10.1007/978-0-387-79061-9
Webb, S. J., Dawson, G., Bernier, R., & Panagiotides, H. (2006). ERP evidence of atypical face processing in young children with autism. Journal of Autism and Developmental Disorders, 36(7), 884–890. https://doi.org/10.1007/s10803-006-0126-x
Wiebe, S. A., Lukowski, A. F., & Bauer, P. J. (2010). Sequence imitation and reaching measures of executive control: A longitudinal examination in the second year of life. Developmental Neuropsychology, 35(5), 522–538. doi.org/10.1080/87565641. 2010.494751 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/30%3A_Research_Methods_in_Developmental_Psychology/30.05%3A_Conclusion.txt |
Personality traits reflect people’s characteristic patterns of thoughts, feelings, and behaviors. Personality traits imply consistency and stability—someone who scores high on a specific trait like Extraversion is expected to be sociable in different situations and over time. Thus, trait psychology rests on the idea that people differ from one another in terms of where they stand on a set of basic trait dimensions that persist over time and across situations. The most widely used system of traits is called the Five-Factor Model. This system includes five broad traits that can be remembered with the acronym OCEAN: Openness, Conscientiousness, Extraversion, Agreeableness, and Neuroticism. Each of the major traits from the Big Five can be divided into facets to give a more fine- grained analysis of someone’s personality. In addition, some trait theorists argue that there are other traits that cannot be completely captured by the Five-Factor Model. Critics of the trait concept argue that people do not act consistently from one situation to the next and that people are very influenced by situational forces. Thus, one major debate in the field concerns the relative power of people’s traits versus the situations in which they find themselves as predictors of their behavior.
When we observe people around us, one of the first things that strikes us is how different people are from one another. Some people are very talkative, whereas others are very quiet. Some are active, whereas others are couch potatoes. Some worry a lot, others almost never seem anxious. Each time we use one of these words—words like “talkative,” “quiet,” “active,” or “anxious”—to describe those around us, we are talking about a person’s personality: the characteristic ways that people differ from one another. Personality psychologists try to describe and understand these differences.
Although there are many ways to think about the personalities that people have, Gordon Allport and other “personol- ogists” claimed that we can best understand the differences between individuals by understanding their personality traits. Personality traits reflect basic dimensions on which people differ (Matthews et al., 2003). According to trait psychologists, there are a limited number of these dimensions (such as Extraversion, Conscientiousness, or Agreeableness), and each individual falls somewhere on each dimension, meaning that they could be low, medium, or high on any specific trait.
An important feature of personality traits is that they reflect continuous distributions rather than distinct personality types. This means that when personality psychologists talk about Introverts and Extraverts, they are not really talking about two distinct types of people who are completely and qualitatively different from one another.
Instead, they are talking about people who score relatively low or relatively high along a continuous distribution. In fact, when personality psychologists measure traits like Extraversion, they typically find that most people score somewhere in the middle, with smaller numbers showing more extreme levels. Figure \(2\) shows the distribution of Extraversion scores from a survey of thousands of people. As you can see, most people report being moderately, but not extremely, extraverted, with fewer people reporting very high or very low scores.
There are three criteria that characterize personality traits:
(1) consistency, (2) stability, and (3) individual differences.
1. To have a personality trait, individuals must be somewhat consistent across situations in their behaviors related to the trait. For example, if they are talkative at home, they tend also to be talkative at work.
2. Individuals with a trait are also somewhat stable over time in behaviors related to the trait. If they are talkative, for example, at age 30, they will also tend to be talkative at age 40.
3. People differ from one another on behaviors related to the trait. Using speech is not a personality trait and neither is walking on two feet—virtually all individuals do these activities, and there are almost no individual differences. But people differ on how frequently they talk and how active they are, and thus personality traits such as Talkativeness and Activity Level do exist.
A challenge of the trait approach was to discover the major traits on which all people differ. Scientists for many decades generated hundreds of new traits, so that it was soon difficult to keep track and make sense of them. For instance, one psychologist might focus on individual differences in “friendliness,” whereas another might focus on the highly related concept of “sociability.” Scientists began seeking ways to reduce the number of traits in some systematic way and to discover the basic traits that describe most of the differences between people.
The way that Gordon Allport and his colleague Henry Odbert approached this was to search the dictionary for all descriptors of personality (Allport & Odbert, 1936). Their approach was guided by the lexical hypothesis, which states that all important personality characteristics should be reflected in the language that we use to describe other people. Therefore, if we want to understand the fundamental ways in which people differ from one another, we can turn to the words that people use to describe one another. So if we want to know what words people use to describe one another, where should we look? Allport and Odbert looked in the most obvious place—the dictionary. Specifically, they took all the personality descriptors that they could find in the dictionary (they started with almost 18,000 words but quickly reduced that list to a more manageable number) and then used statistical techniques to determine which words “went together.” In other words, if everyone who said that they were “friendly” also said that they were “sociable,” then this might mean that personality psychologists would only need a single trait to capture individual differences in these characteristics. Statistical techniques were used to determine whether a small number of dimensions might underlie all of the thousands of words we use to describe people. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/31%3A_Personality_Traits/31.01%3A_Introduction.txt |
Research that used the lexical approach showed that many of the personality descriptors found in the dictionary do indeed overlap. In other words, many of the words that we use to describe people are synonyms. Thus, if we want to know what a person is like, we do not necessarily need to ask how sociable they are, how friendly they are, and how gregarious they are. Instead, because sociable people tend to be friendly and gregarious, we can summarize this personality dimension with a single term. Someone who is sociable, friendly, and gregarious would typically be described as an “Extravert.” Once we know she is an extravert, we can assume that she is sociable, friendly, and gregarious.
Statistical methods (specifically, a technique called factor analysis) helped to determine whether a small number of dimensions underlie the diversity of words that people like Allport and Odbert identified. The most widely accepted system to emerge from this approach was “The Big Five” or “Five-Factor Model” (Goldberg, 1990; McCrae & Costa, 1987; McCrae & John, 1992). The Big Five comprises five major traits shown in Table \(1\). A way to remember these five is with the acronym OCEAN (O is for Openness; C is for Conscientiousness; E is for Extraversion; A is for Agreeableness; N is for Neuroticism). Table \(2\) provides descriptions of people who would score high and low on each of these traits.
Table \(1\): Descriptions of the Big Five Personality Traits
Big Five Trait
Definition
Openness
The tendency to appreciate new art, ideas, values, feelings, and behaviors
Conscientiousness
The tendency to be careful, to be on-time for appointments, to follow rules, and to be hardworking
Extraversion
The tendency to be talkative, to be sociable, and to enjoy others; the tendency to have a dominant style
Agreeableness
The tendency to agree and go along with others rather than to assert one’s own opinions and choices
Neuroticism
The tendency to frequently experience negative emotions such as anger, worry, and sadness; the tendency to be interpersonally sensitive
Table \(2\):Example Behaviors for Those Scoring low and High for the Big Five Traits
Big Five Trait
Example Behavior for LOW Scorers
Example Behavior for HIGH Scorers
Openness
Prefers not to be exposed to alternative moral systems; narrow interests; inartistic; not analytical; down-to-earth
Enjoys seeing people with new types of haircuts and body piercing; curious; imaginative; untraditional
Conscientiousness
Prefers spur-of-the-moment action to planning; unreliable; hedonistic; careless; lax
Never late for a date; organized; hardworking; neat; persevering; punctual; self-disciplined
Extraversion
Prefers a quiet evening reading to a loud party; sober; aloof; unenthusiastic
Is the life of the party; active; optimistic; fun-loving; affectionate
Agreeableness
Quickly and confidently asserts own rights; irritable; manipulative; uncooperative; rude
Agrees with others about political opinions; good- natured; forgiving; gullible; helpful
Neuroticism
Not irritated by small annoyances; calm; unemotional; hardy; secure; self-satisfied
Constantly worries about little things; insecure; hypochondriacal; feels inadequate
Scores on the Big Five traits are mostly independent. That means that a person’s standing on one trait tells very little about their standing on the other traits of the Big Five. For example, a person can be extremely high in Extraversion and be either high or low on Neuroticism. Similarly, a person can be low in Agreeableness and be either high or low in Conscientiousness. Thus, in the Five-Factor Model, you need five scores to describe most of an individual’s personality.
In the Appendix to this reading, we present a short scale to assess the Five-Factor Model of personality (Donnellan et al., 2006). You can take this test to see where you stand in terms of your Big Five scores. John Johnson has also created a helpful website that has personality scales that can be used and taken by the general public:
http://www.personal.psu.edu/j5j/IPIP/ipipneo120.htm
After seeing your scores, you can judge for yourself whether you think such tests are valid.
Traits are important and interesting because they describe stable patterns of behavior that persist for long periods of time (Caspi et al., 2005). Importantly, these stable patterns can have broad-ranging consequences for many areas of our life (Roberts et al., 2007). For instance, think about the factors that determine success in college. If you were asked to guess what factors predict good grades in college, you might guess something like intelligence. This guess would be correct, but we know much more about who is likely to do well. Specifically, personality researchers have also found that personality traits like Conscientiousness play an important role in college and beyond, probably because highly conscientious individuals study hard, get their work done on time, and are less distracted by nonessential activities that take time away from school work. In addition, highly conscientious people are often healthier than people low in conscientiousness because they are more likely to maintain healthy diets, to exercise, and to follow basic safety procedures like wearing seat belts or bicycle helmets. Over the long term, this consistent pattern of behaviors can add up to meaningful differences in health and longevity. Thus, personality traits are not just a useful way to describe people you know; they actually help psychologists predict how good a worker someone will be, how long he or she will live, and the types of jobs and activities the person will enjoy. Thus, there is growing interest in personality psychology among psychologists who work in applied settings, such as health psychology or organizational psychology. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/31%3A_Personality_Traits/31.02%3A_The_Five-Factor_Model_Of_Personality.txt |
So how does it feel to be told that your entire personality can be summarized with scores on just five personality traits? Do you think these five scores capture the complexity of your own and others’ characteristic patterns of thoughts, feelings, and behaviors? Most people would probably say no, point- ing to some exception in their behavior that goes against the general pattern that others might see. For instance, you may know people who are warm and friendly and find it easy to talk with strangers at a party yet are terrified if they have to perform in front of others or speak to large groups of people. The fact that there are different ways of being extraverted or conscientious shows that there is value in considering lower-level units of personality that are more specific than the Big Five traits. These more specific, lower-level units of personality are often called facets.
To give you a sense of what these narrow units are like, Table \(1\) shows facets for each of the Big Five traits. It is important to note that although personality researchers generally agree about the value of the Big Five traits as a way to summarize one’s personality, there is no widely accepted list of facets that should be studied. The list seen here, based on work by researchers Paul Costa and Jeff McCrae, thus reflects just one possible list among many. It should, however, give you an idea of some of the facets making up each of the Five-Factor Model.
Table \(1\): Facets of Big Five Personality Traits
Trait
Facets of Trait
Openness
• Fantasy prone
• Open to feelings
• Open to diverse behaviors
• Open to new and different ideas
• Open to various values and beliefs
Conscientiousness
• Competent
• Orderly
• Dutiful
• Achievement oriented
• Self-disciplined
• Deliberate
Extraversion
• Gregarious (sociable)
• Warm
• Assertive
• Active
• Excitement-seeking
• Positive emotionality
Agreeableness
• Trusting
• Straightforward
• Altruistic
• Compliant
• Modest
• Tender-minded
Neuroticism
• Anxious
• Angry
• Depressed
• Self-conscious
• Impulsive
• Vulnerable
Facets can be useful because they provide more specific descriptions of what a person is like. For instance, if we take our friend who loves parties but hates public speaking, we might say that this person scores high on the “gregariousness” and “warmth” facets of extraversion, while scoring lower on facets such as “assertiveness” or “excitement-seeking.”
This precise profile of facet scores not only provides a better description, it might also allow us to better predict how this friend will do in a variety of different jobs (for example, jobs that require public speaking versus jobs that involve one- on-one interactions with customers) (Paunonen & Ashton, 2001). Because different facets within a broad, global trait like extraversion tend to go together (those who are gregarious are often but not always assertive), the broad trait often provides a useful summary of what a person is like. But when we really want to know a person, facet scores add to our knowledge in important ways.
31.04: Other Traits Beyond The Five-Factor Model
Despite the popularity of the Five-Factor Model, it is certainly not the only model that exists. Some suggest that there are more than five major traits, or perhaps even fewer. For example, in one of the first comprehensive models to be proposed, Hans Eysenck suggested that Extraversion and Neuroticism are most important. Eysenck believed that by combining people’s standing on these two major traits, we could account for many of the differences in personality that we see in people (Eysenck, 1981). So for instance, a neurotic introvert would be shy and nervous, while a stable introvert might avoid social situations and prefer solitary activities, but he may do so with a calm, steady attitude and little anxiety or emotion. Interestingly, Eysenck attempted to link these two major dimensions to underlying differences in people’s biology. For instance, he suggested that introverts experienced too much sensory stimulation and arousal, which made them want to seek out quiet settings and less stimulating environments. More recently, Jeffrey Gray (1981) suggested that these two broad traits are related to fundamental reward and avoidance systems in the brain—extraverts might be motivated to seek reward and thus exhibit assertive, reward-seeking behavior, whereas people high in neuroticism might be motivated to avoid punishment and thus may experience anxiety as a result of their heightened awareness of the threats in the world around them (Gray’s model has since been updated; see Gray & McNaughton, 2000). These early theories have led to a burgeoning interest in identifying the physiological underpinnings of the individual differences that we observe.
Another revision of the Big Five is the HEXACO model of traits (Ashton & Lee, 2007). This model is similar to the Big Five, but it posits slightly different versions of some of the traits, and its proponents argue that one important class of individual differences was omitted from the Five-Factor Model. The HEXACO adds Honesty-Humility as a sixth dimension of personality. People high in this trait are sincere, fair, and modest, whereas those low in the trait are manipulative, narcissistic, and self-centered. Thus, trait theorists are agreed that personality traits are important in understanding behavior, but there are still debates on the exact number and composition of the traits that are most important.
There are other important traits that are not included in comprehensive models like the Big Five. Although the five factors capture much that is important about personality, researchers have suggested other traits that capture interesting aspects of our behavior. In Table \(1\) we present just a few, out of hundreds, of the other traits that have been studied by personologists.
Table \(1\): Other Traits Beyond Those Included in the Big Five
Personality Trait
Description
Machiavellianism
Named after the famous political philosopher, Niccolo Machiavelli, this trait refers to individuals who manipulate the behavior of others, often through duplicity. Machiavellians are often interested in money and power, and pragmatically use others in this quest.
Need for Achievement
Those high in need for achievement want to accomplish a lot and set high standards of excellence for themselves. They are able to work persistently and hard for distant goals. David McClelland argued that economic growth depends in part on citizens with high need for achievement.
Need for Cognition
People high in need for cognition find it rewarding to understand things and are willing to use considerable cognitive effort in this quest. Such individuals enjoy learning and the process of trying to understand new things.
Authoritarianism
Authoritarians believe in strict social hierarchies in which they are totally obedient to those above them and expect complete obedience from their subordinates. Rigid in adherence to rules, the authoritarian personality is very uncomfortable with uncertainty.
Narcissism
The narcissistic personality has self-love that is so strong that it results in high levels of vanity, conceit, and selfishness. The narcissistic individual often has problems feeling empathetic toward others and grateful to others.
Self-esteem
The tendency to evaluate oneself positively. Self-esteem does not imply that one believes that he or she is better than others, only that he or she is a person of worth.
Optimism
The tendency to expect positive outcomes in the future. People who are optimistic expect good things to happen, and indeed they often have more positive outcomes, perhaps because they work harder to achieve them.
Alexithymia
The inability to recognize and label emotions in oneself. The individual also has a difficult time recognizing emotions in others and often has difficulties in relationships.
Not all of the traits in Table \(1\) are currently popular with scientists, yet each of them has experienced popularity in the past. Although the Five-Factor Model has been the tar- get of more rigorous research than some of the traits in the table, these additional personality characteristics give a good idea of the wide range of behaviors and attitudes that traits can cover. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/31%3A_Personality_Traits/31.03%3A_Facets_Of_Traits_%28Subtraits%29.txt |
The ideas described in this module should probably seem familiar, if not obvious to you. When asked to think about what our friends, enemies, family members, and colleagues are like, some of the first things that come to mind are their personality characteristics. We might think about how warm and helpful our first teacher was, how irresponsible and care- less our brother is, or how demanding and insulting our first boss was. Each of these descriptors reflects a personality trait, and most of us generally think that the descriptions we use for individuals accurately reflect their “characteristic pattern of thoughts, feelings, and behaviors,” or in other words, their personality.
But what if this idea were wrong? What if our belief in personality traits were an illusion and people are not consistent from one situation to the next? This was a possibility that shook the foundation of personality psychology in the late 1960s when Walter Mischel published a book called Personality and Assessment (1968). In this book, Mischel suggested that if one looks closely at people’s behavior across many different situations, the consistency is really not that impressive. In other words, children who cheat on tests at school may steadfastly follow all rules when playing games and may never tell a lie to their parents. In other words, he suggested, there may not be any general trait of honesty that links these seemingly related behaviors. Furthermore, Mischel suggested that observers may believe that broad personality traits like honesty exist when, in fact, this belief is an illusion. The debate that followed the publication of Mischel’s book was called the person-situation debate because it pitted the power of personality against the power of situational factors as determinants of the behavior that people exhibit.
Because of the findings that Mischel emphasized, many psychologists focused on an alternative to the trait perspective. Instead of studying broad, context-free descriptions, like the trait terms we’ve described so far, Mischel thought that psychologists should focus on people’s distinctive reactions to specific situations. For instance, although there may not be a broad and general trait of honesty, some children may be especially likely to cheat on a test when the risk of being caught is low and the rewards for cheating are high. Others might be motivated by the sense of risk involved in cheating and may do so even when the rewards are not very high. Thus, the behavior itself results from the child’s unique evaluation of the risks and rewards present at that moment, along with her evaluation of her abilities and values. Because of this, the same child might act very differently in different situations. Thus, Mischel thought that specific behaviors were driven by the interaction between very specific, psychologically meaningful features of the situation in which people found them- selves, the person’s unique way of perceiving that situation, and his or her abilities for dealing with it. Mischel and others argued that it was these social-cognitive processes that under- lie people’s reactions to specific situations that provide some consistency when situational features are the same. If so, then studying these broad traits might be more fruitful than cataloging and measuring narrow, context-free traits like Extra- version or Neuroticism.
In the years after the publication of Mischel’s (1968) book, debates raged about whether personality truly exists, and if so, how it should be studied. And, as is often the case, it turns out that a more moderate middle ground than what the situationists proposed could be reached. It is certainly true, as Mischel pointed out, that a person’s behavior in one specific situation is not a good guide to how that person will behave in a very different specific situation. Someone who is extremely talkative at one specific party may sometimes be reticent to speak up during class and may even act like a wallflower at a different party. But this does not mean that personality does not exist, nor does it mean that people’s behavior is completely determined by situational factors. Indeed, research conducted after the person-situation debate shows that on average, the effect of the “situation” is about as large as that of personality traits. However, it is also true that if psychologists assess a broad range of behaviors across many different situations, there are general tendencies that emerge. Personality traits give an indication about how people will act on average, but frequently they are not so good at predicting how a person will act in a specific situation at a certain moment in time. Thus, to best capture broad traits, one must assess aggregate behaviors, averaged over time and across many different types of situations. Most modern personality researchers agree that there is a place for broad personality traits and for the narrower units such as those studied by Walter Mischel. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/31%3A_Personality_Traits/31.05%3A_The_Person-Situation_Debate_And_Alternatives_To_The_Trait_Perspective.txt |
The Mini-iPiP Scale
(Donnellan et al., 2006)
Instructions: Below are phrases describing people’s behaviors. Please use the rating scale below to describe how accurately each statement describes you. Describe yourself as you generally are now, not as you wish to be in the future. Describe yourself as you honestly see yourself, in relation to other people you know of the same sex as you are, and roughly your same age. Please read each statement carefully, and put a number from 1 to 5 next to it to describe how accurately the statement describes you.
1 = Very inaccurate
2 = Moderately inaccurate
3 = Neither inaccurate nor accurate 4 = Moderately accurate
5 = Very accurate
1. Am the life of the party (E)
2. Sympathize with others’ feelings (A)
3. Get chores done right away (C)
4. Have frequent mood swings (N)
5. Have a vivid imagination (O)
6. Don’t talk a lot (E)
7. Am not interested in other people’s problems (A)
8. Often forget to put things back in their proper place (C)
9. Am relaxed most of the time (N)
10. Am not interested in abstract ideas (O)
11. Talk to a lot of different people at parties (E)
12. Feel others’ emotions (A)
13. Like order (C)
14. Get upset easily (N)
15. Have difficulty understanding abstract ideas (O)
16. Keep in the background (E)
17. Am not really interested in others (A)
18. Make a mess of things (C)
19. Seldom feel blue (N)
20. Do not have a good imagination (O)
Scoring: The first thing you must do is to reverse the items that are worded in the opposite direction. In order to do this, subtract the number you put for that item from 6. So if you put a 4, for instance, it will become a 2. Cross out the score you put when you took the scale, and put the new number in representing your score subtracted from the number 6.
Items to be reversed in this way:
6, 7, 8, 9, 10, 15, 16, 17, 18, 19, 20
Next, you need to add up the scores for each of the five OCEAN scales (including the reversed numbers where relevant). Each OCEAN score will be the sum of four items. Place the sum next to each scale below.
Openness: Add items 5, 10, 15, 20 Conscientiousness: Add items 3, 8, 13, 18 Extraversion: Add items 1, 6, 11, 16 Agreeableness: Add items 2, 7, 12, 17 Neuroticism: Add items 4, 9,14, 19
Compare your scores to the norms below to see where you stand on each scale. If you are low on a trait, it means you are the opposite of the trait label. For example, low on Extraversion is Introversion, low on Openness is Conventional, and low on Agreeableness is Assertive.
19–20 Extremely high
17–18 Very high
14–16 High
11–13 Neither high nor low; in the middle
8–10 Low
6–7 Very low
4–5 Extremely low
REFERENCES
Allport, G. W., & Odbert, H. S. (1936). Trait names: A psycho-lexical study. Psychological Monographs, 47(1), 211. doi.org/10.1037/ h0093360
Ashton, M. C., & Lee, K. (2007). Empirical, theoretical, and practical advantages of the HEXACO model of personality structure. Personality and Social Psychological Review, 11(2), 150–166. https:// doi.org/10.1177/1088868306294907
Caspi, A., Roberts, B. W., & Shiner, R. L. (2005). Personality development: Stability and change. Annual Reviews of Psychology, 56(1), 453–484. https://doi.org/10.1146/annurev.psyc....090902.141913
Donnellan, M. B., Oswald, F. L., Baird, B. M., & Lucas, R. E. (2006). The mini-IPIP scales: Tiny-yet-effective measures of the Big Five factors of personality. Psychological Assessment, 18(2), 192–203. doi. org/10.1037/1040-3590.18.2.192
Eysenck, H. J. (1981). A model for personality. Springer Verlag. Goldberg, L. R. (1990). An alternative “description of personality”: The
Big-Five factor structure. Journal of Personality and Social Psychology, 59(6), 1216–1229. doi.org/10.1037/ 0022-3514.59.6.1216
Gray, J. A. (1981). A critique of Eysenck’s theory of personality. In H. J. Eysenck (Ed.), A model for personality (pp. 246–276). Springer Verlag.
Gray, J. A., & McNaughton, N. (2000). The neuropsychology of anxiety: An enquiry into the functions of the septo-hippocampal system (2nd ed.). Oxford University Press.
Matthews, G., Deary, I. J., & Whiteman, M. C. (2003). Personality traits. Cambridge University Press.
McCrae, R. R., & Costa, P. T. (1987). Validation of the five-factor model of personality across instruments and observers. Journal of Personality and Social Psychology, 52(1), 81–90. doi.org/10.1037/ 0022-3514.52.1.81
McCrae, R. R., & John, O. P. (1992). An introduction to the five-factor model and its applications. Journal of Personality, 60(2), 175–215. doi.org/10.1111/j.1467-6494.1992.tb00970.x
Mischel, W. (1968). Personality and assessment. John Wiley.
Paunonen, S. V., & Ashton, M. S. (2001). Big five factors and facets and the prediction of behavior. Journal of Personality and Social Psychology, 81(3), 524–539. doi.org/10.1037/ 0022-3514.81.3.524
Roberts, B. W., Kuncel, N. R., Shiner, R., Caspi, A., & Golberg, L. R. (2007). The power of personality: The comparative validity of personality traits, socioeconomic status, and cognitive ability for predicting important life outcomes. Perspectives on Psychological Science, 2(4), 313–345. https://doi.org/10.1111/j.1745-6916.2007.00047.x | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/31%3A_Personality_Traits/31.06%3A_Appendix.txt |
Apsychological disorder is a condition characterized by abnormal thoughts, feelings, and behaviors. Psychopathology is the study of psychological disorders, including their symptoms, etiology (i.e., their causes), and treatment. The term psychopathology can also refer to the manifestation of a psychological disorder. Although consensus can be difficult, it is extremely important for mental health professionals to agree on what kinds of thoughts, feelings, and behaviors are truly abnormal in the sense that they genuinely indicate the presence of psychopathology. Certain patterns of behavior and inner experience can easily be labeled as abnormal and clearly signify some kind of psychological disturbance. The person who washes his hands 40 times per day and the person who claims to hear the voices of demons exhibit behaviors and inner experiences that most would regard as abnormal: beliefs and behaviors that suggest the existence of a psychological disorder. But, consider the nervousness a young man feels when talking to attractive women or the loneliness and longing for home a freshman experiences during her first semester of college—these feelings may not be regularly present, but they fall in the range of normal. So, what kinds of thoughts, feelings, and behaviors represent a true psychological disorder? Psychologists work to distinguish psychological disorders from inner experiences and behaviors that are merely situational, idiosyncratic, or unconventional.
32.02: Definition Of A Psychological Disorder
Perhaps the simplest approach to conceptualizing psycho- logical disorders is to label behaviors, thoughts, and inner experiences that are atypical, distressful, dysfunctional, and sometimes even dangerous, as signs of a disorder. For example, if you ask a classmate for a date and you are rejected, you probably would feel a little dejected. Such feelings would be normal. If you felt extremely depressed—so much so that you lost interest in activities, had difficulty eating or sleeping, felt utterly worthless, and contemplated suicide—your feelings would be atypical, would deviate from the norm, and could signify the presence of a psychological disorder. Just because something is atypical, however, does not necessarily mean it is disordered.
For example, only about 4% of people in the United States have red hair, so red hair is considered an atypical characteristic (Figure \(1\)), but it is not considered disordered, it’s just unusual. And it is less unusual in Scotland, where approximately 13% of the population has red hair (“DNA Project Aims,” 2012). As you will learn, some disorders, although not exactly typical, are far from atypical, and the rates in which they appear in the population are surprisingly high.
If we can agree that merely being atypical is an insufficient criterion for a having a psychological disorder, is it reasonable to consider behavior or inner experiences that differ from widely expected cultural values or expectations as disordered? Using this criterion, a woman who walks around a subway platform wearing a heavy winter coat in July while screaming obscenities at strangers may be considered as exhibiting symptoms of a psychological disorder. Her actions and clothes violate socially accepted rules governing appropriate dress and behavior; these characteristics are atypical.
32.03: Cultural Expectations
Violating cultural expectations is not, in and of itself, a satisfactory means of identifying the presence of a psychological disorder. Since behavior varies from one culture to another, what may be expected and considered appropriate in one culture may not be viewed as such in other cultures. For example, returning a stranger’s smile is expected in the United States because a pervasive social norm dictates that we reciprocate friendly gestures. A person who refuses to acknowledge such gestures might be considered socially awkward—perhaps even disordered—for violating this expectation. How- ever, such expectations are not universally shared. Cultural expectations in Japan involve showing reserve, restraint, and a concern for maintaining privacy around strangers. Japanese people are generally unresponsive to smiles from strangers (Patterson et al., 2007). Eye contact provides another example. In the United States and Europe, eye contact with others typically signifies honesty and attention. However, most Latin American, Asian, and African cultures interpret direct eye contact as rude, confrontational, and aggressive (Pazain, 2010). Thus, someone who makes eye contact with you could be considered appropriate and respectful or brazen and offensive, depending on your culture (Figure \(1\)).
Hallucinations (seeing or hearing things that are not physically present) in Western societies is a violation of cultural expectations, and a person who reports such inner experiences is readily labeled as psychologically disordered. In other cultures, visions that, for example, pertain to future events may be regarded as normal experiences that are positively valued (Bourguignon, 1970). Finally, it is important to recognize that cultural norms change over time: what might be considered typical in a society at one time may no longer be viewed this way later, similar to how fashion trends from one era may elicit quizzical looks decades later—imagine how a headband, legwarmers, and the big hair of the 1980s would go over on your campus today. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/32%3A_What_Are_Psychological_Disorders/32.01%3A_Introduction.txt |
If none of the criteria discussed so far is adequate by itself to define the presence of a psychological disorder, how can a disorder be conceptualized? Many efforts have been made to identify the specific dimensions of psychological disorders, yet none is entirely satisfactory. No universal definition of psychological disorder exists that can apply to all situations in which a disorder is thought to be present (Zachar & Kendler, 2007). However, one of the more influential conceptualizations was proposed by Wakefield (1992), who defined psycho- logical disorder as a harmful dysfunction. Wakefield argued that natural internal mechanisms—that is, psychological processes honed by evolution, such as cognition, perception, and learning—have important functions, such as enabling us to experience the world the way others do and to engage in rational thought, problem solving, and communication. For example, learning allows us to associate a fear with a potential danger in such a way that the intensity of fear is roughly equal to the degree of actual danger. Dysfunction occurs when an internal mechanism breaks down and can no longer perform its normal function. But the presence of a dysfunction by itself does not determine a disorder. The dysfunction must be harmful in that it leads to negative consequences for the individual or for others, as judged by the standards of the individual’s culture. The harm may include significant internal anguish (e.g., high levels of anxiety or depression) or problems in day-to-day living (e.g., in one’s social or work life).
To illustrate, Janet has an extreme fear of spiders. Janet’s fear might be considered a dysfunction in that it signals that the internal mechanism of learning is not working correctly (i.e., a faulty process prevents Janet from appropriately associating the magnitude of her fear with the actual threat posed by spiders). Janet’s fear of spiders has a significant negative influence on her life: she avoids all situations in which she suspects spiders to be present (e.g., the basement or a friend’s home), and she quit her job last month and is now unemployed because she saw a spider in the restroom at work.
DIG DEEPER
The Myth of Mental illness
In the 1950s and 1960s, the concept of mental illness was widely criticized. One of the major criticisms focused on the notion that mental illness was a “myth that justifies psychiatric intervention in socially disapproved behavior” (Wakefield, 1992). Thomas Szasz (1960), a noted psychiatrist, was perhaps the biggest proponent of this view. Szasz argued that the notion of mental illness was invented by society (and the mental health establishment) to stigmatize and subjugate people whose behavior violates accepted social and legal norms. Indeed, Szasz suggested that what appear to be symptoms of mental illness are more appropriately characterized as “problems in living” (Szasz, 1960).
In his 1961 book, The Myth of Mental Illness: Foundations of a Theory of Personal Conduct, Szasz expressed his disdain for the concept of mental illness and for the field of psychiatry in general (Oliver, 2006). The basis for Szasz’s attack was his contention that detectable abnormalities in bodily structures and functions (e.g., infections and organ damage or dysfunction) represent the defining features of genuine illness or disease, and because symptoms of purported mental illness are not accompanied by such detect- able abnormalities, so-called psychological disorders are not disorders at all. Szasz (1961/2010) proclaimed that “dis- ease or illness can only affect the body; hence, there can be no mental illness” (p. 267).
Today, we recognize the extreme level of psychological suffering experienced by people with psychological disorders: the painful thoughts and feelings they experience, the disordered behavior they demonstrate, and the levels of distress and impairment they exhibit. This makes it very difficult to deny the reality of mental illness.
However controversial Szasz’s views and those of his supporters might have been, they have influenced the mental health community and society in several ways. First, lay people, politicians, and professionals now often refer to mental illness as mental health “problems,” implicitly acknowledging the “problems in living” perspective Szasz described (Buchanan-Barker & Barker, 2009). Also influential was Szasz’s view of homosexuality. Szasz was perhaps the first psychiatrist to openly challenge the idea that homo- sexuality represented a form of mental illness or disease (Szasz, 1965). By challenging the idea that homosexuality represented a form a mental illness, Szasz helped pave the way for the social and civil rights that gay and lesbian people now have (Barker & Buchanan-Barker, 2010). His work also inspired legal changes that protect the rights of people in psychiatric institutions and allow such individuals a greater degree of influence and responsibility over their lives (Buchanan-Barker & Barker, 2009). ■
According to the harmful dysfunction model, Janet’s condition would signify a disorder because (a) there is a dysfunction in an internal mechanism, and (b) the dysfunction has resulted in harmful consequences. Similar to how the symptoms of physical illness reflect dysfunctions in biological processes, the symptoms of psychological disorders presumably reflect dysfunctions in mental processes. The internal mechanism component of this model is especially appealing because it implies that disorders may occur through a break- down of biological functions that govern various psychological processes, thus supporting contemporary neurobiological models of psychological disorders (Fabrega, 2007). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/32%3A_What_Are_Psychological_Disorders/32.04%3A_Harmful_Dysfunction.txt |
Many of the features of the harmful dysfunction model are incorporated in a formal definition of psychological disorder developed by the American Psychiatric Association (APA). According to the APA (2013), a psychological disorder is a condition that is said to consist of the following:
• There are significant disturbances in thoughts, feelings, and behaviors. A person must experience inner states (e.g., thoughts and/or feelings) and exhibit behaviors that are clearly disturbed—that is, unusual, but in a negative, self-defeating way. Often, such disturbances are troubling to those around the individual who experiences them. For example, an individual who is uncontrollably preoccupied by thoughts of germs spends hours each day bathing, has inner experiences, and displays behaviors that most would consider atypical and negative (disturbed) and that would likely be troubling to family members.
• The disturbances reflect some kind of biological, psychological, or developmental dysfunction. Disturbed patterns of inner experiences and behaviors should reflect some flaw (dysfunction) in the internal biological, psychological, and developmental mechanisms that lead to normal, healthy psychological functioning. For example, the hallucinations observed in schizophrenia could be a sign of brain abnormalities.
• The disturbances lead to significant distress or disability in one’s life. A person’s inner experiences and behaviors are considered to reflect a psychological disorder if they cause the person considerable distress or greatly impair his ability to function as a normal individual (often referred to as functional impairment, or occupational and social impairment). As an illustration, a person’s fear of social situations might be so distressing that it causes the person to avoid all social situations (e.g., preventing that person from being able to attend class or apply for a job).
• The disturbances do not reflect expected or culturally approved responses to certain events. Disturbances in thoughts, feelings, and behaviors must be socially unacceptable responses to certain events that often happen in life. For example, it is perfectly natural (and expected) that a person would experience great sadness and might wish to be left alone following the death of a close family member. Because such reactions are in some ways culturally expected, the individual would not be assumed to signify a mental disorder.
Some believe that there is no essential criterion or set of criteria that can definitively distinguish all cases of disorder from nondisorder (Lilienfeld & Marino, 1999). In truth, no single approach to defining a psychological disorder is adequate by itself, nor is there universal agreement on where the boundary is between disordered and not disordered. From time to time we all experience anxiety, unwanted thoughts, and moments of sadness; our behavior at other times may not make much sense to ourselves or to others. These inner experiences and behaviors can vary in their intensity but are only considered disordered when they are highly disturbing to us and/or others, suggest a dysfunction in normal mental functioning, and are associated with significant distress or disability in social or occupational activities.
REFERENCES
American Psychiatric Association. (2013). Diagnostic and statistical manual of mental disorders (5th ed.). https://doi.org/10.1176/appi. books.9780890425596
Barker, P., & Buchanan-Barker, P. (2010). The legacy of Thomas Szasz. Center for Independent Thought. http://www. centerforindependentthought.org/SzaszLegacy.html
Bourguignon, E. (1970). Hallucinations and trance: An anthropologist’s perspective. In W. Keup (Ed.), Origins and mechanisms of hallucination (pp. 183–190). Plenum Press.
Buchanan-Barker, P., & Barker, P. (2009, February). The convenient myth of Thomas Szasz. Journal of Psychiatric and Mental Health Nursing, 16(1), 87–95. https://doi.org/10.1111/j.1365-2850.2008.01310.x
DNA project aims to count Scots redheads. (2012, November 7). BBC News. http://www.bbc.com/news/uk-scotland-20237511
Fabrega, H. (2007). How psychiatric conditions were made. Psychiatry, 70(2), 130–153. https://doi.org/10.1521/psyc.2007.70.2.130
Lilienfeld, S. O., & Marino, L. (1999). Essentialism revisited: Evolutionary theory and the concept of mental disorder. Journal of Abnormal Psychology, 108(3), 400–411. https://doi.org/10.1037/ 0021-843X.108.3.400
Oliver, J. (2006, July). The myth of Thomas Szasz. The New Atlantis, 13(13), 68–84. http://www.thenewatlantis.com/docLib/TNA13-Oliver.pdf
Patterson, M. L., Iizuka, Y., Tubbs, M. E., Ansel, J., Tsutsumi, M., & Anson, J. (2007). Passing encounters east and west: Comparing Japanese and American pedestrian interactions. Journal of Nonverbal Behavior, 31(3), 155–166. https://doi.org/10.1007/s10919-007-0028-4
Pazain, M. (2010, December 2). To look or not to look? Eye contact differences in different cultures. Examiner.com. https://web.archive. org/web/20160618201038/http://www.examiner.com/article/ to-look-or-not-to-look-eye-contact-differences-different-cultures
Szasz, T. S. (1960). The myth of mental illness. American Psychologist, 15(2), 113–118. https://doi.org/10.1037/h0046535
Szasz, T. S. (1965). Legal and moral aspects of homosexuality. In J. Marmor (Ed.), Sexual inversion: The multiple roots of homosexuality (pp. 124–139). Basic Books.
Szasz, T. S. (2010). The myth of mental illness: Foundations of a theory of personal conduct. HarperCollins. (Original work published 1961)
Wakefield, J. C. (1992). The concept of mental disorder: On the boundary between biological facts and social values. American Psychologist, 47(3), 373–388. https://doi.org/10.1037/0003-066X.47.3.373
Zachar, P., & Kendler, K. S. (2007). Psychiatric disorders: A conceptual taxonomy. American Journal of Psychiatry, 164(4), 557–565. https:// doi.org/10.1176/ajp.2007.164.4.557 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/32%3A_What_Are_Psychological_Disorders/32.05%3A_The_American_Psychiatric_Association_%28APA%29_Definition.txt |
A first step in the study of psychological disorders is care- fully and systematically discerning significant signs and symptoms. How do mental health professionals ascertain whether or not a person’s inner states and behaviors truly represent a psychological disorder? Arriving at a proper diagnosis—that is, appropriately identifying and labeling a set of defined symptoms—is absolutely crucial. This process enables professionals to use a common language with others in the field and aids in communication about the disorder with the patient, colleagues, and the public. A proper diagnosis is an essential element to guide proper and successful treatment. For these reasons, classification systems that organize psycho- logical disorders systematically are necessary.
33.02: The Diagnostic And Statistical Manual Of Menal Disorders (DSM)
Although a number of classification systems have been developed over time, the one that is used by most mental health professionals in the United States is the Diagnostic and Statistical Manual of Mental Disorders (DSM-5), published by the American Psychiatric Association (2013). (Note that the American Psychiatric Association differs from the American Psychological Association, but both are abbreviated APA.) The first edition of the DSM, published in 1952, classified psychological disorders according to a format developed by the U.S. Army during World War II (Clegg, 2012). In the years since, the DSM has undergone numerous revisions and editions. The most recent edition, published in 2013, is the DSM-5 (APA, 2013). The DSM-5 includes many categories of disorders (e.g., anxiety disorders, depressive disorders, and dissociative disorders). Each disorder is described in detail, including an overview of the disorder (diagnostic features), specific symptoms required for diagnosis (diagnostic criteria), prevalence information (what percent of the population is thought to be afflicted with the disorder), and risk factors associated with the disorder. Figure \(1\) shows lifetime prevalence rates—the percentage of people in a population who develop a disorder in their lifetime—of various psychological disorders among U.S. adults. These data were based on a national sample of 9,282 U.S. residents (National Comorbidity Survey, 2007).
The DSM-5 also provides information about comorbidity; the co-occurrence of two disorders. For example, the DSM-5 mentions that 41% of people with obsessive-compulsive disorder (OCD) also meet the diagnostic criteria for major depressive disorder (Figure \(2\)). Drug use is highly comorbid with other mental illnesses; 6 out of 10 people who have a substance use disorder also suffer from another form of mental illness (National Institute on Drug Abuse, 2007).
The DSM has changed considerably in the half-century since it was originally published. The first two editions of the DSM, for example, listed homosexuality as a disorder; how- ever, in 1973, the APA voted to remove it from the manual (Silverstein, 2009). Additionally, beginning with the DSM- III in 1980, mental disorders have been described in much greater detail, and the number of diagnosable conditions has grown steadily, as has the size of the manual itself. DSM-I included 106 diagnoses and was 130 total pages, whereas DSM-III included more than 2 times as many diagnoses (265) and was nearly seven times its size (886 total pages) (Mayes & Horowitz, 2005). Although DSM-5 is longer than DSM-IV, the volume includes only 237 disorders, a decrease from the 297 disorders that were listed in DSM-IV. The latest edition, DSM-5, includes revisions in the organization and naming of categories and in the diagnostic criteria for various disorders (Regier et al., 2012), while emphasizing careful consideration of the importance of gender and cultural difference in the expression of various symptoms (Fisher, 2010).
Some believe that establishing new diagnoses might overpathologize the human condition by turning common human problems into mental illnesses (The Associated Press, 2013). Indeed, the finding that nearly half of all Americans will meet the criteria for a DSM disorder at some point in their life (Kessler et al., 2005) likely fuels much of this skepticism. The DSM-5 is also criticized on the grounds that its diagnostic criteria have been loosened, thereby threatening to “turn our current diagnostic inflation into diagnostic hyperinflation” (Frances, 2012, para. 22). For example, DSM-IV specified that the symptoms of major depressive disorder must not be attributable to normal bereavement (loss of a loved one). The DSM-5, however, has removed this bereavement exclusion, essentially meaning that grief and sadness after a loved one’s death can constitute major depressive disorder. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/33%3A_Diagnosing_and_Classifying_Psychological_Disorders/33.01%3A_Introduction.txt |
A second classification system, the International Classification of Diseases (ICD), is also widely recognized. Published by the World Health Organization (WHO), the ICD was developed in Europe shortly after World War II and, like the DSM, has been revised several times. The categories of psychological disorders in both the DSM and ICD are similar, as are the criteria for specific disorders; however, some differences exist. Although the ICD is used for clinical purposes, this tool is also used to examine the general health of populations and to monitor the prevalence of diseases and other health problems internationally (WHO, 2019). The ICD is in its 11th edition (ICD-11); this latest edition was designed, in conjunction with the changes in DSM-5, to help harmonize the two classification systems as much as possible (APA, 2013).
A study that compared the use of the two classification systems found that worldwide the ICD is more frequently used for clinical diagnosis, whereas the DSM is more valued for research (Mezzich, 2002). Most research findings concerning the etiology and treatment of psychological disor- ders are based on criteria set forth in the DSM (Oltmanns & Castonguay, 2013). The DSM also includes more explicit disorder criteria, along with an extensive and helpful explanatory text (Regier et al., 2012). The DSM is the classification system of choice among U.S. mental health professionals, and this chapter is based on the DSM paradigm.
33.04: The Compassionate View Of Psychological Disorders
As these disorders are outlined, please bear two things in mind. First, remember that psychological disorders represent extremes of inner experience and behavior. If, while reading about these disorders, you feel that these descriptions begin to personally characterize you, do not worry—this moment of enlightenment probably means nothing more than you are normal. Each of us experiences episodes of sadness, anxiety, and preoccupation with certain thoughts—times when we do not quite feel ourselves. These episodes should not be considered problematic unless the accompanying thoughts and behaviors become extreme and have a disruptive effect on one’s life. Second, understand that people with psychological disorders are far more than just embodiments of their disorders. We do not use terms such as schizophrenics, depressives, or phobics because they are labels that objectify people who suffer from these conditions, thus promoting biased and disparaging assumptions about them. It is important to remember that a psychological disorder is not what a person is; it is something that a person has—through no fault of his or her own. As is the case with cancer or diabetes, those with psychological disorders suffer debilitating, often painful conditions that are not of their own choosing. These individuals deserve to be viewed and treated with compassion, understanding, and dignity.
REFERENCES
American Psychiatric Association. (2013). Diagnostic and statistical manual of mental disorders (5th ed.). doi.org/10.1176/appi. books.9780890425596
Clegg, J. W. (2012). Teaching about mental health and illness through the history of the DSM. History of Psychology, 15(4), 364–370. doi. org/10.1037/a0027249
Fisher, C. (2010, February 11). DSM-5 development process includes emphasis on gender and cultural sensitivity [Web log post]. BMED Report. http://www.bmedreport.com/archives/9359
Frances, A. (2012, December 2). DSM 5 is guide not bible—Ignore its ten worst changes [Web log post]. Psychology Today. www. psychologytoday.com/us/blog/dsm5-in-distress/201212/ dsm-5-is-guide-not-bible-ignore-its-ten-worst-changes
Kessler, R. C., Berglund, P., Demler, O., Jin, R., Merikangas, K. P., & Walters, E. F. (2005). Lifetime prevalence and age-of-onset distributions of DSM-IV disorders in the National Comorbidity Survey Replication. Archives of General Psychiatry, 62(6), 593–602. https:// doi.org/10.1001/archpsyc.62.6.593
Mayes, R., & Horowitz, A. V. (2005). DSM-III and the revolution in the classification of mental illness. Journal of the History of the Behavioral Sciences, 41(3), 249–267. doi.org/10.1002/ jhbs.20103
Mezzich, J. E. (2002). International surveys on the use of ICD-10 and related diagnostic systems. Psychopathology, 35(2–3), 72–75. https://doi.org/10.1159/000065122
National Comorbidity Survey. (2007). NCS-R lifetime prevalence estimates [Data file]. www.hcp.med.harvard.edu/ncs/index.php
National Institute on Drug Abuse. (2007, October). Comorbid drug use and mental illness: A research update from the National Institute on Drug Abuse. web.archive.org/web/20161221175505/https:// www.drugabuse.gov/sites/default/files/comorbid.pdf
Oltmanns, T. F., & Castonguay, L. G. (2013). General issues in understanding and treating psychopathology. In L. G. Castonguay & T. F. Oltmanns (Eds.), Psychopathology: From science to clinical practice (pp. 1–16). Guilford Press.
Regier, D. A., Kuhl, E. A., & Kupfer, D. A. (2012). DSM-5: Classification and criteria changes. World Psychiatry, 12(2), 92–98. doi. org/10.1002/wps.20050
Silverstein, C. (2009). The implications of removing homosexuality from the DSM as a mental disorder. Archives of Sexual Behavior, 38(2), 161–163. https://doi.org/10.1007/s10508-008-9442-x
The Associated Press. (2013, May 15). New psychiatric manual, DSM-5, faces criticism for turning “normal” human problems into mental illness. New York Daily News. http://www.nydailynews.com/life-style/ health/shrinks-critics-face-new-psychiatric-manual-article-1.1344935
World Health Organization (WHO). (2019). International statistical classification of diseases and related health problems (11th ed.). https://icd.who.int/ | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/33%3A_Diagnosing_and_Classifying_Psychological_Disorders/33.03%3A_The_International_Classification_Of_Diseases.txt |
Anxiety is a natural part of life and, at normal levels, helps us to function at our best. However, for people with anxiety disorders, anxiety is overwhelming and hard to control. Anxiety disorders develop out of a blend of biological (genetic) and psychological factors that, when combined with stress, may lead to the development of ailments. Primary anxiety-related diagnoses include generalized anxiety disorder, panic disorder, specific phobia, social anxiety disorder (social phobia), post traumatic stress disorder, and obsessive-compulsive disorder. In this module, we summarize the main clinical features of each of these disorders and discuss their similarities and differences with everyday experiences of anxiety.
What is anxiety? Most of us feel some anxiety almost every day of our lives. Maybe you have an important test coming up for school. Or maybe there’s that big game next Saturday, or that first date with someone new you are hoping to impress. Anxiety can be defined as a negative mood state that is accompanied by bodily symptoms such as increased heart rate, muscle tension, a sense of unease, and apprehension about the future (American Psychiatric Association [APA], 2013; Barlow, 2002).
Anxiety is what motivates us to plan for the future, and in this sense, anxiety is actually a good thing. It’s the nagging feeling that motivates us to study for that test, practice harder for that game, or be at our very best on that date. But some people experience anxiety so intensely that it is no longer helpful or useful. They may become so overwhelmed and distracted by anxiety that they actually fail their test, fumble the ball, or spend the whole date fidgeting and avoiding eye contact. If anxiety begins to interfere in the person’s life in a significant way, it is considered a disorder.
Anxiety and closely related disorders emerge from “triple vulnerabilities,” a combination of biological, psychological, and specific factors that increase our risk for developing a disorder (Barlow, 2002; Suárez et al., 2009). Biological vulnerabilities refer to specific genetic and neurobiological factors that might predispose someone to develop anxiety disorders. No single gene directly causes anxiety or panic, but our genes may make us more susceptible to anxiety and influence how our brains react to stress (Drabant et al., 2012; Gelernter & Stein, 2009; Smoller et al., 2009). Psychological vulnerabilities refer to the influences that our early experiences have on how we view the world. If we were confronted with unpredictable stressors or traumatic experiences at younger ages, we may come to view the world as unpredictable and uncontrollable, even dangerous (Chorpita & Barlow, 1998; Gunnar & Fisher, 2006). Specific vulnerabilities refer to how our experiences lead us to focus and channel our anxiety (Suárez et al., 2009). If we learned that physical illness is dangerous, maybe through witnessing our family’s reaction whenever anyone got sick, we may focus our anxiety on physical sensations. If we learned that disapproval from others has negative, even dangerous consequences, such as being yelled at or severely punished for even the slightest offense, we might focus our anxiety on social evaluation. If we learn that the “other shoe might drop” at any moment, we may focus our anxiety on worries about the future. None of these vulnerabilities directly causes anxiety disorders on its own—instead, when all of these vulnerabilities are present, and we experience some triggering life stress, an anxiety disorder may be the result (Barlow, 2002; Suárez et al., 2009). In the next sections, we will briefly explore each of the major anxiety-based disorders, found in the fifth edition of the Diagnostic and Statistical Manual of Mental Disorders (DSM-5) (APA, 2013). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/34%3A_Anxiety_and_Related_Disorders/34.01%3A_Introduction.txt |
Most of us worry some of the time, and this worry can actually be useful in helping us plan for the future or make sure we remember to do something important. Most of us can set aside our worries when we need to focus on other things or stop worrying altogether whenever a problem has passed. However, for someone with generalized anxiety disorder (GAD), these worries become difficult, or even impossible, to turn off. They may find themselves worrying excessively about a number of different things, both minor and catastrophic. Their worries also come with a host of other symptoms such as muscle tension, fatigue, agitation or restlessness, irritability, difficulties with sleep (either falling asleep, staying asleep, or both), or difficulty concentrating. The DSM-5 criteria specify that at least six months of excessive anxiety and worry of this type must be ongoing, happening more days than not for a good proportion of the day, to receive a diagnosis of GAD. About 5.7% of the population has met criteria for GAD at some point during their lifetime (Kessler, Berglund, et al., 2005), making it one of the most common anxiety disorders (see Table \(1\)).
Table \(1\): Prevalence Rates for Major Anxiety Disorders [1] Kessler, Berglund, et al. (2005), [2] Kessler, Chiu, et al. (2005), [3] Kessler et al. (1995), [4] Craske et al. (2006).
Disorder
1-Year Prevalence Rates [1]
Lifetime Prevalence Rates [2]
Prevalence by Gender
Median Age of Onset
Generalized Anxiety Disorder
3.1%
5.7%
67% female
31 years
OCD
1.0%
1.6%
55% female
19 years
Panic Disorder
2.7%
4.7%
67% female
24 years
PTSD
3.5%
6.8%
52% female [3]
23 years
Social Anxiety
6.8%
12.1%
50% female
13 years
Specific Phobia
8.7%
12.5%
60%–90% female [4]
7–9 years
What makes a person with GAD worry more than the average person? Research shows that individuals with GAD are more sensitive and vigilant toward possible threats than people who are not anxious (Aikins & Craske, 2001; Bar- low, 2002; Bradley et al., 1999). This may be related to early stressful experiences, which can lead to a view of the world as an unpredictable, uncontrollable, and even dangerous place. Some have suggested that people with GAD worry as a way to gain some control over these otherwise uncontrollable or unpredictable experiences and against uncertain outcomes (Dugas et al., 1998). By repeatedly going through all of the possible “What if?” scenarios in their mind, the person might feel like they are less vulnerable to an unexpected outcome, giving them the sense that they have some control over the situation (Wells, 2002). Others have suggested people with GAD worry as a way to avoid feeling distressed (Borkovec et al., 2004). For example, Borkovec and Hu (1990) found that those who worried when confronted with a stressful situation had less physiological arousal than those who didn’t worry, maybe because the worry “distracted” them in some way.
The problem is, all of this “what if?”-ing doesn’t get the person any closer to a solution or an answer and, in fact, might take them away from important things they should be paying attention to in the moment, such as finishing an important project. Many of the catastrophic outcomes people with GAD worry about are very unlikely to happen, so when the catastrophic event doesn’t materialize, the act of worrying gets reinforced (Borkovec et al., 1999). For example, if a mother spends all night worrying about whether her teenage daughter will get home safe from a night out and the daughter returns home without incident, the mother could easily attribute her daughter’s safe return to her successful “vigil.” What the mother hasn’t learned is that her daughter would have returned home just as safe if she had been focusing on the movie she was watching with her husband, rather than being preoccupied with worries. In this way, the cycle of worry is perpetuated, and, subsequently, people with GAD often miss out on many otherwise enjoyable events in their lives. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/34%3A_Anxiety_and_Related_Disorders/34.02%3A_Generalized_Anxiety_Disorder.txt |
Have you ever gotten into a near-accident or been taken by surprise in some way? You may have felt a flood of physical sensations, such as a racing heart, shortness of breath, or tingling sensations. This alarm reaction is called the “fight or flight” response (Cannon, 1929) and is your body’s natural reaction to fear, preparing you to either fight or escape in response to threat or danger. It’s likely you weren’t too concerned with these sensations, because you knew what was causing them. But imagine if this alarm reaction came “out of the blue,” for no apparent reason, or in a situation in which you didn’t expect to be anxious or fearful. This is called an “unexpected” panic attack or a false alarm. Because there is no apparent reason or cue for the alarm reaction, you might react to the sensations with intense fear, maybe thinking you are having a heart attack, or going crazy, or even dying. You might begin to associate the physical sensations you felt during this attack with this fear and may start to go out of your way to avoid having those sensations again.
Unexpected panic attacks such as these are at the heart of panic disorder (PD). However, to receive a diagnosis of PD, the person must not only have unexpected panic attacks but also must experience continued intense anxiety and avoidance related to the attack for at least one month, causing significant distress or interference in their lives. People with panic disorder tend to interpret even normal physical sensations in a catastrophic way, which triggers more anxiety and, ironically, more physical sensations, creating a vicious cycle of panic (Clark, 1986, 1996). The person may begin to avoid a number of situations or activities that produce the same physiological arousal that was present during the beginnings of a panic attack. For example, someone who experienced a racing heart during a panic attack might avoid exercise or caffeine. Someone who experienced choking sensations might avoid wearing high-necked sweaters or necklaces. Avoidance of these internal bodily or somatic cues for panic has been termed interoceptive avoidance (Barlow & Craske, 2007; Brown et al., 2005; Craske & Barlow, 2008; Shear et al., 1997).
The individual may also have experienced an overwhelming urge to escape during the unexpected panic attack. This can lead to a sense that certain places or situations—particularly situations where escape might not be possible—are not “safe.” These situations become external cues for panic. If the person begins to avoid several places or situations, or still endures these situations but does so with a significant amount of apprehension and anxiety, then the person also has agoraphobia (Barlow, 2002; Craske & Barlow, 1988, 2008). Agoraphobia can cause significant disruption to a person’s life, causing them to go out of their way to avoid situations, such as adding hours to a commute to avoid taking the train or only ordering take-out to avoid having to enter a grocery store. In one tragic case seen by our clinic, a woman suffering from agoraphobia had not left her apartment for 20 years and had spent the past 10 years confined to one small area of her apartment, away from the view of the outside. In some cases, agoraphobia develops in the absence of panic attacks and therefor is a separate disorder in DSM-5. But agoraphobia often accompanies panic disorder.
About 4.7% of the population has met criteria for PD or agoraphobia over their lifetime (Kessler, Chiu, et al., 2005; Kessler et al., 2006) (see Table 34.2.1). In all of these cases of panic disorder, what was once an adaptive natural alarm reaction now becomes a learned, and much feared, false alarm.
34.04: Specific Phobia
The majority of us might have certain things we fear, such as bees, or needles, or heights (Myers et al., 1984). But what if this fear is so consuming that you can’t go out on a summer’s day, or get vaccines needed to go on a special trip, or visit your doctor in her new office on the 26th floor? To meet criteria for a diagnosis of specific phobia, there must be an irrational fear of a specific object or situation that substantially interferes with the person’s ability to function. For example, a patient at our clinic turned down a prestigious and coveted artist residency because it required spending time near a wooded area, bound to have insects. Another patient purposely left her house two hours early each morning so she could walk past her neighbor’s fenced yard before they let their dog out in the morning.
The list of possible phobias is staggering, but four major subtypes of specific phobia are recognized: blood-injury-injection (BII) type, situational type (such as planes, elevators, or enclosed places), natural environment type for events one may encounter in nature (for example, heights, storms, and water), and animal type.
A fifth category, “other,” includes phobias that do not fit any of the four major subtypes (for example, fears of choking, vomiting, or contracting an illness). Most phobic reactions cause a surge of activity in the sympathetic nervous system and increased heart rate and blood pressure, maybe even a panic attack. However, people with BII type phobias usually experience a marked drop in heart rate and blood pressure and may even faint. In this way, those with BII phobias almost always differ in their physiological reaction from people with other types of phobia (Barlow & Liebowitz, 1995; Craske et al., 2006; Hofmann et al., 2009; Öst, 1992). BII phobia also runs in families more strongly than any phobic disorder we know (Antony & Barlow, 2002; Page & Martin, 1998). Specific phobia is one of the most common psychological disorders in the United States, with 12.5% of the population reporting a lifetime history of fears significant enough to be considered a “phobia” (Arrindell et al., 2003; Kessler, Berglund, et al., 2005) (see TABLE 34.2.1). Most people who suffer from specific phobia tend to have multiple phobias of several types (Hofmann et al., 1997). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/34%3A_Anxiety_and_Related_Disorders/34.03%3A_Panic_Disorder_And_Agoraphobia.txt |
With stories of war, natural disasters, and physical and sexual assault dominating the news, it is clear that trauma is a reality for many people. Many individual traumas that occur every day never even make the headlines, such as a car accident, domestic abuse, or the death of a loved one. Yet, while many people face traumatic events, not everyone who faces a trauma develops a disorder. Some, with the help of family and friends, are able to recover and continue on with their lives (Friedman, 2009). For some, however, the months and years following a trauma are filled with intrusive reminders of the event, a sense of intense fear that another traumatic event might occur, or a sense of isolation and emotional numbing. They may engage in a host of behaviors intended to protect themselves from being vulnerable or unsafe, such as constantly scanning their surroundings to look for signs of potential danger, never sitting with their back to the door, or never allowing themselves to be anywhere alone. This lasting reaction to trauma is what characterizes posttraumatic stress disorder (PTSD).
A diagnosis of PTSD begins with the traumatic event itself. An individual must have been exposed to an event that involves actual or threatened death, serious injury, or sexual violence. To receive a diagnosis of PTSD, exposure to the event must include either directly experiencing the event, witness- ing the event happening to someone else, learning that the event occurred to a close relative or friend, or having repeated or extreme exposure to details of the event (such as in the case of first responders). The person subsequently re-experiences the event through both intrusive memories and nightmares. Some memories may come back so vividly that the person feels like they are experiencing the event all over again, what is known as having a flashback. The individual may avoid anything that reminds them of the trauma, including conversations, places, or even specific types of people. They may feel emotionally numb or restricted in their ability to feel, which may interfere in their interpersonal relationships. The person may not be able to remember certain aspects of what happened during the event. They may feel a sense of a foreshortened future, that they will never marry, have a family, or live a long, full life. They may be jumpy or easily startled, hypervigilant to their surroundings, and quick to anger. The prevalence of PTSD among the population as a whole is relatively low, with 6.8% having experienced PTSD at some point in their life (Kessler, Berglund, et al., 2005) (see TABLE 34.2.1). Combat and sexual assault are the most common precipitating traumas (Kessler et al., 1995). Whereas PTSD was previously categorized as an Anxiety Disorder, in the most recent version of the DSM (DSM-5) (APA, 2013) it has been reclassified under the more specific category of Trauma- and Stressor-Related Disorders.
A person with PTSD is particularly sensitive to both internal and external cues that serve as reminders of their traumatic experience. For example, as we saw in PD, the physical sensations of arousal present during the initial trauma can become threatening in and of themselves, becoming a powerful reminder of the event. Someone might avoid watching intense or emotional movies in order to prevent the experience of emotional arousal. Avoidance of conversations, reminders, or even of the experience of emotion itself may also be an attempt to avoid triggering internal cues. External stimuli that were present during the trauma can also become strong triggers. For example, if a woman is raped by a man wearing a red t-shirt, she may develop a strong alarm reaction to the sight of red shirts, or perhaps even more indiscriminately to anything with a similar color red. A combat veteran who experienced a strong smell of gasoline during a road- side bomb attack may have an intense alarm reaction when pumping gas back at home. Individuals with a psychological vulnerability toward viewing the world as uncontrollable and unpredictable may particularly struggle with the possibility of additional future, unpredictable traumatic events, fueling their need for hypervigilance and avoidance and perpetuating the symptoms of PTSD. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/34%3A_Anxiety_and_Related_Disorders/34.06%3A_Posttraumatic_Stress_Disorder.txt |
Have you ever had a strange thought pop into your mind, such as picturing the stranger next to you naked? Or maybe you walked past a crooked picture on the wall and couldn’t resist straightening it. Most people have occasional strange thoughts and may even engage in some “compulsive” behaviors, especially when they are stressed (Boyer & Liénard, 2008; Fullana et al., 2009). But for most people, these thoughts are nothing more than a passing oddity, and the behaviors are done (or not done) without a second thought. For someone with obsessive-compulsive disorder (OCD), however, these thoughts and compulsive behaviors don’t just come and go. Instead, strange or unusual thoughts are taken to mean something much more important and real, maybe even something dangerous or frightening. The urge to engage in some behavior, such as straightening a picture, can become so intense that it is nearly impossible not to carry it out, or causes significant anxiety if it can’t be carried out. Further, someone with OCD might become preoccupied with the possibility that the behavior wasn’t carried out to completion and feel compelled to repeat the behavior again and again, maybe several times before they are “satisfied.”
To receive a diagnosis of OCD, a person must experience obsessive thoughts and/or compulsions that seem irrational or nonsensical but that keep coming into their mind. Some examples of obsessions include doubting thoughts (such as doubting a door is locked or an appliance is turned off), thoughts of contamination (such as thinking that touching almost anything might give you cancer), or aggressive thoughts or images that are unprovoked or nonsensical. Compulsions may be carried out in an attempt to neutralize some of these thoughts, providing temporary relief from the anxiety the obsessions cause, or they may be nonsensical in and of themselves. Either way, compulsions are distinct in that they must be repetitive or excessive, the person feels “driven” to carry out the behavior, and the person feels a great deal of distress if they can’t engage in the behavior. Some examples of compulsive behaviors are repetitive washing (often in response to contamination obsessions), repetitive checking (locks, door handles, appliances, often in response to doubting obsessions), ordering and arranging things to ensure symmetry, or doing things according to a specific ritual or sequence (such as getting dressed or ready for bed in a specific order). To meet diagnostic criteria for OCD, engaging in obsessions and/or compulsions must take up a significant amount of the person’s time, at least an hour per day, and must cause significant distress or impairment in functioning. About 1.6% of the population has met criteria for OCD over the course of a lifetime (Kessler, Berglund, et al., 2005) (see TABLE 34.2.1). Whereas OCD was previously categorized as an Anxiety Disorder, in the most recent version of the DSM (DSM-5) (APA, 2013) it has been reclassified under the more specific category of Obsessive-Compulsive and Related Disorders.
People with OCD often confuse having an intrusive thought with their potential for carrying out the thought. Whereas most people when they have a strange or frightening thought are able to let it go, a person with OCD may become “stuck” on the thought and be intensely afraid that they might somehow lose control and act on it. Or worse, they believe that having the thought is just as bad as doing it. This is called thought-action fusion. For example, one patient of ours was plagued by thoughts that she would cause harm to her young daughter. She experienced intrusive images of throwing hot coffee in her daughter’s face or pushing her face underwater when she was giving her a bath. These images were so terrifying to the patient that she would no longer allow herself any physical contact with her daughter and would leave her daughter in the care of a babysitter if her husband or another family was not available to “supervise” her. In reality, the last thing she wanted to do was harm her daughter, and she had no intention or desire to act on the aggressive thoughts and images, nor does anybody with OCD act on these thoughts, but these thoughts were so horrifying to her that she made every attempt to prevent herself from the potential of carrying them out, even if it meant not being able to hold, cradle, or cuddle her daughter. These are the types of struggles people with OCD face every day. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/34%3A_Anxiety_and_Related_Disorders/34.07%3A_Obsessive-Compulsive_Disorder.txt |
Many successful treatments for anxiety and related disorders have been developed over the years. Medications (anti-anxiety drugs and antidepressants) have been found to be beneficial for disorders other than specific phobia, but relapse rates are high once medications are stopped (Heimberg et al., 1998; Hollon et al., 2005), and some classes of medications (minor tranquilizers or benzodiazepines) can be habit forming.
Exposure-based cognitive-behavioral therapies (CBT) are effective psychosocial treatments for anxiety disorders, and many show greater treatment effects than medication in the long term (Barlow et al., 2000, 2007). In CBT, patients are taught skills to help identify and change problematic thought processes, beliefs, and behaviors that tend to worsen symptoms of anxiety, and practice applying these skills to real-life situations through exposure exercises. Patients learn how the automatic “appraisals” or thoughts they have about a situation affect both how they feel and how they behave. Similarly, patients learn how engaging in certain behaviors, such as avoiding situations, tends to strengthen the belief that the situation is something to be feared. A key aspect of CBT is exposure exercises, in which the patient learns to gradually approach situations they find fearful or distressing, in order to challenge their beliefs and learn new, less fearful associations about these situations.
Typically 50% to 80% of patients receiving drugs or CBT will show a good initial response, with the effect of CBT more durable. Newer developments in the treatment of anxiety disorders are focusing on novel interventions, such as the use of certain medications to enhance learning during CBT (Otto et al., 2010) and transdiagnostic treatments targeting core, underlying vulnerabilities (Barlow et al., 2011). As we advance our understanding of anxiety and related disorders, so too will our treatments advance, with the hopes that for the many people suffering from these disorders, anxiety can once again become something useful and adaptive, rather than something debilitating.
REFERENCES
American Psychiatric Association. (2013). Diagnostic and statistical manual of mental disorders (5th ed.). https://doi.org/10.1176/appi. books.9780890425596
Aikins, D. E., & Craske, M. G. (2001). Cognitive theories of generalized anxiety disorder. Psychiatric Clinics of North America, 24(1), 57–74. https://doi.org/10.1016/S0193-953X(05)70206-9
Antony, M. M., & Barlow, D. H. (2002). Specific phobias. In D. H. Barlow (Ed.), Anxiety and its disorders: The nature and treatment of anxiety and panic (2nd ed.). Guilford Press.
Arrindell, W. A., Eisemann, M., Richter, J., Oei, T. P. S., Caballo, V. E., van der Ende, J., Sanavio, E., Bagés, N., Feldman, L., Torres, B., Sica, C., Iwawaki, S., Edelmann, R. J., Crozier, W. R., Furnham, A., Hudson,
B. L., & Cultural Clinical Psychology Study Group. (2003). Phobic anxiety in 11 nations: Part I: Dimensional constancy of the five-factor model. Behaviour Research and Therapy, 41(4), 461–479. https://doi. org/10.1016/S0005-7967(02)00047-5
Barlow, D. H. (2002). Anxiety and its disorders: The nature and treatment of anxiety and panic (2nd ed.). Guilford Press.
Barlow, D. H., & Craske, M. G. (2007). Mastery of your anxiety and panic (4th ed.). Oxford University Press.
Barlow, D. H., & Liebowitz, M. R. (1995). Specific and social phobias. In H. I. Kaplan & B. J. Sadock (Eds.), Comprehensive textbook of psychiatry/VI (pp. 1204–1217). Williams & Wilkins.
Barlow, D. H., Allen, L. B., & Basden, S. (2007). Pscyhological treatments for panic disorders, phobias, and generalized anxiety disorder. In
P. E. Nathan & J. M. Gorman (Eds.), A guide to treatments that work (3rd ed.). Oxford University Press.
Barlow, D. H., Ellard, K. K., Fairholme, C. P., Farchione, T. J., Boisseau,
C. L., Allen, L. B., & Ehrenreich-May, J. (2011). Unified protocol for transdiagnostic treatment of emotional disorders: Workbook. Oxford University Press.
Barlow, D. H., Gorman, J. M., Shear, M. K., & Woods, S. W. (2000). Cognitive-behavioral therapy, imipramine, or their combination for panic disorder: A randomized controlled trial. Journal of the American Medical Association, 283(19), 2529–2536. https://doi.org/ 10.1001/jama.283.19.2529
Bögels, S. M., Alden, L., Beidel, D. C., Clark, L. A., Pine, D. S., Stein, M. B., & Voncken, M. (2010). Social anxiety disorder: questions and answers for the DSM-V. Depression and Anxiety, 27(2), 168–189. https://doi. org/10.1002/da.20670
Borkovec, T. D., & Hu, S. (1990). The effect of worry on cardiovascular response to phobic imagery. Behaviour Research and Therapy, 28(1), 69–73. https://doi.org/10.1016/0005-7967(90)90056-O
Borkovec, T. D., Alcaine, O. M., & Behar, E. (2004). Avoidance theory of worry and generalized anxiety disorder. In R. G. Heimberg, C. L. Turk, & D. S. Mennin (Eds.), Generalized anxiety disorder: Advances in research and practice (pp. 77–108). Guilford Press.
Borkovec, T. D., Hazlett-Stevens, H., & Diaz, M. L. (1999). The role of positive beliefs about worry in generalized anxiety disorder and its treatment. Clinical Psychology & Pscyhotherapy, 6(2), 69–73. https:// doi.org/10.1002/(SICI)1099-0879(199905)6:2%3C126::AID- CPP193%3E3.0.CO;2-M
Boyer, P., & Liénard, P. (2008). Ritual behavior in obsessive and normal individuals: Moderating anxiety and reorganizing the flow of action. Current Directions in Psychological Science, 17(4), 291–294. https:// doi.org/10.1111/j.1467-8721.2008.00592.x
Bradley, B. P., Mogg, K., White, J., Groom, C., & de Bono, J. (1999). Attentional bias for emotional faces in generalized anxiety disorder. British Journal of Clinical Psychology, 38(3), 267–278. https://doi. org/10.1348/014466599162845
Brown, T. A., White, K. S., & Barlow, D. H. (2005). A psychometric reanalysis of the Albany Panic and Phobia Questionnaire. Behaviour Research and Therapy, 43(3), 337–355. https://doi.org/10.1016/j. brat.2004.03.004
Bruch, M. A., & Heimberg, R. G. (1994). Differences in perceptions of parental and personal characteristics between generalized and nongeneralized social phobics. Journal of Anxiety Disorders, 8(2), 155–168. https://doi.org/10.1016/0887-6185(94)90013-2
Cannon, W. B. (1929). Bodily changes in pain, hunger, fear and rage. Appleton.
Chorpita, B. F., & Barlow, D. H. (1998). The development of anxiety: the role of control in the early environment. Psychological Bulletin, 124(1), 3–21. https://doi.org/10.1037/0033-2909.124.1.3
Clark, D. M. (1996). Panic disorder: From theory to therapy. In P. Salkovskis (Ed.), Fronteirs of cognitive therapy (pp. 318–344). Guilford Press.
Clark, D. M. (1986). A cognitive approach to panic. Behaviour Research and Therapy, 24(4), 461–470. https://doi.org/10.1016/ 0005-7967(86)90011-2
Craske, M. G., & Barlow, D. H. (1988). A review of the relationship between panic and avoidance. Clinical Pscyhology Review, 8(6), 667–685. https://doi.org/10.1016/0272-7358(88)90086-4
Craske, M. G., & Barlow, D. H. (2008). Panic disorder and agoraphobia. Guilford Press.
Craske, M. G., Antony, M. M., & Barlow, D. H. (2006). Mastering your fears and phobias: Therapist guide. Oxford University Press.
Drabant, E. M., Ramel, W., Edge, M. D., Hyde, L. W., Kuo, J. R., Goldin,
P. R., Hariri, A. R., & Gross, J. J. (2012). Neural mechanisms underlying 5-HTTLPR-related sensitivity to acute stress. American Journal of Psychiatry, 169(4), 397–405. https://doi.org/10.1176/appi. ajp.2011.10111699
Dugas, M. J., Gagnon, F., Ladouceur, R., & Freeston, M. H. (1998). Generalized anxiety disorder: a preliminary test of a conceptual model. Behaviour Research and Therapy, 36(2), 215–226. https://doi. org/10.1016/S0005-7967(97)00070-3
Friedman, M. J. (2009). Phenomenology of postraumatic stress disorder and acute stress disorder. In M. M. Anthony & M. B. Stein (Eds.), Oxford Handbook of Anxiety and Related Disorders. Oxford University Press.
Fullana, M. A., Mataix-Cols, D., Caspi, A., Harrington, H., Grisham, J. R., Moffitt, T. E., & Poulton, R. (2009). Obsessions and compulsions in the community: prevalence, interference, help-seeking, developmental stability, and co-occurring psychiatric conditions. American Journal of Psychiatry, 166(3), 329–336. https://doi.org/10.1176/appi. ajp.2008.08071006
Gelernter, J., & Stein, M. B. (2009). Heritability and genetics of anxiety disorders. In M. M. Antony & M. B. Stein (Eds.), Oxford handbook of anxiety and related disorders. Oxford University Press.
Gunnar, M. R., & Fisher, P. A. (2006). Bringing basic research on early experience and stress neurobiology to bear on preventive interventions for neglected and maltreated children. Developmental Psychopathology, 18(3), 651–677. https://doi.org/10.1017/ S0954579406060330
Heimberg, R. G., Liebowitz, M. R., Hope, D. A., Schneier, F. R., Holt, C. S., Welkowitz, L. A., Juster, H. R., Campeas, R., Bruch, M. A., Cloitre, M., Fallon, B., & Klein, D. F. (1998). Cognitive behavioral group therapy vs phenelzine therapy for social phobia: 12-week outcome. Archives of General Psychiatry, 55(12), 1133–1141. https://doi.org/10.1001/ archpsyc.55.12.1133
Hofmann, S. G., Alpers, G. W., & Pauli, P. (2009). Phenomenology of panic and phobic disorders. In M. M. Antony & M. B. Stein (Eds.), Oxford handbook of anxiety and related disorders (pp. 34–46). Oxford University Press.
Hofmann, S. G., Lehman, C. L., & Barlow, D. H. (1997). How specific are specific phobias? Journal of Behavior Therapy and Experimental Psychiatry, 28(3), 233–240. https://doi.org/10.1016/ S0005-7916(97)00012-8
Hollon, S. D., DeRubeis, R. J., Shelton, R. C., Amsterdam, J. D., Salomon, R. M., O’Reardon, J. P., Lovett, M. L., Young, P. R., Haman, K. L., Freeman, B. B., & Gallop, R. (2005). Prevention of relapse following cognitive therapy vs medications in moderate to severe depression. Archives of General Psychiatry, 62(4), 417–422. https://doi.org/ 10.1001/archpsyc.62.4.417
Kessler, R. C., Berglund, P., Demler, O., Jin, R., Merikangas, K. R., & Walters, E. E. (2005). Lifetime prevalence and age-of-onset distributions of DSM-IV disorders in the National Comorbidity Survey Replication. Archives of General Psychiatry, 62(6), 593–602. https:// doi.org/10.1001/archpsyc.62.6.593
Kessler, R. C., Chiu, W. T., Demler, O., Merikangas, K. R., & Walters, E. E. (2005). Prevalence, severity, and comorbidity of 12-month DSM-IV disorders in the National Comorbidity Survey Replication. Archives of General Psychiatry, 62(6), 617–627. https://doi.org/10.1001/ archpsyc.62.6.617
Kessler, R. C., Chiu, W. T., Jin, R., Ruscio, A. M., Shear, K., & Walters, E. E. (2006). The epidemiology of panic attacks, panic disorder, and agoraphobia in the National Comorbidity Survey Replication. Archives of General Psychiatry, 63(4), 415–424. https://doi.org/10.1001/ archpsyc.63.4.415
Kessler, R. C., Sonnega, A., Bromet, E., Hughes, M., & Nelson, C. B. (1995). Posttraumatic stress disorder in the National Comorbidity Survey. Archives of General Psychiatry, 52(12), 1048–1060. https:// doi.org/10.1001/archpsyc.1995.03950240066012
Lieb, R., Wittchen, H.-U., Höfler, M., Fuetsch, M., Stein, M. B., & Merikangas, K. R. (2000). Parental psychopathology, parenting styles, and the risk of social phobia in offspring: a prospective- longitudinal community study. Archives of General Psychiatry, 57(9), 859–866. https://doi.org/10.1001/archpsyc.57.9.859
McCabe, R. E., Antony, M. M., Summerfeldt, L. J., Liss, A., & Swinson, R. P. (2003). Preliminary examination of the relationship between anxiety disorders in adults and self-reported history of teasing or bullying experiences. Cognitive Behavior Therapy, 32(4), 187–193. https://doi. org/10.1080/16506070310005051
Myers, J. K., Weissman, M. M., Tischler, G. L., Holzer, C. E., III, Leaf, P. J., Orvaschel, H., Anthony, J. C., Boyd, J. H., Burke, J. D., Jr., Kramer, M., & Stoltzman, R. (1984). Six-month prevalence of psychiatric disorders in three communities. Archives of General Psychiatry, 41(10), 959–967. https://doi.org/10.1001/archpsyc.1984.01790210041006
Öst, L.-G. (1992). Blood and injection phobia: Background and cognitive, physiological, and behavioral variables. Journal of Abnormal Psychology, 101(1), 68–74. https://dx.doi.org/10.1037/0021-843X.101.1.68
Otto, M. W., Tolin, D. F., Simon, N. M., Pearlson, G. D., Basden, S., Meunier, S. A., Hofmann, S. G., Eisenmenger, K., Krystal, J. H., & Pollack, M. H. (2010). Efficacy of d-cycloserine for enhancing response to cognitive- behavior therapy for panic disorder. Biological Psychiatry, 67(4), 365–370. https://doi.org/10.1016/j.biopsych.2009.07.036
Page, A. C., & Martin, N. G. (1998). Testing a genetic structure of blood-injury-injection fears. American Journal of Medical Genetics, 81(5), 377–384. https://doi.org/10.1002/(SICI)1096-8628(19980907) 81:5%3C377::AID-AJMG5%3E3.0.CO;2-Q
Rapee, R. M., & Melville, L. F. (1997). Recall of family factors in social phobia and panic disorder: comparison of mother and offspring reports. Depression and Anxiety, 5(1), 7–11. https://doi.org/10.1002/ (SICI)1520-6394(1997)5:1%3C7::AID-DA2%3E3.0.CO;2-E
Schneier, F. R., Leibowitz, M. R., Beidel, D. C., J., Fyer A., George, M. S., Heimberg, R. G., . . . Versiani, M. (1996). Social phobia. In T. A. Widiger, A. J. Frances, H. A. Pincus, R. Ross, M. B. First & W. W. Davis (Eds.), DSM-IV sourcebook (Vol. 2, pp. 507–548). American Psychiatric Association.
Shear, M. K., Brown, T. A., Barlow, D. H., Money, R., Sholomskas, D. E., Woods, S. W., Gorman, J. M., & Papp, L. A. (1997). Multicenter collaborative panic disorder severity scale. American Journal of Psychiatry, 154(11), 1571–1575. https://doi.org/10.1176/ajp.154.11.1571
Smoller, J. W., Block, S. R., & Young, M. M. (2009). Genetics of anxiety disorders: the complex road from DSM to DNA. Depression and Anxiety, 26(11), 965–975. https://doi.org/10.1002/da.20623
Suárez, L, Bennett, S., Goldstein, C., & Barlow, D. H. (2009). Understanding anxiety disorders from a “triple vulnerabilities” framework. In M. M. Antony & M. B. Stein (Eds.), Oxford handbook of anxiety and related disorders (pp. 153–172). Oxford University Press.
Wells, A. (2002). GAD, metacognition, and mindfulness: An information processing analysis. Clinical Psychology: Science and Practice, 9(1), 95–100. https://doi.org/10.1093/clipsy.9.1.95 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/34%3A_Anxiety_and_Related_Disorders/34.08%3A_Treaments_For_Anxiety_And_Related_Disorders.txt |
Everyone feels down or euphoric from time to time, but this is different from having a mood disorder such as major depressive disorder or bipolar disorder. Mood disorders are extended periods of depressed, euphoric, or irritable moods that in combination with other symptoms cause the person significant distress and interfere with his or her daily life, often resulting in social and occupational difficulties. In this module, we describe major mood disorders, including their symptom presentations, general prevalence rates, and how and why the rates of these disorders tend to vary by age, gender, and race. In addition, biological and environmental risk factors that have been implicated in the development and course of mood disorders, such as heritability and stressful life events, are reviewed. Finally, we provide an overview of treatments for mood disorders, covering treatments with demonstrated effectiveness, as well as new treatment options showing promise.
The actress Brooke Shields (2005) published a memoir titled Down Came the Rain: My Journey through Postpartum Depression, in which she described her struggles with depression following the birth of her daughter. Despite the fact that about one in 20 women experience depression after the birth of a baby (American Psychiatric Association [APA], 2013), postpartum depression—recently renamed “perinatal depression”—continues to be veiled by stigma, owing in part to a widely held expectation that motherhood should be a time of great joy. In an opinion piece in the New York Times, Shields revealed that entering motherhood was a profoundly overwhelming experience for her. She vividly describes experiencing a sense of “doom” and “dread” in response to her newborn baby. Because motherhood is conventionally thought of as a joyous event and not associated with sadness and hopelessness, responding to a newborn baby in this way can be shocking to the new mother as well as those close to her. It may also involve a great deal of shame for the mother, making her reluctant to divulge her experience to others, including her doctors and family.
Feelings of shame are not unique to perinatal depression. Stigma applies to other types of depressive and bipolar disorders and contributes to people not always receiving the necessary support and treatment for these disorders. In fact, the World Health Organization ranks both major depressive disorder (MDD) and bipolar disorder (BD) among the top 10 leading causes of disability worldwide. Further, MDD and BD carry a high risk of suicide. It is estimated that 25% to 50% of people diagnosed with BD will attempt suicide at least once in their lifetimes (Goodwin & Jamison, 2007).
35.02: What Are Mood Didorders
Mood Episodes
Everyone experiences brief periods of sadness, irritability, or euphoria. This is different than having a mood disorder, such as MDD or BD, which are characterized by a constellation of symptoms that causes people significant distress or impairs their everyday functioning.
Major Depressive Episode
A major depressive episode (MDE) refers to symptoms that co-occur for at least two weeks and cause significant distress or impairment in functioning, such as interfering with work, school, or relationships. Core symptoms include feeling down or depressed or experiencing anhedonia—loss of interest or pleasure in things that one typically enjoys. According to the fifth edition of the Diagnostic and Statistical Manual (DSM-5) (APA, 2013), the criteria for an MDE require five or more of the following symptoms, including one or both of the first two symptoms, for most of the day, nearly every day:
1. Depressed mood
2. Diminished interest or pleasure in almost all activities
3. Significant weight loss or gain or an increase or decrease in appetite
4. Insomnia or hypersomnia
5. Psychomotor agitation or retardation
6. Fatigue or loss of energy
7. Feeling worthless or excessive or inappropriate guilt
8. Diminished ability to concentrate or indecisiveness
9. Recurrent thoughts of death, suicidal ideation, or a suicide attempt
These symptoms cannot be caused by physiological effects of a substance or a general medical condition (e.g., hypothyroidism).
Manic or Hypomanic Episode
The core criterion for a manic or hypomanic episode is a distinct period of abnormally and persistently euphoric, expansive, or irritable mood and persistently increased goal-directed activity or energy. The mood disturbance must be present for one week or longer in mania (unless hospitalization is required) or four days or longer in hypomania. Concurrently, at least three of the following symptoms must be present in the context of euphoric mood (or at least four in the context of irritable mood):
1. Inflated self-esteem or grandiosity
2. Increased goal-directed activity or psychomotor agitation
3. Reduced need for sleep
4. Racing thoughts or flight of ideas
5. Distractibility
6. Increased talkativeness
7. Excessive involvement in risky behaviors
Manic episodes are distinguished from hypomanic episodes by their duration and associated impairment; whereas manic episodes must last one week and are defined by a significant impairment in functioning, hypomanic episodes are shorter and not necessarily accompanied by impairment in functioning.
Mood Disorders
Unipolar Mood Disorders
Two major types of unipolar disorders described by the DSM-5 (APA, 2013) are major depressive disorder and persistent depressive disorder (PDD; dysthymia). MDD is defined by one or more MDEs but no history of manic or hypomanic episodes. Criteria for PDD are feeling depressed most of the day for more days than not, for at least two years. At least two of the following symptoms are also required to meet criteria for PDD:
1. Poor appetite or overeating
2. Insomnia or hypersomnia
3. Low energy or fatigue
4. Low self-esteem
5. Poor concentration or difficulty making decisions
6. Feelings of hopelessness
Like MDD, these symptoms need to cause significant distress or impairment and cannot be due to the effects of a substance or a general medical condition. To meet criteria for PDD, a person cannot be without symptoms for more than two months at a time. PDD has overlapping symptoms with MDD. If someone meets criteria for an MDE during a PDD episode, the person will receive diagnoses of PDD and MDD.
Bipolar Mood Disorders
Three major types of BDs are described by the DSM-5 (APA, 2013). Bipolar I Disorder (BD I), which was previously known as manic-depression, is characterized by a single (or recurrent) manic episode. A depressive episode is not necessary but commonly present for the diagnosis of BD I. Bipolar II Disorder is characterized by single (or recurrent) hypomanic episodes and depressive episodes. Another type of BD is cyclothymic disorder, characterized by numerous and alternating periods of hypomania and depression, lasting at least two years. To qualify for cyclothymic disorder, the periods of depression cannot meet full diagnostic criteria for an MDE; the person must experience symptoms at least half the time with no more than two consecutive symptom-free months; and the symptoms must cause significant distress or impairment.
It is important to note that the DSM-5 was published in 2013, and findings based on the updated manual will be forthcoming. Consequently, the research presented below was largely based on a similar, but not identical, conceptualization of mood disorders drawn from the DSM-IV (APA, 2000). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/35%3A_Mood_Disorders/35.01%3A_Introduction.txt |
Depressive Disorders
In a nationally representative sample, lifetime prevalence rate for MDD is 16.6% (Kessler et al., 2005). This means that nearly one in five Americans will meet the criteria for MDD during their lifetime. The 12-month prevalence—the proportion of people who meet criteria for a disorder during a 12-month period—for PDD is approximately 0.5% (APA, 2013).
Although the onset of MDD can occur at any time throughout the lifespan, the average age of onset is mid- 20s, with the age of onset decreasing with people born more recently (APA, 2000). Prevalence of MDD among older adults is much lower than it is for younger cohorts (Kessler et al., 2010). The duration of MDEs varies widely. Recovery begins within three months for 40% of people with MDD and within 12 months for 80% (APA, 2013). MDD tends to be a recurrent disorder with about 40% to 50% of those who experience one MDE experiencing a second MDE (Monroe & Harkness, 2011). An earlier age of onset predicts a worse course.
About 5% to 10% of people who experience an MDE will later experience a manic episode (APA, 2000), thus no longer meeting criteria for MDD but instead meeting them for BD I. Diagnoses of other disorders across the lifetime are common for people with MDD: 59% experience an anxiety disorder; 32% experience an impulse control disorder, and 24% experi- ence a substance use disorder (Kessler et al., 2007).
Women experience two to three times higher rates of MDD than do men (Nolen-Hoeksema & Hilt, 2009). This gender difference emerges during puberty (Conley & Rudolph, 2009). Before puberty, boys exhibit similar or higher prevalence rates of MDD than do girls (Twenge & Nolen-Hoeksema, 2002). MDD is inversely correlated with socioeconomic status (SES), a person’s economic and social position based on income, education, and occupation. Higher prevalence rates of MDD are associated with lower SES (Lorant et al., 2003), particularly for adults over 65 years old (Kessler et al., 2010). Independent of SES, results from a nationally representative sample found that European Americans had a higher prevalence rate of MDD than did African Americans and Hispanic Americans, whose rates were similar (Breslau et al., 2006). The course of MDD for African Americans is often more severe and less often treated than it is for European Americans, how- ever (Williams et al., 2007). Native Americans have a higher prevalence rate than do European Americans, African Americans, or Hispanic Americans (Hasin et al., 2005). Depression is not limited to industrialized or western cultures; it is found in all countries that have been examined, although the symptom presentation as well as prevalence rates vary across cultures (Chentsova-Dutton & Tsai, 2009).
SPECIFIERS
Both MDEs and manic episodes can be further described using standardized tags based on the timing of, or other symptoms that are occurring during, the mood episode, to increase diagnostic specificity and inform treatment. Psychotic features is specified when the episodes are accompanied by delusions (rigidly held beliefs that are false) or hallucinations (perceptual disturbances that are not based in reality). Seasonal pattern is specified when a mood episode occurs at the same time of the year for two consecutive years—most commonly occurring in the fall and winter. Peripartum onset is specified when a mood episode has an onset during pregnancy or within four weeks of the birth of a child. Approximately 3% to 6% of women who have a child experience an MDE with peripartum onset (APA, 2013). This is less frequent and different from the baby blues or when women feel transient mood symptoms usually within 10 days of giving birth, which are experienced by most women (Nolen-Hoeksema & Hilt, 2009). ■
Bipolar Disorders
The lifetime prevalence rate of bipolar spectrum disorders in the general U.S. population is estimated at approximately 4.4%, with BD I constituting about 1% of this rate (Merikangas et al., 2007). Prevalence estimates, however, are highly dependent on the diagnostic procedures used (e.g., interviews vs. self-report) and whether or not sub-threshold forms of the disorder are included in the estimate. BD often co-occurs with other psychiatric disorders. Approximately 65% of people with BD meet diagnostic criteria for at least one additional psychiatric disorder, most commonly anxiety disorders and substance use disorders (McElroy et al., 2001). The co-occurrence of BD with other psychiatric disorders is associated with poorer illness course, including higher rates of suicidality (Leverich et al., 2003). A recent cross-national study sample of more than 60,000 adults from 11 countries estimated the worldwide prevalence of BD at 2.4%, with BD I constituting 0.6% of this rate (Merikangas et al., 2011). In this study, the prevalence of BD varied somewhat by country. Whereas the United States had the highest lifetime prevalence (4.4%), India had the lowest (0.1%). Variation in prevalence rates was not necessarily related to SES, as in the case of Japan, a high-income country with a very low prevalence rate of BD (0.7%).
With regard to ethnicity, data from studies not confounded by SES or inaccuracies in diagnosis are limited, but available reports suggest rates of BD among European Americans are similar to those found among African Americans (Blazer et al., 1985) and Hispanic Americans (Breslau et al., 2005). Another large community-based study found that although prevalence rates of mood disorders were similar across ethnic groups, Hispanic Americans and African Americans with a mood disorder were more likely to remain persistently ill than European Americans (Breslau et al., 2005). Compared with European Americans with BD, African Americans tend to be underdiagnosed for BD (and over-diagnosed for schizophrenia) (Kilbourne et al., 2004; Minsky et al., 2003), and Hispanic Americans with BD have been shown to receive fewer psychiatric medication prescriptions and specialty treatment visits (Gonzalez et al., 2007). Misdiagnosis of BD can result in the underutilization of treatment or the utilization of inappropriate treatment, and thus profoundly impact the course of illness.
As with MDD, adolescence is known to be a significant risk period for BD; mood symptoms start by adolescence in roughly half of BD cases (Leverich et al., 2007; Perlis et al., 2004). Longitudinal studies show that those diagnosed with BD prior to adulthood experience a more pernicious course of illness relative to those with adult onset, including more episode recurrence, higher rates of suicidality, and profound social, occupational, and economic repercussions (e.g., Lewinsohn et al., 2002). The prevalence of BD is substantially lower in older adults compared with younger adults (1% vs. 4%) (Merikangas et al., 2007). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/35%3A_Mood_Disorders/35.03%3A_How_Common_Are_Mood_Disorders_Who_Develops_Mood_Disorders.txt |
Mood disorders are complex disorders resulting from multiple factors. Causal explanations can be attempted at various levels, including biological and psychosocial levels. Below, several of the key factors that contribute to onset and course of mood disorders are highlighted.
Depressive Disorders
Research across family and twin studies has provided support that genetic factors are implicated in the development of MDD. Twin studies suggest that familial influence on MDD is mostly due to genetic effects and that individual-specific environmental effects (e.g., romantic relationships) play an important role, too. By contrast, the contribution of shared environmental effect by siblings is negligible (Sullivan et al., 2000). The mode of inheritance is not fully understood, although no single genetic variation has been found to increase the risk of MDD significantly. Instead, several genetic variants and environmental factors most likely contribute to the risk for MDD (Lohoff, 2010).
One environmental stressor that has received much support in relation to MDD is stressful life events. In particular, severe stressful life events—those that have long-term consequences and involve loss of a significant relationship (e.g., divorce) or economic stability (e.g., unemployment) are strongly related to depression (Brown & Harris, 1989; Monroe et al., 2009). Stressful life events are more likely to predict the first MDE than subsequent episodes (Lewinsohn et al., 1999). In contrast, minor events may play a larger role in subsequent episodes than the initial episodes (Monroe & Harkness, 2005).
Depression research has not been limited to examining reactivity to stressful life events. Much research, particularly brain imaging research using functional magnetic resonance imaging (fMRI), has centered on examining neural circuitry—the interconnections that allow multiple brain regions to perceive, generate, and encode information in concert. A meta-analysis of neuroimaging studies showed that when viewing negative stimuli (e.g., picture of an angry face, picture of a car accident), compared with healthy control participants, participants with MDD have greater activation in brain regions involved in stress response and reduced activation of brain regions involved in positively motivated behaviors (Hamilton et al., 2012).
Other environmental factors related to increased risk for MDD include experiencing early adversity (e.g., childhood abuse or neglect) (Spatz Widom et al., 2007), chronic stress (e.g., poverty), and interpersonal factors. For example, marital dissatisfaction predicts increases in depressive symptoms in both men and women. On the other hand, depressive symptoms also predict increases in marital dissatisfaction (Whisman & Uebelacker, 2009). Research has found that people with MDD generate some of their interpersonal stress (Hammen, 2005). People with MDD whose relatives or spouses can be described as critical and emotionally overinvolved have higher relapse rates than do those living with people who are less critical and emotionally overinvolved (Butzlaff & Hooley, 1998).
People’s attributional styles or their general ways of thinking, interpreting, and recalling information have also been examined in the etiology of MDD (Gotlib & Joorman, 2010). People with a pessimistic attributional style tend to make internal (versus external), global (versus specific), and stable (versus unstable) attributions to negative events, serving as a vulnerability to developing MDD. For example, someone who, when he fails an exam, thinks that it was his fault (internal), that he is stupid (global), and that he will always do poorly (stable) has a pessimistic attributional style. Several influential theories of depression incorporate attributional styles (Abramson et al., 1978, 1989).
Bipolar Disorders
Although there have been important advances in research on the etiology, course, and treatment of BD, there remains a need to understand the mechanisms that contribute to episode onset and relapse. There is compelling evidence for biological causes of BD, which is known to be highly heritable (McGuffin et al., 2003). It may be argued that a high rate of heritability demonstrates that BD is fundamentally a biological phenomenon. However, there is much variability in the course of BD both within a person across time and across people (Johnson, 2005). The triggers that determine how and when this genetic vulnerability is expressed are not yet understood; however, there is evidence to suggest that psychosocial triggers may play an important role in BD risk (e.g., Johnson et al., 2008; Malkoff-Schwartz et al., 1998).
In addition to the genetic contribution, biological explanations of BD have also focused on brain function. Many of the studies using fMRI techniques to characterize BD have focused on the processing of emotional stimuli based on the idea that BD is fundamentally a disorder of emotion (APA, 2000). Findings show that regions of the brain thought to be involved in emotional processing and regulation are activated differently in people with BD relative to healthy controls (e.g., Altshuler et al., 2008; Hassel et al., 2008; Lennox et al., 2004).
However, there is little consensus as to whether a particular brain region becomes more or less active in response to an emotional stimulus among people with BD compared with healthy controls. Mixed findings are in part due to samples consisting of participants who are at various phases of illness at the time of testing (manic, depressed, inter-episode). Sample sizes tend to be relatively small, making comparisons between subgroups difficult. Additionally, the use of a standardized stimulus (e.g., facial expression of anger) may not elicit a sufficiently strong response. Personally engaging stimuli, such as recalling a memory, may be more effective in inducing strong emotions (Isaacowitz et al., 2013).
Within the psychosocial level, research has focused on the environmental contributors to BD. A series of studies shows that environmental stressors, particularly severe stressors (e.g., loss of a significant relationship), can adversely impact the course of BD. People with BD have substantially increased risk of relapse (Ellicott et al., 1990) and suffer more depressive symptoms (Johnson et al., 1999) following a severe life stressor. Interestingly, positive life events can also adversely impact the course of BD. People with BD suffer more manic symptoms after life events involving attainment of a desired goal (Johnson et al., 2008). Such findings suggest that people with BD may have a hypersensitivity to rewards.
Evidence from the life stress literature has also suggested that people with mood disorders may have a circadian vulnerability that renders them sensitive to stressors that disrupt their sleep or rhythms. According to social zeitgeber theory (Ehlers et al., 1988; Frank et al., 1994), stressors that disrupt sleep or that disrupt the daily routines that entrain the biological clock (e.g., meal times) can trigger episode relapse. Consistent with this theory, studies have shown that life events that involve a disruption in sleep and daily routines, such as overnight travel, can increase bipolar symptoms in people with BD (Malkoff-Schwartz et al., 1998). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/35%3A_Mood_Disorders/35.04%3A_What_Are_Some_Of_The_Factors_Implicated_In_The_Development_And_Course_Of_Mood_Disorders.txt |
Depressive Disorders
There are many treatment options available for people with MDD. First, a number of antidepressant medications are available, all of which target one or more of the neurotransmitters implicated in depression. The earliest antidepressant medications were monoamine oxidase inhibitors (MAOIs). MAOIs inhibit monoamine oxidase, an enzyme involved in deactivating dopamine, norepinephrine, and serotonin. Although effective in treating depression, MAOIs can have serious side effects. Patients taking MAOIs may develop dangerously high blood pressure if they take certain drugs (e.g., antihistamines) or eat foods containing tyramine, an amino acid commonly found in foods such as aged cheeses, wine, and soy sauce. Tricyclics, the second-oldest class of antide- pressant medications, block the reabsorption of norepinephrine, serotonin, or dopamine at synapses, resulting in their increased availability. Tricyclics are most effective for treating vegetative and somatic symptoms of depression. Like MAOIs, they have serious side effects, the most concerning of which is cardiotoxicity. Selective serotonin reuptake inhibitors (SSRIs; e.g., fluoxetine) and serotonin and norepinephrine reuptake inhibitors (SNRIs; e.g., duloxetine) are the most recently introduced antidepressant medications. SSRIs, the most commonly prescribed antidepressant medication, block the reabsorption of serotonin, whereas SNRIs block the reabsorption of serotonin and norepinephrine. SSRIs and SNRIs have fewer serious side effects than do MAOIs and tricyclics. In particular, they are less cardiotoxic, less lethal in overdose, and produce fewer cognitive impairments. They are not, however, without their own side effects, which include but are not limited to difficulty having orgasms, gastrointestinal issues, and insomnia. It should be noted that antidepressant medication may not work equally for all people. This approach to treatment often involves experimentation with several medications and dosages, and may be more effective when paired with physical exercise and psychotherapy.
Other biological treatments for people with depression include electroconvulsive therapy (ECT), transcranial magnetic stimulation (TMS), and deep brain stimulation. ECT involves inducing a seizure after a patient takes muscle relaxants and is under general anesthesia. ECT is a viable treatment for patients with severe depression or who show resistance to antidepressants although the mechanisms through which it works remain unknown. A common side effect is confusion and memory loss, usually short-term (Schulze-Rauschen- bach et al., 2005). Repetitive TMS is a noninvasive technique administered while a patient is awake. Brief pulsating magnetic fields are delivered to the cortex, inducing electrical activity. TMS has fewer side effects than ECT (Schulze- Rauschenbach et al., 2005), and while outcome studies are mixed, there is evidence that TMS is a promising treatment for patients with MDD who have shown resistance to other treatments (Rosa et al., 2006). Most recently, deep brain stimulation is being examined as a treatment option for patients who do not respond to more traditional treatments like those already described. Deep brain stimulation involves implanting an electrode in the brain. The electrode is connected to an implanted neurostimulator, which electrically stimulates that particular brain region. Although there is some evidence of its effectiveness (Mayberg et al., 2005), additional research is needed.
Several psychosocial treatments have received strong empirical support, meaning that independent investigations have achieved similarly positive results—a high threshold for examining treatment outcomes. These treatments include but are not limited to behavior therapy, cognitive therapy, and interpersonal therapy. Behavior therapies focus on increasing the frequency and quality of experiences that are pleasant or help the patient achieve mastery. Cognitive therapies primarily focus on helping patients identify and change distorted automatic thoughts and assumptions (e.g., Beck, 1967). Cognitive-behavioral therapies are based on the rationale that thoughts, behaviors, and emotions affect and are affected by each other. Interpersonal therapy for depression focuses largely on improving interpersonal relationships by targeting problem areas, specifically unresolved grief, interpersonal role disputes, role transitions, and interpersonal deficits. Finally, there is also some support for the effectiveness of short-term psychodynamic therapy for depression (Leichsenring, 2001). The short-term treatment focuses on a limited number of important issues, and the therapist tends to be more actively involved than in more traditional psychodynamic therapy.
Bipolar Disorders
Patients with BD are typically treated with pharmacotherapy. Antidepressants such as SSRIs and SNRIs are the primary choice of treatment for depression, whereas for BD, lithium is the first-line treatment choice. This is because SSRIs and SNRIs have the potential to induce mania or hypomania in patients with BD. Lithium acts on several neurotransmitter systems in the brain through complex mechanisms, including reduction of excitatory (dopamine and glutamate) neurotransmission, and increasing of inhibitory (GABA) neurotransmission (Lenox & Hahn, 2000). Lithium has strong efficacy for the treatment of BD (Geddes et al., 2004). However, a number of side effects can make lithium treatment difficult for patients to tolerate. Side effects include impaired cognitive function (Wingo et al., 2009), as well as physical symptoms such as nausea, tremor, weight gain, and fatigue (Dunner, 2000). Some of these side effects can improve with continued use; however, medication noncompliance remains an ongoing concern in the treatment of patients with BD. Anticonvulsant medications (e.g., carbamazepine, valproate) are also commonly used to treat patients with BD, either alone or in conjunction with lithium.
There are several adjunctive treatment options for people with BD. Interpersonal and social rhythm therapy (IPSRT) (Frank et al., 1994) is a psychosocial intervention focused on addressing the mechanism of action posited in social zeitgeber theory to predispose patients who have BD to relapse, namely sleep disruption. A growing body of literature provides support for the central role of sleep dysregulation in BD (Harvey, 2008). Consistent with this literature, IPSRT aims to increase rhythmicity of patients’ lives and encourage vigilance in maintaining a stable rhythm. The therapist and patient work to develop and maintain a healthy balance of activity and stimulation such that the patient does not become overly active (e.g., by taking on too many projects) or inactive (e.g., by avoiding social contact). The efficacy of IPSRT has been demonstrated in that patients who received this treatment show reduced risk of episode recurrence and are more likely to remain well (Frank et al., 2005). | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/35%3A_Mood_Disorders/35.05%3A_What_Are_Some_Of_The_Well-Supported_Treatments_For_Mood_Disorders.txt |
Everyone feels down or euphoric from time to time. For some people, these feelings can last for long periods of time and can also co-occur with other symptoms that, in combination, interfere with their everyday lives. When people experience an MDE or a manic episode, they see the world differently. During an MDE, people often feel hopeless about the future and may even experience suicidal thoughts. During a manic episode, people often behave in ways that are risky or place them in danger. They may spend money excessively or have unprotected sex, often expressing deep shame over these decisions after the episode. MDD and BD cause significant problems for people at school, at work, and in their relationships and affect people regardless of gender, age, nationality, race, religion, or sexual orientation. If you or someone you know is suffering from a mood disorder, it is important to seek help. Effective treatments are available and continually improving. If you have an interest in mood disorders, there are many ways to contribute to their understanding, prevention, and treatment, whether by engaging in research or clinical work.
REFERENCES
Abramson, L. Y., Seligman, M. E. P., & Teasdale, J. D. (1978). Learned helplessness in humans: Critique and reformulation. Journal of Abnormal Psychology, 87(1), 49–74. https://doi.org/10.1037/ 0021-843X.87.1.49
Abramson, L. Y., Metalsky, G. I., & Alloy, L. B. (1989). Hopelessness depression: A theory-based subtype of depression. Psychological Review, 96(2), 358–373. https://doi.org/10.1037/0033-295X.96.2.358
Altshuler, L., Bookheimer, S., Townsend, J., Proenza, M. A., Sabb, F., Mintz, J., & Cohen, M. S. (2008). Regional brain changes in bipolar I depression: A functional magnetic resonance imaging study. Bipolar Disorders, 10(6), 708–717. https://doi.org/10.1111/j.1399- 5618.2008.00617.x
American Psychiatric Association. (2000). Diagnostic and statistical manual of mental disorders (4th ed., text rev.).
American Psychiatric Association. (2013). Diagnostic and statistical manual of mental disorders (5th ed.). https://doi.org/10.1176/appi. books.9780890425596
Beck, A. T. (1967). Depression: Clinical, experimental, and theoretical aspects. Hoeber.
Blazer, D., George, L. K., Landerman, R., Pennybacker, M., Melville, M. L., Woodbury, M., Manton, K. G., & Jordan, K. (1985). Psychiatric disorders: A rural/urban comparison. Archives of General Psychiatry, 42(7), 651–656. https://doi.org/10.1001/archpsyc.1985. 01790300013002
Breslau, J., Aguilar-Gaxiola, S., Kendler, K. S., Su, M., Williams, D., & Kessler, R. C. (2006). Specifying race-ethnic differences in risk for psychiatric disorder in a USA national sample. Psychological Medicine, 36(1), 57–68. https://doi.org/10.1017/S0033291705006161
Breslau, J., Kendler, K. S., Su, M., Gaxiola-Aguilar, S., & Kessler, R. C. (2005). Lifetime risk and persistence of psychiatric disorders across ethnic groups in the United States. Psychological Medicine, 35(3), 317–327. https://doi.org/10.1017/S0033291704003514
Brown, G. W., & Harris, T. O. (1989). Life events and illness. Guilford Press.
Butzlaff, R. L., & Hooley, J. M. (1998). Expressed emotion and psychiatric relapse: A meta-analysis. Archives of General Psychiatry, 55(6), 547–552. https://doi.org/10.1001/archpsyc.55.6.547
Chentsova-Dutton, Y. E., & Tsai, J. L. (2009). Understanding depression across cultures. In I. H. Gotlib & C. L. Hammen (Eds.), Handbook of depression (2nd ed., pp. 363–385). Guilford Press.
Conley, C. S., & Rudolph, K. D. (2009). The emerging sex difference in adolescent depression: Interacting contributions of puberty and peer stress. Development and Psychopathology, 21(2), 593–620. https:// doi.org/10.1017/S0954579409000327
Dunner, D. L. (2000). Optimizing lithium treatment. Journal of Clinical Psychiatry, 61(Suppl. 9), 76–81.
Ehlers, C. L., Frank, E., & Kupfer, D. J. (1988). Social zeitgebers and biological rhythms: A unified approach to understanding the etiology of depression. Archives of General Psychiatry, 45(10), 948–952. https://doi.org/10.1001/archpsyc.1988.01800340076012
Ellicott, A., Hammen, C., Gitlin, M., Brown, G., & Jamison, K. (1990). Life events and the course of bipolar disorder. American Journal of Psychiatry, 147(9), 1194–1198. https://doi.org/10.1176/ajp.147.9.1194
Frank, E., Kupfer, D. J., Ehlers, C. L., Monk, T. H., Cornes, C., Carter, S., & Frankel, D. (1994). Interpersonal and social rhythm therapy for bipolar disorder: Integrating interpersonal and behavioral approaches. Behavior Therapy, 17, 143–149.
Frank, E., Kupfer, D. J., Thase, M. E., Mallinger, A. G., Swartz, H. A., Fagiolini, A. M., Grochocinski, V., Houck, P., Scott, J., Thompson, W., & Monk, T. (2005). Two-year outcomes for interpersonal and social rhythm therapy in individuals with bipolar I disorder. Archives of General Psychiatry, 62(9), 996–1004. https://doi.org/10.1001/ archpsyc.62.9.996
Geddes, J. R., Burgess, S., Hawton, K., Jamison, K., & Goodwin, G. M. (2004). Long-term lithium therapy for bipolar disorder: Systematic review and meta-analysis of randomized controlled trials. American Journal of Psychiatry, 161(2), 217–222. https://doi.org/10.1176/appi.ajp.161.2.217
Gonzalez, J. M., Perlick, D. A., Miklowitz, D. J., Kaczynski, R., Hernandez, M., Rosenheck, R. A., Culver, J. L., Ostacher, M. J., & Bowden, C. L. (2007). Factors associated with stigma among caregivers of patients with bipolar disorder in the STEP-BD study. Psychiatric Services, 58(1), 41–48. https://doi.org/10.1176/ps.2007.58.1.41
Goodwin, F. K., & Jamison, K. R. (2007). Manic-depressive illness: Bipolar disorders and recurrent depression. Oxford University Press.
Gotlib, I. H., & Joormann, J. (2010). Cognition and depression: Current status and future directions. Annual Review of Clinical Psychology, 6(1), 285–312. https://doi.org/10.1146/annurev.clinpsy.121208.131305
Hamilton, J. P., Etkin, A., Furman, D. F., Lemus, M. G., Johnson, R. F., & Gotlib, I. H. (2012). Functional neuroimaging of major depressive disorder: A meta-analysis and new integration of baseline activation and neural response data. American Journal of Psychiatry, 169(7), 693–703. https://doi.org/10.1176/appi.ajp.2012.11071105
Hammen, C. (2005). Stress and depression. Annual Review of Clinical Psychology, 1(1), 293–319. https://doi.org/10.1146/annurev. clinpsy.1.102803.143938
Harvey, A. G. (2008). Sleep and circadian rhythms in bipolar disorder: Seeking synchrony, harmony and regulation. American Journal of Psychiatry, 165(7), 820–829. https://doi.org/10.1176/appi. ajp.2008.08010098
Hasin, D. S., Goodwin, R. D., Sintson, F. S., & Grant, B. F. (2005). Epidemiology of major depressive disorder: Results from the National Epidemiological Survey on Alcoholism and Related Conditions. Archives of General Psychiatry, 62(10), 1097–1106. https://doi. org/10.1001/archpsyc.62.10.1097
Hassel, S., Almeida, J. R. C., Kerr, N., Nau, S., Ladouceur, C. D., Fissell, K., Kupfer, D. J., & Phillips, M. L. (2008). Elevated striatal and decreased dorsolateral prefrontal cortical activity in response to emotional stimuli in euthymic bipolar disorder: No associations with psychotropic medication load. Bipolar Disorders, 10(8), 916–927. https://doi.org/10.1111/j.1399-5618.2008.00641.x
Isaacowitz, D. M., Gershon, A., Allard, E. S., & Johnson, S. L. (2013). Emotion in aging and bipolar disorder: Similarities, differences and lessons for further research. Emotion Review, 5(3), 312–320. https:// doi.org/10.1177/1754073912472244
Johnson, S. L. (2005). Mania and dysregulation in goal pursuit: A review. Clinical Psychology Review, 25(2), 241–262. https://doi.org/10.1016/j. cpr.2004.11.002
Johnson, S. L., Cueller, A. K., Ruggero, C., Winett-Perlman, C., Goodnick, P., White, R., & Miller, I. (2008). Life events as predictors of mania and depression in bipolar I disorder. Journal of Abnormal Psychology, 117(2), 268–277. https://doi.org/10.1037/0021-843X.117.2.268
Johnson, S. L., Winett, C. A., Meyer, B., Greenhouse, W. J., & Miller, I. (1999). Social support and the course of bipolar disorder. Journal of Abnormal Psychology, 108(4), 558–566. https://doi.org/10.1037/ 0021-843X.108.4.558
Kessler, R. C., Berglund, P., Demler, O., Jim, R., Merikangas, K. R., & Walters, E. E. (2005). Lifetime prevalence and age-of-onset distributions of DSM-IV disorders in the National Comorbidity Survey Replication. Archives of General Psychiatry, 62(6), 593–602. https:// doi.org/10.1001/archpsyc.62.6.593
Kessler, R. C., Birnbaum, H., Bromet, E., Hwang, I., Sampson, N., & Shahly, V. (2010). Age differences in major depression: Results from the National Comorbidity Surveys Replication (NCS-R). Psychological Medicine, 40(2), 225–237. https://doi.org/10.1017/ S0033291709990213
Kessler, R. C., Merikangas, K. R., & Wang, P. S. (2007). Prevalence, comorbidity, and service utilization for mood disorders in the United States at the beginning of the twenty-first century. Annual Review of Clinical Psychology, 3(1), 137–158. https://doi.org/10.1146/annurev. clinpsy.3.022806.091444
Kilbourne, A. M., Haas, G. L., Mulsant, B. H., Bauer, M. S., & Pincus, H. A. (2004). Concurrent psychiatric diagnoses by age and race among persons with bipolar disorder. Psychiatric Services, 55(8), 931–933. https://doi.org/10.1176/appi.ps.55.8.931
Leichsenring, F. (2001). Comparative effects of short-term psychodynamic psychotherapy and cognitive-behavioral therapy in depression: A meta-analytic approach. Clinical Psychology Review, 21(3), 401–419. https://doi.org/10.1016/S0272-7358(99)00057-4
Lennox, B. R., Jacob, R., Calder, A. J., Lupson, V., & Bullmore, E. T. (2004). Behavioural and neurocognitive responses to sad facial affect are attenuated in patients with mania. Psychological Medicine, 34(5), 795–802. https://doi.org/10.1017/S0033291704002557
Lenox, R. H., & Hahn C.-G. (2000). Overview of the mechanism of action of lithium in the brain: Fifty-year update. Journal of Clinical Psychiatry, 61(Suppl. 9), 5–15.
Leverich, G. S., Altshuler, L. L., Frye, M. A., Suppes, T., Keck, P. E., Jr, McElroy, S. L., Denicoff, K. D., Obrocea, G., Nolen, W. A., Kupka, R., Walden, J., Grunze, H., Perez, S., Luckenbaugh, D. A., & Post, R. M. (2003). Factors associated with suicide attempts in 648 patients with bipolar disorder in the Stanley Foundation Bipolar Network. Journal of Clinical Psychiatry, 64(5), 506–515. https://doi.org/10.4088/JCP. v64n0503
Leverich, G. S., Post, R. M., Keck, P. E., Jr, Altshuler, L. L., Frye, M. A., Kupka, R. W., Nolen, W. A., Suppes, T., McElroy, S. L., Grunze, H., Denicoff, K., Moravec, M. K. M., & Luckenbaugh, D. (2007). The poor prognosis of childhood-onset bipolar disorder. Journal of Pediatrics, 150(5), 485–490. https://doi.org/10.1016/j.jpeds.2006.10.070
Lewinsohn, P. M., Allen, N. B., Seeley, J. R., & Gotlib, I. H. (1999). First onset versus recurrence of depression: Differential processes of psychosocial risk. Journal of Abnormal Psychology, 108(3), 483–489. https://doi.org/10.1037/0021-843X.108.3.483
Lewinsohn, P. M., Seeley, J. R., Buckley, M. E., & Klein, D. N. (2002). Bipolar disorder in adolescence and young adulthood. Child and Adolescent Psychiatric Clinics of North America, 11(3), 461–475. https://doi.org/10.1016/S1056-4993(02)00005-6
Lohoff, F. W. (2010). Overview of genetics of major depressive disorder. Current Psychiatry Reports, 12(6), 539–546. https://doi.org/10.1007/ s11920-010-0150-6
Lorant, V., Deliège, D., Eaton, W., Robert, A., Philippot, P., & Ansseau, A. (2003). Socioeconomic inequalities in depression: A meta-analysis. American Journal of Epidemiology, 157(2), 98–112. https://doi.org/ 10.1093/aje/kwf182
Malkoff-Schwartz, S., Frank, E., Anderson, B. P., Sherrill, J. T., Siegel, L., Patterson, D., & Kupfer, D. J. (1998). Stressful life events and social rhythm disruption in the onset of manic and depressive bipolar episodes: A preliminary investigation. Archives of General Psychiatry, 55(8), 702–707. https://doi.org/10.1001/archpsyc.55.8.702
Mayberg, H. S., Lozano, A. M., Voon, V., McNeely, H. E., Seminowixz, D., Hamani, C., Schwalb, J. M., & Kennedy, S. H. (2005). Deep brain stimulation for treatment-resistant depression. Neuron, 45(5), 651–660. https://doi.org/10.1016/j.neuron.2005.02.014
McElroy, S. L., Altshuler, L. L., Suppes, T., Keck, P. E., Jr, Frye, M. A., Denicoff, K. D., Nolen, W. A., Kupka, R. W., Leverich, G. S., Rochussen, J. R., Rush, A. J., & Post, R. M. (2001). Axis I psychiatric comorbidity and its relationship to historical illness variables in 288 patients with bipolar disorder. American Journal of Psychiatry, 158(3), 420–426. https://doi.org/10.1176/appi.ajp.158.3.420
McGuffin, P., Rijsdijk, F., Andrew, M., Sham, P., Katz, R., Cardno, A. (2003). The heritability of bipolar affective disorder and the genetic relationship to unipolar depression. Archives of General Psychiatry, 60(5), 497–502. https://doi.org/10.1001/archpsyc.60.5.497
Merikangas, K. R., Akiskal, H. S., Angst, J., Greenberg, P. E., Hirschfeld, R. M. A., Petukhova, M., & Kessler, R. C. (2007). Lifetime and 12-month prevalence of bipolar spectrum disorder in the National Comorbidity Survey replication. Archives of General Psychiatry, 64(5), 543–552. https://doi.org/10.1001/archpsyc.64.5.543
Merikangas, K. R., Jin, R., He, J.-P., Kessler, R. C., Lee, S., Sampson, N. A., Viana, M. C., Andrade, L. H., Hu, C., Karam, E. G., Ladea, M., Medina-Mora, M. E., Ono, Y., Posada-Villa, J., Sagar, R., Wells, J. E., & Zarkov, Z. (2011). Prevalence and correlates of bipolar spectrum disorder in the world mental health survey initiative. Archives of General Psychiatry, 68(3), 241–251. https://doi.org/10.1001/ archgenpsychiatry.2011.12
Minsky, S., Vega, W., Miskimen, T., Gara, M., & Escobar, J. (2003). Diagnostic patterns in Latino, African American, and European American psychiatric patients. Archives of General Psychiatry, 60(6), 637–644. https://doi.org/10.1001/archpsyc.60.6.637
Monroe, S. M., & Harkness, K. L. (2011). Recurrence in major depression: A conceptual analysis. Psychological Review, 118(4), 655–674. https:// doi.org/10.1037/a0025190
Monroe, S. M., & Harkness, K. L. (2005). Life stress, the “Kindling” hypothesis, and the recurrence of depression: Considerations from a life stress perspective. Psychological Review, 112(2), 417–445. https:// doi.org/10.1037/0033-295X.112.2.417
Monroe, S. M., Slavich, G. M., & Georgiades, K. (2009). The social environment and life stress in depression. In I. H. Gotlib & C. L. Hammen (Eds.), Handbook of depression (2nd ed., pp. 340–360). Guilford Press.
Nolen-Hoeksema, S., & Hilt, L. M. (2009). Gender differences in depression. In I. H. Gotlib & C. L. Hammen (Eds.), Handbook of depression (2nd ed., pp. 386–404). Guilford Press.
Perlis, R. H., Miyahara, S., Marangell, L. B., Wisniewski, S. R., Ostacher, M., DelBello, M. P., Bowden, C. L., Sachs, G. S., & Nierenberg, A. A. (2004). Long-term implications of early onset in bipolar disorder: data from the first 1000 participants in the systematic treatment enhancement program for bipolar disorder (STEP-BD). Biological Psychiatry, 55(9), 875–881. https://doi.org/10.1016/j. biopsych.2004.01.022
Rosa, M. A., Gattaz, W. F., Pascual-Leone, A., Fregni, F., Rosa, M. O., Rumi, D. O., Myczkowski, M., Silva, M. F., Mansur, C., Rigonatti, S. P., Teixeira, M. J., & Marcolin, M. A. (2006). Comparison of repetitive transcranial magnetic stimulation and electroconvulsive therapy in unipolar non-psychotic refractory depression: A randomized, single-blind study. International Journal of Neuropsychopharma- cology, 9(6), 667–676. https://doi.org/10.1017/S1461145706007127
Schulze-Rauschenbach, S. C., Harms, U., Schlaepfer, T. E., Maier, W., Falkai, P., & Wagner, M. (2005). Distinctive neurocognitive effects of repetitive transcranial magnetic stimulation and electroconvulsive therapy in major depression. British Journal of Psychiatry, 186(5), 410–416. https://doi.org/10.1192/bjp.186.5.410
Shields, B. (2005). Down came the rain: My journey through postpartum depression. Hyperion.
Spatz Widom, C., DuMont, K., & Czaja, S. J. (2007). A prospective investigation of major depressive disorder and comorbidity in abused and neglected children grown up. Archives of General Psychiatry, 64(1), 49–56. https://doi.org/10.1001/archpsyc.64.1.49
Sullivan, P., Neale, M. C., & Kendler, K. S. (2000). Genetic epidemiology of major depression: Review and meta-analysis. American Journal of Psychiatry, 157(10), 1552–1562. https://doi.org/10.1176/appi. ajp.157.10.1552
Twenge, J. M., & Nolen-Hoeksema, S. (2002). Age, gender, race, socioeconomic status, and birth cohort differences on the Children’s Depression Inventory: A meta-analysis. Journal of Abnormal Psychology, 111(4), 578–588. https://doi.org/10.1037/ 0021-843X.111.4.578
Whisman, M. A., & Uebelacker, L. A. (2009). Prospective associations between marital discord and depressive symptoms in middle-aged and older adults. Psychology and Aging, 24(1), 184–189. https://doi. org/10.1037/a0014759
Williams, D. R., Gonzalez, H. M., Neighbors, H., Nesse, R., Abelson, J. M., Sweetman, J., & Jackson, J. S. (2007). Prevalence and distribution of major depressive disorder in African Americans, Caribbean blacks, and non-Hispanic whites: Results from the National Survey of American Life. Archives of General Psychiatry, 64(3), 305–315. https://doi.org/10.1001/archpsyc.64.3.305
Wingo, A. P., Wingo, T. S., Harvey, P. D., & Baldessarini, R. J. (2009). Effects of lithium on cognitive performance: A meta-analysis. Journal of Clinical Psychiatry, 70(11), 1588–1597. https://doi.org/10.4088/ JCP.08r04972 | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/35%3A_Mood_Disorders/35.06%3A_Conclusion.txt |
Schizophrenia and the other psychotic disorders are some of the most impairing forms of psychopathology, frequently associated with a profound negative effect on the individual’s educational, occupational, and social function. Sadly, these disorders often manifest right at the time of transition from adolescence to adulthood, just as young people should be evolving into independent young adults. The spectrum of psychotic disorders includes schizophrenia, schizoaffective disorder, delusional disorder, schizotypal personality disorder, schizophreniform disorder, brief psychotic disorder, and psychosis associated with substance use or medical conditions. In this module, we summarize the primary clinical features of these disorders, describe the known cognitive and neurobiological changes associated with schizophrenia, describe potential risk factors and/or causes for the development of schizophrenia, and describe currently available treatments for schizophrenia.
36.02: The Phenomenology Of Schizophrenia And Related Psychotic Disorders
Most of you have probably had the experience of walking down the street in a city and seeing a person you thought was acting oddly. They may have been dressed in an unusual way, perhaps disheveled or wearing an unusual collection of clothes, makeup, or jewelry that did not seem to fit any particular group or subculture. They may have been talking to themselves or yelling at someone you could not see. If you tried to speak to them, they may have been difficult to follow or understand, or they may have acted paranoid or started telling a bizarre story about the people who were plotting against them. If so, chances are that you have encountered an individual with schizophrenia or another type of psychotic disorder. If you have watched the movie A Beautiful Mind or The Fisher King, you have also seen a portrayal of someone thought to have schizophrenia. Sadly, a few of the individuals who have committed some of the recently highly publicized mass murders may have had schizophrenia, though most people who commit such crimes do not have schizophrenia. It is also likely that you have met people with schizophrenia without ever knowing it, as they may suffer in silence or stay isolated to protect themselves from the horrors they see, hear, or believe are operating in the outside world. As these examples begin to illustrate, psychotic disorders involve many different types of symptoms, including delusions, hallucinations, disorganized speech and behavior, abnormal motor behavior (including catatonia), and negative symptoms such anhedonia or amotivation and blunted affect/reduced speech.
Delusions are false beliefs that are often fixed, hard to change even when the person is presented with conflicting information, and often culturally influenced in their content (e.g., delusions involving Jesus in Judeo-Christian cultures, delusions involving Allah in Muslim cultures). They can be terrifying for the person, who may remain convinced that they are true even when loved ones and friends present them with clear information that they cannot be true. There are many different types or themes to delusions.
The most common delusions are persecutory and involve the belief that individuals or groups are trying to harm or plot against the person in some way. These can be people that the person knows (people at work, the neighbors, family members), or more abstract groups (the FBI, the CIA, aliens, etc.). Other types of delusions include grandiose delusions, where the person believes they have some special power or ability (e.g., I am the new Buddha, I am a rock star); referential delusions, where the person believes that events or objects in the environment have special meaning for them (e.g., that song on the radio is being played specifically for me); or other types of delusions where the person may believe that others are controlling their thoughts and actions, that their thoughts are being broadcast aloud, or that others can read their mind (or they can read other people’s minds).
When you see a person on the street talking to themselves or shouting at other people, they are experiencing hallucinations. These are perceptual experiences that occur even when there is no stimulus in the outside world generating the experiences. They can be auditory, visual, olfactory (smell), gustatory (taste), or somatic (touch). The most common hallucinations in psychosis (at least in adults) are auditory and can involve one or more voices talking about the person, commenting on the person’s behavior, or giving them orders. The content of the hallucinations is frequently negative (“you are a loser,” “that drawing is stupid,” “you should go kill your- self ”) and can be the voice of someone the person knows or a complete stranger. Sometimes the voices sound as if they are coming from outside the person’s head. Other times the voices seem to be coming from inside the person’s head, but are not experienced the same as the person’s inner thoughts or inner speech.
Talking to someone with schizophrenia is sometimes difficult, as their speech may be difficult to follow, either because their answers do not clearly flow from your questions, or because one sentence does not logically follow from another. This is referred to as disorganized speech, and it can be present even when the person is writing. Disorganized behavior can include odd dress, odd makeup (e.g., lipstick outlining a mouth for 1 inch), or unusual rituals (e.g., repetitive hand gestures). Abnormal motor behavior can include catatonia, which refers to a variety of behaviors that seem to reflect a reduction in responsiveness to the external environment. This can include holding unusual postures for long periods of time, failing to respond to verbal or motor prompts from another person, or excessive and seemingly purposeless motor activity.
Some of the most debilitating symptoms of schizophrenia are difficult for others to see. These include what people refer to as “negative symptoms” or the absence of certain things we typically expect most people to have. For example, anhedonia and amotivation reflect a lack of apparent interest in or drive to engage in social or recreational activities. These symptoms can manifest as a great amount of time spent in physical immobility. Importantly, anhedonia and amotivation do not seem to reflect a lack of enjoyment in pleasurable activities or events (Cohen & Minor, 2010; Kring & Moran, 2008; Llerena et al., 2012) but rather a reduced drive or ability to take the steps necessary to obtain the potentially positive outcomes (Barch & Dowd, 2010). Flat affect and reduced speech (alogia) reflect a lack of showing emotions through facial expressions, gestures, and speech intonation, as well as a reduced amount of speech and increased pause frequency and duration.
In many ways, the types of symptoms associated with psychosis are the most difficult for us to understand, as they may seem far outside the range of our normal experiences. Unlike depression or anxiety, many of us may not have had experiences that we think of as on the same continuum as psychosis. However, just like many of the other forms of psychopathology described in this book, the types of psychotic symptoms that characterize disorders like schizophrenia are on a continuum with “normal” mental experiences. For example, work by Jim van Os in the Netherlands has shown that a surprisingly large percentage of the general population (10%+) experience psychotic-like symptoms, though many fewer have multiple experiences and most will not continue to experience these symptoms in the long run (Verdoux & van Os, 2002). Similarly, work in a general population of adolescents and young adults in Kenya has also shown that a relatively high percentage of individuals experience one or more psychotic-like experiences (~19%) at some point in their lives (Mamah et al., 2012; Ndetei et al., 2012), although again most will not go on to develop a full-blown psychotic disorder.
Schizophrenia is the primary disorder that comes to mind when we discuss “psychotic” disorders (see Table \(1\) for diagnostic criteria), though there are a number of other disorders that share one or more features with schizophrenia. In the remainder of this module, we will use the terms “psycho- sis” and “schizophrenia” somewhat interchangeably, given that most of the research has focused on schizophrenia. In addition to schizophrenia (see Table \(1\)), other psychotic disorders include schizophreniform disorder (a briefer version of schizophrenia), schizoaffective disorder (a mixture of psycho- sis and depression/mania symptoms), delusional disorder (the experience of only delusions), and brief psychotic disorder (psychotic symptoms that last only a few days or weeks).
Table \(1\): Types of Psychotic Disorders Simplified from the Diagnostic and Statistical Manual, 5th edition (DSM-5) (American Psychiatric Association [APA], 2013)
Schizophrenia (lifetime prevalence about 0.3% to 0.7%)
• Two or more of the following for at least 1 month: hallucinations, delusions, disorganized speech, grossly disorganized or catatonic behavior, negative symptoms.
• Impairment in one or more areas of function (social, occupational, educational self-care) for a significant period of time since the onset of the illness.
• Continuous signs of the illness for at least 6 months (this can include prodromal or residual symptoms, which are attenuated forms of the symptoms described above).
Schizophreniform Disorder (lifetime prevalence similar to Schizophrenia)
• The same symptoms of schizophrenia described above that are present for at least 1 month but less than 6 months.
Schizoaffective Disorder (lifetime prevalence above 0.3%)
• A period of illness where the person has both the psychotic symptoms necessary to meet criteria for schizophrenia and either a major depression or manic episode.
• The person experiences either delusions or hallucinations for at least 2 weeks when they are not having a depressive or manic episode.
• The symptoms that meet criteria for depressive or manic episodes are present for over half of the illness duration.
Delusional Disorder (lifetime prevalence about 0.2%)
• The presence of at least one delusion for at least a month.
• The person has never met criteria for schizophrenia.
• The person’s function is not impaired outside the specific impact of the delusion.
• The duration of any depressive or manic episodes have been brief relative to the duration of the delusion(s).
Brief Psychotic Disorder (lifetime prevalence unclear)
• One or more of the following symptoms present for at least 1 day but less than 1 month: delusions, hallucinations, disorganized speech, grossly disordered or catatonic behavior.
Attenuated Psychotic Disorder (in Section III of the DSM-5, lifetime prevalence unclear)
• One or more of the following symptoms in an “attenuated” form: delusions, hallucinations, or disorganized speech.
• The symptoms must have occurred at least once a week for the past month and must have started or gotten worse in the past year.
• The symptoms must be severe enough to distress or disable the individual or to suggest to others that the person needs clinical help.
• The person has never met the diagnostic criteria for a psychotic disorder, and the symptoms are not better attributed to another disorder, to substance use, or to a medical condition. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/36%3A_Schizophrenia_Spectrum_Disorders/36.01%3A_Introductoin.txt |
As described above, when we think of the core symptoms of psychotic disorders such as schizophrenia, we think of people who hear voices, see visions, and have false beliefs about reality (i.e., delusions). However, problems in cognitive function are also a critical aspect of psychotic disorders and of schizophrenia in particular. This emphasis on cognition in schizophrenia is in part due to the growing body of research suggesting that cognitive problems in schizophrenia are a major source of disability and loss of functional capacity (Green, 2006; Nuechterlein et al., 2011). The cognitive deficits that are present in schizophrenia are widespread and can include problems with episodic memory (the ability to learn and retrieve new information or episodes in one’s life), working memory (the ability to maintain information over a short period of time, such as 30 seconds), and other tasks that require one to “control” or regulate one’s behavior (Barch & Ceaser, 2012; Bora et al., 2009a; Fioravanti et al., 2005; Forbes et al., 2009; Mesholam-Gately et al., 2009). Individuals with schizophrenia also have difficulty with what is referred to as “processing speed” and are frequently slower than healthy individuals on almost all tasks. Importantly, these cognitive deficits are present prior to the onset of the illness (Fusar-Poli et al., 2007) and are also present, albeit in a milder form, in the first-degree relatives of people with schizophrenia (Snitz et al., 2006). This suggests that cognitive impairments in schizophrenia reflect part of the risk for the development of psychosis, rather than being an outcome of developing psychosis. Furthermore, people with schizophrenia who have more severe cognitive problems also tend to have more severe negative symptoms and more disorganized speech and behavior (Barch et al., 1999, 2003; de Gracia Dominguez et al., 2009; Ventura et al., 2009, 2010). In addition, people with more cognitive problems have worse function in everyday life (Bowie et al., 2006, 2008; Fett et al., 2011).
Some people with schizophrenia also show deficits in what is referred to as social cognition, though it is not clear whether such problems are separate from the cognitive problems described above or the result of them (Hoe et al., 2012; Kerr & Neale, 1993; van Hooren et al., 2008). This includes problems with the recognition of emotional expressions on the faces of other individuals (Kohler et al., 2010) and problems inferring the intentions of other people (theory of mind) (Bora et al., 2009b).
Individuals with schizophrenia who have more problems with social cognition also tend to have more negative and disorganized symptoms (Ventura et al., 2011), as well as worse community function (Fett et al., 2011).
The advent of neuroimaging techniques such as structural and functional magnetic resonance imaging and positron emission tomography opened up the ability to try to understand the brain mechanisms of the symptoms of schizophrenia as well as the cognitive impairments found in psychosis. For example, a number of studies have suggested that delusions in psychosis may be associated with problems in “salience” detection mechanisms supported by the ventral striatum (Jensen & Kapur, 2009; Jensen et al., 2008; Kapur, 2003, Kapur et al., 2005; Murray et al., 2008) and the anterior prefrontal cortex (Corlett, Honey, et al., 2006; Corlett, Honey, & Fletcher, 2007; Corlett, Murray, et al., 2007). These are regions of the brain that normally increase their activity when something important (aka “salient”) happens in the environment. If these brain regions misfire, it may lead individuals with psychosis to mistakenly attribute importance to irrelevant or unconnected events. Furthermore, there is good evidence that problems in working memory and cognitive control in schizophrenia are related to problems in the function of a region of the brain called the dorsolateral prefrontal cortex (DLPFC) (Minzenberg et al., 2009; Ragland et al., 2009). These problems include changes in how the DLPFC works when people are doing working-memory or cognitive-control tasks, and how this brain region is connected to other brain regions important for working memory and cognitive control, including the posterior parietal cortex (e.g., Karlsgodt et al., 2008; Kim et al., 2003; Schlösser et al., 2003), the anterior cingulate (Repovš & Barch, 2012), and temporal cortex (e.g., Fletcher et al., 1995; Meyer-Lindenberg et al., 2001). In terms of understanding episodic memory problems in schizophrenia, many researchers have focused on medial temporal lobe deficits, with a specific focus on the hippocampus (e.g., Heckers & Konradi, 2010). This is because there is much data from humans and animals showing that the hippocampus is important for the creation of new memories (Squire, 1992). However, it has become increasingly clear that problems with the DLPFC also make important contributions to episodic memory deficits in schizophrenia (Ragland et al., 2009), probably because this part of the brain is important for controlling our use of memory.
In addition to problems with regions such as the DLFPC and medial temporal lobes in schizophrenia described above, magnitude resonance neuroimaging studies have also identified changes in cellular architecture, white matter connectivity, and gray matter volume in a variety of regions that include the prefrontal and temporal cortices (Bora et al., 2011). People with schizophrenia also show reduced overall brain volume, and reductions in brain volume as people get older may be larger in those with schizophrenia than in healthy people (Olabi et al., 2011). Taking antipsychotic medications or taking drugs such as marijuana, alcohol, and tobacco may cause some of these structural changes. However, these structural changes are not completely explained by medications or sub- stance use alone. Furthermore, both functional and structural brain changes are seen, again to a milder degree, in the first-degree relatives of people with schizophrenia (Boos et al., 2007; Brans et al., 2008; Fusar-Poli et al., 2007; MacDonald et al., 2009). This again suggests that neural changes associated with schizophrenia are related to a genetic risk for this illness. | textbooks/socialsci/Psychology/Introductory_Psychology/General_Psychology_for_Honors_Students_(Votaw)/36%3A_Schizophrenia_Spectrum_Disorders/36.03%3A_The_Cognitive_Neuroscience_Of_Schizophrenia.txt |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.