Stanley Milgram Shock Experiment

Saul McLeod, PhD

Editor-in-Chief for Simply Psychology

BSc (Hons) Psychology, MRes, PhD, University of Manchester

Saul McLeod, PhD., is a qualified psychology teacher with over 18 years of experience in further and higher education. He has been published in peer-reviewed journals, including the Journal of Clinical Psychology.

Learn about our Editorial Process

Olivia Guy-Evans, MSc

Associate Editor for Simply Psychology

BSc (Hons) Psychology, MSc Psychology of Education

Olivia Guy-Evans is a writer and associate editor for Simply Psychology. She has previously worked in healthcare and educational sectors.

On This Page:

Stanley Milgram, a psychologist at Yale University, carried out one of the most famous studies of obedience in psychology.

He conducted an experiment focusing on the conflict between obedience to authority and personal conscience.

Milgram (1963) examined justifications for acts of genocide offered by those accused at the World War II, Nuremberg War Criminal trials. Their defense often was based on obedience  – that they were just following orders from their superiors.

The experiments began in July 1961, a year after the trial of Adolf Eichmann in Jerusalem. Milgram devised the experiment to answer the question:

Could it be that Eichmann and his million accomplices in the Holocaust were just following orders? Could we call them all accomplices?” (Milgram, 1974).

Milgram (1963) wanted to investigate whether Germans were particularly obedient to authority figures, as this was a common explanation for the Nazi killings in World War II.

Milgram selected participants for his experiment by newspaper advertising for male participants to take part in a study of learning at Yale University.

The procedure was that the participant was paired with another person and they drew lots to find out who would be the ‘learner’ and who would be the ‘teacher.’  The draw was fixed so that the participant was always the teacher, and the learner was one of Milgram’s confederates (pretending to be a real participant).

stanley milgram generator scale

The learner (a confederate called Mr. Wallace) was taken into a room and had electrodes attached to his arms, and the teacher and researcher went into a room next door that contained an electric shock generator and a row of switches marked from 15 volts (Slight Shock) to 375 volts (Danger: Severe Shock) to 450 volts (XXX).

The shocks in Stanley Milgram’s obedience experiments were not real. The “learners” were actors who were part of the experiment and did not actually receive any shocks.

However, the “teachers” (the real participants of the study) believed the shocks were real, which was crucial for the experiment to measure obedience to authority figures even when it involved causing harm to others.

Milgram’s Experiment (1963)

Milgram (1963) was interested in researching how far people would go in obeying an instruction if it involved harming another person.

Stanley Milgram was interested in how easily ordinary people could be influenced into committing atrocities, for example, Germans in WWII.

Volunteers were recruited for a controlled experiment investigating “learning” (re: ethics: deception). 

Participants were 40 males, aged between 20 and 50, whose jobs ranged from unskilled to professional, from the New Haven area. They were paid $4.50 for just turning up.

Milgram

At the beginning of the experiment, they were introduced to another participant, a confederate of the experimenter (Milgram).

They drew straws to determine their roles – learner or teacher – although this was fixed, and the confederate was always the learner. There was also an “experimenter” dressed in a gray lab coat, played by an actor (not Milgram).

Two rooms in the Yale Interaction Laboratory were used – one for the learner (with an electric chair) and another for the teacher and experimenter with an electric shock generator.

Milgram Obedience: Mr Wallace

The “learner” (Mr. Wallace) was strapped to a chair with electrodes.

After he has learned a list of word pairs given to him to learn, the “teacher” tests him by naming a word and asking the learner to recall its partner/pair from a list of four possible choices.

The teacher is told to administer an electric shock every time the learner makes a mistake, increasing the level of shock each time. There were 30 switches on the shock generator marked from 15 volts (slight shock) to 450 (danger – severe shock).

Milgram Obedience IV Variations

The learner gave mainly wrong answers (on purpose), and for each of these, the teacher gave him an electric shock. When the teacher refused to administer a shock, the experimenter was to give a series of orders/prods to ensure they continued.

There were four prods, and if one was not obeyed, then the experimenter (Mr. Williams) read out the next prod, and so on.

Prod 1 : Please continue. Prod 2: The experiment requires you to continue. Prod 3 : It is absolutely essential that you continue. Prod 4 : You have no other choice but to continue.

These prods were to be used in order, and begun afresh for each new attempt at defiance (Milgram, 1974, p. 21). The experimenter also had two special prods available. These could be used as required by the situation:

  • Although the shocks may be painful, there is no permanent tissue damage, so please go on’ (ibid.)
  • ‘Whether the learner likes it or not, you must go on until he has learned all the word pairs correctly. So please go on’ (ibid., p. 22).

65% (two-thirds) of participants (i.e., teachers) continued to the highest level of 450 volts. All the participants continued to 300 volts.

Milgram did more than one experiment – he carried out 18 variations of his study.  All he did was alter the situation (IV) to see how this affected obedience (DV).

Conclusion 

The individual explanation for the behavior of the participants would be that it was something about them as people that caused them to obey, but a more realistic explanation is that the situation they were in influenced them and caused them to behave in the way that they did.

Some aspects of the situation that may have influenced their behavior include the formality of the location, the behavior of the experimenter, and the fact that it was an experiment for which they had volunteered and been paid.

Ordinary people are likely to follow orders given by an authority figure, even to the extent of killing an innocent human being.  Obedience to authority is ingrained in us all from the way we are brought up.

People tend to obey orders from other people if they recognize their authority as morally right and/or legally based. This response to legitimate authority is learned in a variety of situations, for example in the family, school, and workplace.

Milgram summed up in the article “The Perils of Obedience” (Milgram 1974), writing:

“The legal and philosophic aspects of obedience are of enormous import, but they say very little about how most people behave in concrete situations. I set up a simple experiment at Yale University to test how much pain an ordinary citizen would inflict on another person simply because he was ordered to by an experimental scientist. Stark authority was pitted against the subjects’ [participants’] strongest moral imperatives against hurting others, and, with the subjects’ [participants’] ears ringing with the screams of the victims, authority won more often than not. The extreme willingness of adults to go to almost any lengths on the command of an authority constitutes the chief finding of the study and the fact most urgently demanding explanation.”

Milgram’s Agency Theory

Milgram (1974) explained the behavior of his participants by suggesting that people have two states of behavior when they are in a social situation:

  • The autonomous state – people direct their own actions, and they take responsibility for the results of those actions.
  • The agentic state – people allow others to direct their actions and then pass off the responsibility for the consequences to the person giving the orders. In other words, they act as agents for another person’s will.

Milgram suggested that two things must be in place for a person to enter the agentic state:

  • The person giving the orders is perceived as being qualified to direct other people’s behavior. That is, they are seen as legitimate.
  • The person being ordered about is able to believe that the authority will accept responsibility for what happens.
According to Milgram, when in this agentic state, the participant in the obedience studies “defines himself in a social situation in a manner that renders him open to regulation by a person of higher status. In this condition the individual no longer views himself as responsible for his own actions but defines himself as an instrument for carrying out the wishes of others” (Milgram, 1974, p. 134).

Agency theory says that people will obey an authority when they believe that the authority will take responsibility for the consequences of their actions. This is supported by some aspects of Milgram’s evidence.

For example, when participants were reminded that they had responsibility for their own actions, almost none of them were prepared to obey.

In contrast, many participants who were refusing to go on did so if the experimenter said that he would take responsibility.

According to Milgram (1974, p. 188):

“The behavior revealed in the experiments reported here is normal human behavior but revealed under conditions that show with particular clarity the danger to human survival inherent in our make-up.

And what is it we have seen? Not aggression, for there is no anger, vindictiveness, or hatred in those who shocked the victim….

Something far more dangerous is revealed: the capacity for man to abandon his humanity, indeed, the inevitability that he does so, as he merges his unique personality into larger institutional structures.”

Milgram Experiment Variations

The Milgram experiment was carried out many times whereby Milgram (1965) varied the basic procedure (changed the IV).  By doing this Milgram could identify which factors affected obedience (the DV).

Obedience was measured by how many participants shocked to the maximum 450 volts (65% in the original study). Stanley Milgram conducted a total of 23 variations (also called conditions or experiments) of his original obedience study:

In total, 636 participants were tested in 18 variation studies conducted between 1961 and 1962 at Yale University.

In the original baseline study – the experimenter wore a gray lab coat to symbolize his authority (a kind of uniform).

The lab coat worn by the experimenter in the original study served as a crucial symbol of scientific authority that increased obedience. The lab coat conveyed expertise and legitimacy, making participants see the experimenter as more credible and trustworthy.

Milgram carried out a variation in which the experimenter was called away because of a phone call right at the start of the procedure.

The role of the experimenter was then taken over by an ‘ordinary member of the public’ ( a confederate) in everyday clothes rather than a lab coat. The obedience level dropped to 20%.

Change of Location:  The Mountain View Facility Study (1963, unpublished)

Milgram conducted this variation in a set of offices in a rundown building, claiming it was associated with “Research Associates of Bridgeport” rather than Yale.

The lab’s ordinary appearance was designed to test if Yale’s prestige encouraged obedience. Participants were led to believe that a private research firm experimented.

In this non-university setting, obedience rates dropped to 47.5% compared to 65% in the original Yale experiments. This suggests that the status of location affects obedience.

Private research firms are viewed as less prestigious than certain universities, which affects behavior. It is easier under these conditions to abandon the belief in the experimenter’s essential decency.

The impressive university setting reinforced the experimenter’s authority and conveyed an implicit approval of the research.

Milgram filmed this variation for his documentary Obedience , but did not publish the results in his academic papers. The study only came to wider light when archival materials, including his notes, films, and data, were studied by later researchers like Perry (2013) in the decades after Milgram’s death.

Two Teacher Condition

When participants could instruct an assistant (confederate) to press the switches, 92.5% shocked to the maximum of 450 volts.

Allowing the participant to instruct an assistant to press the shock switches diffused personal responsibility and likely reduced perceptions of causing direct harm.

By attributing the actions to the assistant rather than themselves, participants could more easily justify shocking to the maximum 450 volts, reflected in the 92.5% obedience rate.

When there is less personal responsibility, obedience increases. This relates to Milgram’s Agency Theory.

Touch Proximity Condition

The teacher had to force the learner’s hand down onto a shock plate when the learner refused to participate after 150 volts. Obedience fell to 30%.

Forcing the learner’s hand onto the shock plate after 150 volts physically connected the teacher to the consequences of their actions. This direct tactile feedback increased the teacher’s personal responsibility.

No longer shielded from the learner’s reactions, the proximity enabled participants to more clearly perceive the harm they were causing, reducing obedience to 30%. Physical distance and indirect actions in the original setup made it easier to rationalize obeying the experimenter.

The participant is no longer buffered/protected from seeing the consequences of their actions.

Social Support Condition

When the two confederates set an example of defiance by refusing to continue the shocks, especially early on at 150 volts, it permitted the real participant also to resist authority.

Two other participants (confederates) were also teachers but refused to obey. Confederate 1 stopped at 150 volts, and Confederate 2 stopped at 210 volts.

Their disobedience provided social proof that it was acceptable to disobey. This modeling of defiance lowered obedience to only 10% compared to 65% without such social support. It demonstrated that social modeling can validate challenging authority.

The presence of others who are seen to disobey the authority figure reduces the level of obedience to 10%.

Absent Experimenter Condition 

It is easier to resist the orders from an authority figure if they are not close by. When the experimenter instructed and prompted the teacher by telephone from another room, obedience fell to 20.5%.

Many participants cheated and missed out on shocks or gave less voltage than ordered by the experimenter. The proximity of authority figures affects obedience.

The physical absence of the authority figure enabled participants to act more freely on their own moral inclinations rather than the experimenter’s commands. This highlighted the role of an authority’s direct presence in influencing behavior.

A key reason the obedience studies fascinate people is Milgram presented them as a scientific experiment, contrasting himself as an “empirically grounded scientist” compared to philosophers. He claimed he systematically varied factors to alter obedience rates.

However, recent scholarship using archival records shows Milgram’s account of standardizing the procedure was misleading. For example, he published a list of standardized prods the experimenter used when participants questioned continuing. Milgram said these were delivered uniformly in a firm but polite tone.

Analyzing audiotapes, Gibson (2013) found considerable variation from the published protocol – the prods differed across trials. The point is not that Milgram did poor science, but that the archival materials reveal the limitations of the textbook account of his “standardized” procedure.

The qualitative data like participant feedback, Milgram’s notes, and researchers’ actions provide a fuller, messier picture than the obedience studies’ “official” story. For psychology students, this shows how scientific reporting can polish findings in a way that strays from the less tidy reality.

Critical Evaluation

Inaccurate description of the prod methodology:.

A key reason the obedience studies fascinate people is Milgram (1974) presented them as a scientific experiment, contrasting himself as an “empirically grounded scientist” compared to philosophers. He claimed he systematically varied factors to alter obedience rates.

However, recent scholarship using archival records shows Milgram’s account of standardizing the procedure was misleading. For example, he published a list of standardized prods the experimenter used when participants questioned continuing. Milgram said these were delivered uniformly in a firm but polite tone (Gibson, 2013; Perry, 2013; Russell, 2010).

Perry’s (2013) archival research revealed another discrepancy between Milgram’s published account and the actual events. Milgram claimed standardized prods were used when participants resisted, but Perry’s audiotape analysis showed the experimenter often improvised more coercive prods beyond the supposed script.

This off-script prodding varied between experiments and participants, and was especially prevalent with female participants where no gender obedience difference was found – suggesting the improvisation influenced results. Gibson (2013) and Russell (2009) corroborated the experimenter’s departures from the supposed fixed prods. 

Prods were often combined or modified rather than used verbatim as published.

Russell speculated the improvisation aimed to achieve outcomes the experimenter believed Milgram wanted. Milgram seemed to tacitly approve of the deviations by not correcting them when observing.

This raises significant issues around experimenter bias influencing results, lack of standardization compromising validity, and ethical problems with Milgram misrepresenting procedures.

Milgram’s experiment lacked external validity:

The Milgram studies were conducted in laboratory-type conditions, and we must ask if this tells us much about real-life situations.

We obey in a variety of real-life situations that are far more subtle than instructions to give people electric shocks, and it would be interesting to see what factors operate in everyday obedience. The sort of situation Milgram investigated would be more suited to a military context.

Orne and Holland (1968) accused Milgram’s study of lacking ‘experimental realism,”’ i.e.,” participants might not have believed the experimental set-up they found themselves in and knew the learner wasn’t receiving electric shocks.

“It’s more truthful to say that only half of the people who undertook the experiment fully believed it was real, and of those two-thirds disobeyed the experimenter,” observes Perry (p. 139).

Milgram’s sample was biased:

  • The participants in Milgram’s study were all male. Do the findings transfer to females?
  • Milgram’s study cannot be seen as representative of the American population as his sample was self-selected. This is because they became participants only by electing to respond to a newspaper advertisement (selecting themselves).
  • They may also have a typical “volunteer personality” – not all the newspaper readers responded so perhaps it takes this personality type to do so.

Yet a total of 636 participants were tested in 18 separate experiments across the New Haven area, which was seen as being reasonably representative of a typical American town.

Milgram’s findings have been replicated in a variety of cultures and most lead to the same conclusions as Milgram’s original study and in some cases see higher obedience rates.

However, Smith and Bond (1998) point out that with the exception of Jordan (Shanab & Yahya, 1978), the majority of these studies have been conducted in industrialized Western cultures, and we should be cautious before we conclude that a universal trait of social behavior has been identified.

Selective reporting of experimental findings:

Perry (2013) found Milgram omitted findings from some obedience experiments he conducted, reporting only results supporting his conclusions. A key omission was the Relationship condition (conducted in 1962 but unpublished), where participant pairs were relatives or close acquaintances.

When the learner protested being shocked, most teachers disobeyed, contradicting Milgram’s emphasis on obedience to authority.

Perry argued Milgram likely did not publish this 85% disobedience rate because it undermined his narrative and would be difficult to defend ethically since the teacher and learner knew each other closely.

Milgram’s selective reporting biased interpretations of his findings. His failure to publish all his experiments raises issues around researchers’ ethical obligation to completely and responsibly report their results, not just those fitting their expectations.

Unreported analysis of participants’ skepticism and its impact on their behavior:

Perry (2013) found archival evidence that many participants expressed doubt about the experiment’s setup, impacting their behavior. This supports Orne and Holland’s (1968) criticism that Milgram overlooked participants’ perceptions.

Incongruities like apparent danger, but an unconcerned experimenter likely cued participants that no real harm would occur. Trust in Yale’s ethics reinforced this. Yet Milgram did not publish his assistant’s analysis showing participant skepticism correlated with disobedience rates and varied by condition.

Obedient participants were more skeptical that the learner was harmed. This selective reporting biased interpretations. Additional unreported findings further challenge Milgram’s conclusions.

This highlights issues around thoroughly and responsibly reporting all results, not just those fitting expectations. It shows how archival evidence makes Milgram’s study a contentious classic with questionable methods and conclusions.

Ethical Issues

What are the potential ethical concerns associated with Milgram’s research on obedience?

While not a “contribution to psychology” in the traditional sense, Milgram’s obedience experiments sparked significant debate about the ethics of psychological research.

Baumrind (1964) criticized the ethics of Milgram’s research as participants were prevented from giving their informed consent to take part in the study. 

Participants assumed the experiment was benign and expected to be treated with dignity.

As a result of studies like Milgram’s, the APA and BPS now require researchers to give participants more information before they agree to take part in a study.

The participants actually believed they were shocking a real person and were unaware the learner was a confederate of Milgram’s.

However, Milgram argued that “illusion is used when necessary in order to set the stage for the revelation of certain difficult-to-get-at-truths.”

Milgram also interviewed participants afterward to find out the effect of the deception. Apparently, 83.7% said that they were “glad to be in the experiment,” and 1.3% said that they wished they had not been involved.

Protection of participants 

Participants were exposed to extremely stressful situations that may have the potential to cause psychological harm. Many of the participants were visibly distressed (Baumrind, 1964).

Signs of tension included trembling, sweating, stuttering, laughing nervously, biting lips and digging fingernails into palms of hands. Three participants had uncontrollable seizures, and many pleaded to be allowed to stop the experiment.

Milgram described a businessman reduced to a “twitching stuttering wreck” (1963, p. 377),

In his defense, Milgram argued that these effects were only short-term. Once the participants were debriefed (and could see the confederate was OK), their stress levels decreased.

“At no point,” Milgram (1964) stated, “were subjects exposed to danger and at no point did they run the risk of injurious effects resulting from participation” (p. 849).

To defend himself against criticisms about the ethics of his obedience research, Milgram cited follow-up survey data showing that 84% of participants said they were glad they had taken part in the study.

Milgram used this to claim that the study caused no serious or lasting harm, since most participants retrospectively did not regret their involvement.

Yet archival accounts show many participants endured lasting distress, even trauma, refuting Milgram’s insistence the study caused only fleeting “excitement.” By not debriefing all, Milgram misled participants about the true risks involved (Perry, 2013).

However, Milgram did debrief the participants fully after the experiment and also followed up after a period of time to ensure that they came to no harm.

Milgram debriefed all his participants straight after the experiment and disclosed the true nature of the experiment.

Participants were assured that their behavior was common, and Milgram also followed the sample up a year later and found no signs of any long-term psychological harm.

The majority of the participants (83.7%) said that they were pleased that they had participated, and 74% had learned something of personal importance.

Perry’s (2013) archival research found Milgram misrepresented debriefing – around 600 participants were not properly debriefed soon after the study, contrary to his claims. Many only learned no real shocks occurred when reading a mailed study report months later, which some may have not received.

Milgram likely misreported debriefing details to protect his credibility and enable future obedience research. This raises issues around properly informing and debriefing participants that connect to APA ethics codes developed partly in response to Milgram’s study.

Right to Withdrawal 

The BPS states that researchers should make it plain to participants that they are free to withdraw at any time (regardless of payment).

When expressing doubts, the experimenter assured them all was well. Trusting Yale scientists, many took the experimenter at his word that “no permanent tissue damage” would occur, and continued administering shocks despite reservations.

Did Milgram give participants an opportunity to withdraw? The experimenter gave four verbal prods which mostly discouraged withdrawal from the experiment:

  • Please continue.
  • The experiment requires that you continue.
  • It is absolutely essential that you continue.
  • You have no other choice, you must go on.

Milgram argued that they were justified as the study was about obedience, so orders were necessary.

Milgram pointed out that although the right to withdraw was made partially difficult, it was possible as 35% of participants had chosen to withdraw.

Replications

Direct replications have not been possible due to current ethical standards . However, several researchers have conducted partial replications and variations that aim to reproduce some aspects of Milgram’s methods ethically.

One important replication was conducted by Jerry Burger in 2009. Burger’s partial replication included several safeguards to protect participant welfare, such as screening out high-risk individuals, repeatedly reminding participants they could withdraw, and stopping at the 150-volt shock level. This was the point where Milgram’s participants first heard the learner’s protests.

As 79% of Milgram’s participants who went past 150 volts continued to the maximum 450 volts, Burger (2009) argued that 150 volts provided a reasonable estimate for obedience levels. He found 70% of participants continued to 150 volts, compared to 82.5% in Milgram’s comparable condition.

Another replication by Thomas Blass (1999) examined whether obedience rates had declined over time due to greater public awareness of the experiments. Blass correlated obedience rates from replication studies between 1963 and 1985 and found no relationship between year and obedience level. He concluded that obedience rates have not systematically changed, providing evidence against the idea of “enlightenment effects”.

Some variations have explored the role of gender. Milgram found equal rates of obedience for male and female participants. Reviews have found most replications also show no gender difference, with a couple of exceptions (Blass, 1999). For example, Kilham and Mann (1974) found lower obedience in female participants.

Partial replications have also examined situational factors. Having another person model defiance reduced obedience compared to a solo participant in one study, but did not eliminate it (Burger, 2009). The authority figure’s perceived expertise seems to be an influential factor (Blass, 1999). Replications have supported Milgram’s observation that stepwise increases in demands promote obedience.

Personality factors have been studied as well. Traits like high empathy and desire for control correlate with some minor early hesitation, but do not greatly impact eventual obedience levels (Burger, 2009). Authoritarian tendencies may contribute to obedience (Elms, 2009).

In sum, the partial replications confirm Milgram’s degree of obedience. Though ethical constraints prevent full reproductions, the key elements of his procedure seem to consistently elicit high levels of compliance across studies, samples, and eras. The replications continue to highlight the power of situational pressures to yield obedience.

Milgram (1963) Audio Clips

Below you can also hear some of the audio clips taken from the video that was made of the experiment. Just click on the clips below.

Why was the Milgram experiment so controversial?

The Milgram experiment was controversial because it revealed people’s willingness to obey authority figures even when causing harm to others, raising ethical concerns about the psychological distress inflicted upon participants and the deception involved in the study.

Would Milgram’s experiment be allowed today?

Milgram’s experiment would likely not be allowed today in its original form, as it violates modern ethical guidelines for research involving human participants, particularly regarding informed consent, deception, and protection from psychological harm.

Did anyone refuse the Milgram experiment?

Yes, in the Milgram experiment, some participants refused to continue administering shocks, demonstrating individual variation in obedience to authority figures. In the original Milgram experiment, approximately 35% of participants refused to administer the highest shock level of 450 volts, while 65% obeyed and delivered the 450-volt shock.

How can Milgram’s study be applied to real life?

Milgram’s study can be applied to real life by demonstrating the potential for ordinary individuals to obey authority figures even when it involves causing harm, emphasizing the importance of questioning authority, ethical decision-making, and fostering critical thinking in societal contexts.

Were all participants in Milgram’s experiments male?

Yes, in the original Milgram experiment conducted in 1961, all participants were male, limiting the generalizability of the findings to women and diverse populations.

Why was the Milgram experiment unethical?

The Milgram experiment was considered unethical because participants were deceived about the true nature of the study and subjected to severe emotional distress. They believed they were causing harm to another person under the instruction of authority.

Additionally, participants were not given the right to withdraw freely and were subjected to intense pressure to continue. The psychological harm and lack of informed consent violates modern ethical guidelines for research.

Baumrind, D. (1964). Some thoughts on ethics of research: After reading Milgram’s” Behavioral study of obedience.”.  American Psychologist ,  19 (6), 421.

Blass, T. (1999). The Milgram paradigm after 35 years: Some things we now know about obedience to authority 1.  Journal of Applied Social Psychology ,  29 (5), 955-978.

Brannigan, A., Nicholson, I., & Cherry, F. (2015). Introduction to the special issue: Unplugging the Milgram machine.  Theory & Psychology ,  25 (5), 551-563.

Burger, J. M. (2009). Replicating Milgram: Would people still obey today? American Psychologist, 64 , 1–11.

Elms, A. C. (2009). Obedience lite. American Psychologist, 64 (1), 32–36.

Gibson, S. (2013). Milgram’s obedience experiments: A rhetorical analysis. British Journal of Social Psychology, 52, 290–309.

Gibson, S. (2017). Developing psychology’s archival sensibilities: Revisiting Milgram’s obedience’ experiments.  Qualitative Psychology ,  4 (1), 73.

Griggs, R. A., Blyler, J., & Jackson, S. L. (2020). Using research ethics as a springboard for teaching Milgram’s obedience study as a contentious classic.  Scholarship of Teaching and Learning in Psychology ,  6 (4), 350.

Haslam, S. A., & Reicher, S. D. (2018). A truth that does not always speak its name: How Hollander and Turowetz’s findings confirm and extend the engaged followership analysis of harm-doing in the Milgram paradigm. British Journal of Social Psychology, 57, 292–300.

Haslam, S. A., Reicher, S. D., & Birney, M. E. (2016). Questioning authority: New perspectives on Milgram’s ‘obedience’ research and its implications for intergroup relations. Current Opinion in Psychology, 11 , 6–9.

Haslam, S. A., Reicher, S. D., Birney, M. E., Millard, K., & McDonald, R. (2015). ‘Happy to have been of service’: The Yale archive as a window into the engaged followership of participants in Milgram’s ‘obedience’ experiment. British Journal of Social Psychology, 54 , 55–83.

Kaplan, D. E. (1996). The Stanley Milgram papers: A case study on appraisal of and access to confidential data files. American Archivist, 59 , 288–297.

Kaposi, D. (2022). The second wave of critical engagement with Stanley Milgram’s ‘obedience to authority’experiments: What did we learn?.  Social and Personality Psychology Compass ,  16 (6), e12667.

Kilham, W., & Mann, L. (1974). Level of destructive obedience as a function of transmitter and executant roles in the Milgram obedience paradigm. Journal of Personality and Social Psychology, 29 (5), 696–702.

Milgram, S. (1963). Behavioral study of obedience . Journal of Abnormal and Social Psychology , 67, 371-378.

Milgram, S. (1964). Issues in the study of obedience: A reply to Baumrind. American Psychologist, 19 , 848–852.

Milgram, S. (1965). Some conditions of obedience and disobedience to authority . Human Relations, 18(1) , 57-76.

Milgram, S. (1974). Obedience to authority: An experimental view . Harpercollins.

Miller, A. G. (2009). Reflections on” Replicating Milgram”(Burger, 2009), American Psychologis t, 64 (1):20-27

Nicholson, I. (2011). “Torture at Yale”: Experimental subjects, laboratory torment and the “rehabilitation” of Milgram’s “obedience to authority”. Theory & Psychology, 21 , 737–761.

Nicholson, I. (2015). The normalization of torment: Producing and managing anguish in Milgram’s “obedience” laboratory. Theory & Psychology, 25 , 639–656.

Orne, M. T., & Holland, C. H. (1968). On the ecological validity of laboratory deceptions. International Journal of Psychiatry, 6 (4), 282-293.

Orne, M. T., & Holland, C. C. (1968). Some conditions of obedience and disobedience to authority. On the ecological validity of laboratory deceptions. International Journal of Psychiatry, 6 , 282–293.

Perry, G. (2013). Behind the shock machine: The untold story of the notorious Milgram psychology experiments . New York, NY: The New Press.

Reicher, S., Haslam, A., & Miller, A. (Eds.). (2014). Milgram at 50: Exploring the enduring relevance of psychology’s most famous studies [Special issue]. Journal of Social Issues, 70 (3), 393–602

Russell, N. (2014). Stanley Milgram’s obedience to authority “relationship condition”: Some methodological and theoretical implications. Social Sciences, 3, 194–214

Shanab, M. E., & Yahya, K. A. (1978). A cross-cultural study of obedience. Bulletin of the Psychonomic Society .

Smith, P. B., & Bond, M. H. (1998). Social psychology across cultures (2nd Edition) . Prentice Hall.

Further Reading

  • The power of the situation: The impact of Milgram’s obedience studies on personality and social psychology
  • Seeing is believing: The role of the film Obedience in shaping perceptions of Milgram’s Obedience to Authority Experiments
  • Replicating Milgram: Would people still obey today?

Learning Check

Which is true regarding the Milgram obedience study?
  • The aim was to see how obedient people would be in a situation where following orders would mean causing harm to another person.
  • Participants were under the impression they were part of a learning and memory experiment.
  • The “learners” in the study were actual participants who volunteered to be shocked as part of the experiment.
  • The “learner” was an actor who was in on the experiment and never actually received any real shocks.
  • Although the participant could not see the “learner”, he was able to hear him clearly through the wall
  • The study was directly influenced by Milgram’s observations of obedience patterns in post-war Europe.
  • The experiment was designed to understand the psychological mechanisms behind war crimes committed during World War II.
  • The Milgram study was universally accepted in the psychological community, and no ethical concerns were raised about its methodology.
  • When Milgram’s experiment was repeated in a rundown office building in Bridgeport, the percentage of the participants who fully complied with the commands of the experimenter remained unchanged.
  • The experimenter (authority figure) delivered verbal prods to encourage the teacher to continue, such as ‘Please continue’ or ‘Please go on’.
  • Over 80% of participants went on to deliver the maximum level of shock.
  • Milgram sent participants questionnaires after the study to assess the effects and found that most felt no remorse or guilt, so it was ethical.
  • The aftermath of the study led to stricter ethical guidelines in psychological research.
  • The study emphasized the role of situational factors over personality traits in determining obedience.

Answers : Items 3, 8, 9, and 11 are the false statements.

Short Answer Questions
  • Briefly explain the results of the original Milgram experiments. What did these results prove?
  • List one scenario on how an authority figure can abuse obedience principles.
  • List one scenario on how an individual could use these principles to defend their fellow peers.
  • In a hospital, you are very likely to obey a nurse. However, if you meet her outside the hospital, for example in a shop, you are much less likely to obey. Using your knowledge of how people resist pressure to obey, explain why you are less likely to obey the nurse outside the hospital.
  • Describe the shock instructions the participant (teacher) was told to follow when the victim (learner) gave an incorrect answer.
  • State the lowest voltage shock that was labeled on the shock generator.
  • What would likely happen if Milgram’s experiment included a condition in which the participant (teacher) had to give a high-level electric shock for the first wrong answer?
Group Activity

Gather in groups of three or four to discuss answers to the short answer questions above.

For question 2, review the different scenarios you each came up with. Then brainstorm on how these situations could be flipped.

For question 2, discuss how an authority figure could instead empower those below them in the examples your groupmates provide.

For question 3, discuss how a peer could do harm by using the obedience principles in the scenarios your groupmates provide.

Essay Topic
  • What’s the most important lesson of Milgram’s Obedience Experiments? Fully explain and defend your answer.
  • Milgram selectively edited his film of the obedience experiments to emphasize obedient behavior and minimize footage of disobedience. What are the ethical implications of a researcher selectively presenting findings in a way that fits their expected conclusions?

Print Friendly, PDF & Email

  • Bipolar Disorder
  • Therapy Center
  • When To See a Therapist
  • Types of Therapy
  • Best Online Therapy
  • Best Couples Therapy
  • Managing Stress
  • Sleep and Dreaming
  • Understanding Emotions
  • Self-Improvement
  • Healthy Relationships
  • Student Resources
  • Personality Types
  • Sweepstakes
  • Guided Meditations
  • Verywell Mind Insights
  • 2024 Verywell Mind 25
  • Mental Health in the Classroom
  • Editorial Process
  • Meet Our Review Board
  • Crisis Support

Understanding the Milgram Experiment in Psychology

A closer look at Milgram's controversial studies of obedience

Isabelle Adam (CC BY-NC-ND 2.0) via Flickr

Factors That Influence Obedience

  • Ethical Concerns
  • Replications

How far do you think people would go to obey an authority figure? Would they refuse to obey if the order went against their values or social expectations? Those questions were at the heart of an infamous and controversial study known as the Milgram obedience experiments.

Yale University  psychologist   Stanley Milgram  conducted these experiments during the 1960s. They explored the effects of authority on obedience. In the experiments, an authority figure ordered participants to deliver what they believed were dangerous electrical shocks to another person. These results suggested that people are highly influenced by authority and highly obedient . More recent investigations cast doubt on some of the implications of Milgram's findings and even the results and procedures themselves. Despite its problems, the study has, without question, made a significant impact on psychology .

At a Glance

Milgram's experiments posed the question: Would people obey orders, even if they believed doing so would harm another person? Milgram's findings suggested the answer was yes, they would. The experiments have long been controversial, both because of the startling findings and the ethical problems with the research. More recently, experts have re-examined the studies, suggesting that participants were often coerced into obeying and that at least some participants recognized that the other person was just pretending to be shocked. Such findings call into question the study's validity and authenticity, but some replications suggest that people are surprisingly prone to obeying authority.

History of the Milgram Experiments

Milgram started his experiments in 1961, shortly after the trial of the World War II criminal Adolf Eichmann had begun. Eichmann’s defense that he was merely following instructions when he ordered the deaths of millions of Jews roused Milgram’s interest.

In his 1974 book "Obedience to Authority," Milgram posed the question, "Could it be that Eichmann and his million accomplices in the Holocaust were just following orders? Could we call them all accomplices?"

Procedure in the Milgram Experiment

The participants in the most famous variation of the Milgram experiment were 40 men recruited using newspaper ads. In exchange for their participation, each person was paid $4.50.

Milgram developed an intimidating shock generator, with shock levels starting at 15 volts and increasing in 15-volt increments all the way up to 450 volts. The many switches were labeled with terms including "slight shock," "moderate shock," and "danger: severe shock." The final three switches were labeled simply with an ominous "XXX."

Each participant took the role of a "teacher" who would then deliver a shock to the "student" in a neighboring room whenever an incorrect answer was given. While participants believed that they were delivering real shocks to the student, the “student” was a confederate in the experiment who was only pretending to be shocked.

As the experiment progressed, the participant would hear the learner plead to be released or even complain about a heart condition. Once they reached the 300-volt level, the learner would bang on the wall and demand to be released.

Beyond this point, the learner became completely silent and refused to answer any more questions. The experimenter then instructed the participant to treat this silence as an incorrect response and deliver a further shock.

Most participants asked the experimenter whether they should continue. The experimenter then responded with a series of commands to prod the participant along:

  • "Please continue."
  • "The experiment requires that you continue."
  • "It is absolutely essential that you continue."
  • "You have no other choice; you must go on."

Results of the Milgram Experiment

In the Milgram experiment, obedience was measured by the level of shock that the participant was willing to deliver. While many of the subjects became extremely agitated, distraught, and angry at the experimenter, they nevertheless continued to follow orders all the way to the end.

Milgram's results showed that 65% of the participants in the study delivered the maximum shocks. Of the 40 participants in the study, 26 delivered the maximum shocks, while 14 stopped before reaching the highest levels.

Why did so many of the participants in this experiment perform a seemingly brutal act when instructed by an authority figure? According to Milgram, there are some situational factors that can explain such high levels of obedience:

  • The physical presence of an authority figure dramatically increased compliance .
  • The fact that Yale (a trusted and authoritative academic institution) sponsored the study led many participants to believe that the experiment must be safe.
  • The selection of teacher and learner status seemed random.
  • Participants assumed that the experimenter was a competent expert.
  • The shocks were said to be painful, not dangerous.

Later experiments conducted by Milgram indicated that the presence of rebellious peers dramatically reduced obedience levels. When other people refused to go along with the experimenter's orders, 36 out of 40 participants refused to deliver the maximum shocks.

More recent work by researchers suggests that while people do tend to obey authority figures, the process is not necessarily as cut-and-dried as Milgram depicted it.

In a 2012 essay published in PLoS Biology , researchers suggested that the degree to which people are willing to obey the questionable orders of an authority figure depends largely on two key factors:

  • How much the individual agrees with the orders
  • How much they identify with the person giving the orders

While it is clear that people are often far more susceptible to influence, persuasion , and obedience than they would often like to be, they are far from mindless machines just taking orders. 

Another study that analyzed Milgram's results concluded that eight factors influenced the likelihood that people would progress up to the 450-volt shock:

  • The experimenter's directiveness
  • Legitimacy and consistency
  • Group pressure to disobey
  • Indirectness of proximity
  • Intimacy of the relation between the teacher and learner
  • Distance between the teacher and learner

Ethical Concerns in the Milgram Experiment

Milgram's experiments have long been the source of considerable criticism and controversy. From the get-go, the ethics of his experiments were highly dubious. Participants were subjected to significant psychological and emotional distress.

Some of the major ethical issues in the experiment were related to:

  • The use of deception
  • The lack of protection for the participants who were involved
  • Pressure from the experimenter to continue even after asking to stop, interfering with participants' right to withdraw

Due to concerns about the amount of anxiety experienced by many of the participants, everyone was supposedly debriefed at the end of the experiment. The researchers reported that they explained the procedures and the use of deception.

Critics of the study have argued that many of the participants were still confused about the exact nature of the experiment, and recent findings suggest that many participants were not debriefed at all.

Replications of the Milgram Experiment

While Milgram’s research raised serious ethical questions about the use of human subjects in psychology experiments , his results have also been consistently replicated in further experiments. One review further research on obedience and found that Milgram’s findings hold true in other experiments. In one study, researchers conducted a study designed to replicate Milgram's classic obedience experiment. The researchers made several alterations to Milgram's experiment.

  • The maximum shock level was 150 volts as opposed to the original 450 volts.
  • Participants were also carefully screened to eliminate those who might experience adverse reactions to the experiment.

The results of the new experiment revealed that participants obeyed at roughly the same rate that they did when Milgram conducted his original study more than 40 years ago.

Some psychologists suggested that in spite of the changes made in the replication, the study still had merit and could be used to further explore some of the situational factors that also influenced the results of Milgram's study. But other psychologists suggested that the replication was too dissimilar to Milgram's original study to draw any meaningful comparisons.

One study examined people's beliefs about how they would do compared to the participants in Milgram's experiments. They found that most people believed they would stop sooner than the average participants. These findings applied to both those who had never heard of Milgram's experiments and those who were familiar with them. In fact, those who knew about Milgram's experiments actually believed that they would stop even sooner than other people.

Another novel replication involved recruiting participants in pairs and having them take turns acting as either an 'agent' or 'victim.' Agents then received orders to shock the victim. The results suggest that only around 3.3% disobeyed the experimenter's orders.

Recent Criticisms and New Findings

Psychologist Gina Perry suggests that much of what we think we know about Milgram's famous experiments is only part of the story. While researching an article on the topic, she stumbled across hundreds of audiotapes found in Yale archives that documented numerous variations of Milgram's shock experiments.

Participants Were Often Coerced

While Milgram's reports of his process report methodical and uniform procedures, the audiotapes reveal something different. During the experimental sessions, the experimenters often went off-script and coerced the subjects into continuing the shocks.

"The slavish obedience to authority we have come to associate with Milgram’s experiments comes to sound much more like bullying and coercion when you listen to these recordings," Perry suggested in an article for Discover Magazine .

Few Participants Were Really Debriefed

Milgram suggested that the subjects were "de-hoaxed" after the experiments. He claimed he later surveyed the participants and found that 84% were glad to have participated, while only 1% regretted their involvement.

However, Perry's findings revealed that of the 700 or so people who took part in different variations of his studies between 1961 and 1962, very few were truly debriefed.

A true debriefing would have involved explaining that the shocks weren't real and that the other person was not injured. Instead, Milgram's sessions were mainly focused on calming the subjects down before sending them on their way.

Many participants left the experiment in a state of considerable distress. While the truth was revealed to some months or even years later, many were simply never told a thing.

Variations Led to Differing Results

Another problem is that the version of the study presented by Milgram and the one that's most often retold does not tell the whole story. The statistic that 65% of people obeyed orders applied only to one variation of the experiment, in which 26 out of 40 subjects obeyed.

In other variations, far fewer people were willing to follow the experimenters' orders, and in some versions of the study, not a single participant obeyed.

Participants Guessed the Learner Was Faking

Perry even tracked down some of the people who took part in the experiments, as well as Milgram's research assistants. What she discovered is that many of his subjects had deduced what Milgram's intent was and knew that the "learner" was merely pretending.

Such findings cast Milgram's results in a new light. It suggests that not only did Milgram intentionally engage in some hefty misdirection to obtain the results he wanted but that many of his participants were simply playing along.

An analysis of an unpublished study by Milgram's assistant, Taketo Murata, found that participants who believed they were really delivering a shock were less likely to obey, while those who did not believe they were actually inflicting pain were more willing to obey. In other words, the perception of pain increased defiance, while skepticism of pain increased obedience.

A review of Milgram's research materials suggests that the experiments exerted more pressure to obey than the original results suggested. Other variations of the experiment revealed much lower rates of obedience, and many of the participants actually altered their behavior when they guessed the true nature of the experiment.

Impact of the Milgram Experiment

Since there is no way to truly replicate the experiment due to its serious ethical and moral problems, determining whether Milgram's experiment really tells us anything about the power of obedience is impossible to determine.

So why does Milgram's experiment maintain such a powerful hold on our imaginations, even decades after the fact? Perry believes that despite all its ethical issues and the problem of never truly being able to replicate Milgram's procedures, the study has taken on the role of what she calls a "powerful parable."

Milgram's work might not hold the answers to what makes people obey or even the degree to which they truly obey. It has, however, inspired other researchers to explore what makes people follow orders and, perhaps more importantly, what leads them to question authority.

Recent findings undermine the scientific validity of the study. Milgram's work is also not truly replicable due to its ethical problems. However, the study has led to additional research on how situational factors can affect obedience to authority.

Milgram’s experiment has become a classic in psychology , demonstrating the dangers of obedience. The research suggests that situational variables have a stronger sway than personality factors in determining whether people will obey an authority figure. However, other psychologists argue that both external and internal factors heavily influence obedience, such as personal beliefs and overall temperament.

Milgram S.  Obedience to Authority: An Experimental View.  Harper & Row.

Russell N, Gregory R. The Milgram-Holocaust linkage: challenging the present consensus . State Crim J. 2015;4(2):128-153.

Russell NJC. Milgram's obedience to authority experiments: origins and early evolution . Br J Soc Psychol . 2011;50:140-162. doi:10.1348/014466610X492205

Haslam SA, Reicher SD. Contesting the "nature" of conformity: What Milgram and Zimbardo's studies really show . PLoS Biol. 2012;10(11):e1001426. doi:10.1371/journal.pbio.1001426

Milgram S. Liberating effects of group pressure . J Person Soc Psychol. 1965;1(2):127-234. doi:10.1037/h0021650

Haslam N, Loughnan S, Perry G. Meta-Milgram: an empirical synthesis of the obedience experiments .  PLoS One . 2014;9(4):e93927. doi:10.1371/journal.pone.0093927

Perry G. Deception and illusion in Milgram's accounts of the obedience experiments . Theory Appl Ethics . 2013;2(2):79-92.

Blass T. The Milgram paradigm after 35 years: some things we now know about obedience to authority . J Appl Soc Psychol. 1999;29(5):955-978. doi:10.1111/j.1559-1816.1999.tb00134.x

Burger J. Replicating Milgram: Would people still obey today? . Am Psychol . 2009;64(1):1-11. doi:10.1037/a0010932

Elms AC. Obedience lite . American Psychologist . 2009;64(1):32-36. doi:10.1037/a0014473

Miller AG. Reflections on “replicating Milgram” (Burger, 2009) . American Psychologist . 2009;64(1):20-27. doi:10.1037/a0014407

Grzyb T, Dolinski D. Beliefs about obedience levels in studies conducted within the Milgram paradigm: Better than average effect and comparisons of typical behaviors by residents of various nations .  Front Psychol . 2017;8:1632. doi:10.3389/fpsyg.2017.01632

Caspar EA. A novel experimental approach to study disobedience to authority .  Sci Rep . 2021;11(1):22927. doi:10.1038/s41598-021-02334-8

Haslam SA, Reicher SD, Millard K, McDonald R. ‘Happy to have been of service’: The Yale archive as a window into the engaged followership of participants in Milgram’s ‘obedience’ experiments . Br J Soc Psychol . 2015;54:55-83. doi:10.1111/bjso.12074

Perry G, Brannigan A, Wanner RA, Stam H. Credibility and incredulity in Milgram’s obedience experiments: A reanalysis of an unpublished test . Soc Psychol Q . 2020;83(1):88-106. doi:10.1177/0190272519861952

By Kendra Cherry, MSEd Kendra Cherry, MS, is a psychosocial rehabilitation specialist, psychology educator, and author of the "Everything Psychology Book."

The Milgram Experiment: How Far Will You Go to Obey an Order?

Understand the infamous study and its conclusions about human nature

  • Archaeology
  • Ph.D., Psychology, University of California - Santa Barbara
  • B.A., Psychology and Peace & Conflict Studies, University of California - Berkeley

A brief Milgram experiment summary is as follows: In the 1960s, psychologist Stanley Milgram conducted a series of studies on the concepts of obedience and authority. His experiments involved instructing study participants to deliver increasingly high-voltage shocks to an actor in another room, who would scream and eventually go silent as the shocks became stronger. The shocks weren't real, but study participants were made to believe that they were.

Today, the Milgram experiment is widely criticized on both ethical and scientific grounds. However, Milgram's conclusions about humanity's willingness to obey authority figures remain influential and well-known.

Key Takeaways: The Milgram Experiment

  • The goal of the Milgram experiment was to test the extent of humans' willingness to obey orders from an authority figure.
  • Participants were told by an experimenter to administer increasingly powerful electric shocks to another individual. Unbeknownst to the participants, shocks were fake and the individual being shocked was an actor.
  • The majority of participants obeyed, even when the individual being shocked screamed in pain.
  • The experiment has been widely criticized on ethical and scientific grounds.

Detailed Milgram’s Experiment Summary

In the most well-known version of the Milgram experiment, the 40 male participants were told that the experiment focused on the relationship between punishment, learning, and memory. The experimenter then introduced each participant to a second individual, explaining that this second individual was participating in the study as well. Participants were told that they would be randomly assigned to roles of "teacher" and "learner." However, the "second individual" was an actor hired by the research team, and the study was set up so that the true participant would always be assigned to the "teacher" role.

During the Milgram experiment, the learner was located in a separate room from the teacher (the real participant), but the teacher could hear the learner through the wall. The experimenter told the teacher that the learner would memorize word pairs and instructed the teacher to ask the learner questions. If the learner responded incorrectly to a question, the teacher would be asked to administer an electric shock. The shocks started at a relatively mild level (15 volts) but increased in 15-volt increments up to 450 volts. (In actuality, the shocks were fake, but the participant was led to believe they were real.)

Participants were instructed to give a higher shock to the learner with each wrong answer. When the 150-volt shock was administered, the learner would cry out in pain and ask to leave the study. He would then continue crying out with each shock until the 330-volt level, at which point he would stop responding.

During this process, whenever participants expressed hesitation about continuing with the study, the experimenter would urge them to go on with increasingly firm instructions, culminating in the statement, "You have no other choice, you must go on." The study ended when participants refused to obey the experimenter’s demand, or when they gave the learner the highest level of shock on the machine (450 volts).

Milgram found that participants obeyed the experimenter at an unexpectedly high rate: 65% of the participants gave the learner the 450-volt shock.

Critiques of the Milgram Experiment

The Milgram experiment has been widely criticized on ethical grounds. Milgram’s participants were led to believe that they acted in a way that harmed someone else, an experience that could have had long-term consequences. Moreover, an investigation by writer Gina Perry uncovered that some participants appear to not have been fully debriefed after the study —they were told months later, or not at all, that the shocks were fake and the learner wasn’t harmed. Milgram’s studies could not be perfectly recreated today, because researchers today are required to pay much more attention to the safety and well-being of human research subjects.

Researchers have also questioned the scientific validity of Milgram’s results. In her examination of the study, Perry found that Milgram’s experimenter may have gone off script and told participants to obey many more times than the script specified. Additionally, some research suggests that participants may have figured out that the learner was not harmed: in interviews conducted after the Milgram experiment, some participants reported that they didn’t think the learner was in any real danger. This mindset is likely to have affected their behavior in the study.

Variations on the Milgram Experiment

Milgram and other researchers conducted numerous versions of the experiment over time. The participants' levels of compliance with the experimenter’s demands varied greatly from one study to the next. For example, when participants were in closer proximity to the learner (e.g. in the same room), they were less likely to give the learner the highest level of shock.

Another version of the Milgram experiment brought three "teachers" into the experiment room at once. One was a real participant, and the other two were actors hired by the research team. During the experiment, the two non-participant teachers would quit as the level of shocks began to increase. Milgram found that these conditions made the real participant far more likely to "disobey" the experimenter, too: only 10% of participants gave the 450-volt shock to the learner.

In yet another version of the Milgram experiment, two experimenters were present, and during the experiment, they would begin arguing with one another about whether it was right to continue the study. In this version, none of the participants gave the learner the 450-volt shock.

Replicating the Milgram Experiment

Researchers have sought to replicate Milgram's original study with additional safeguards in place to protect participants. In 2009, Jerry Burger replicated Milgram’s famous experiment at Santa Clara University with new safeguards in place: the highest shock level was 150 volts, and participants were told that the shocks were fake immediately after the experiment ended. Additionally, participants were screened by a clinical psychologist before the experiment began, and those found to be at risk of a negative reaction to the study were deemed ineligible to participate.

Burger found that participants obeyed at similar levels as Milgram’s participants: 82.5% of Milgram’s participants gave the learner the 150-volt shock, and 70% of Burger’s participants did the same.

The Legacy of the Milgram Experiment

Milgram’s interpretation of his research was that everyday people are capable of carrying out unthinkable actions in certain circumstances. His research has been used to explain atrocities such as the Holocaust and the Rwandan genocide, though these applications are by no means widely accepted or agreed upon.

Importantly, not all participants obeyed the experimenter’s demands , and Milgram’s studies shed light on the factors that enable people to stand up to authority. In fact, as sociologist Matthew Hollander writes, we may be able to learn from the participants who disobeyed, as their strategies may enable us to respond more effectively to an unethical situation. The Milgram experiment suggested that human beings are susceptible to obeying authority, but it also demonstrated that obedience is not inevitable.

  • Baker, Peter C. “Electric Schlock: Did Stanley Milgram's Famous Obedience Experiments Prove Anything?” Pacific Standard (2013, Sep. 10). https://psmag.com/social-justice/electric-schlock-65377
  • Burger, Jerry M. "Replicating Milgram: Would People Still Obey Today?."  American Psychologist 64.1 (2009): 1-11. http://psycnet.apa.org/buy/2008-19206-001
  • Gilovich, Thomas, Dacher Keltner, and Richard E. Nisbett. Social Psychology . 1st edition, W.W. Norton & Company, 2006.
  • Hollander, Matthew. “How to Be a Hero: Insight From the Milgram Experiment.” HuffPost Contributor Network (2015, Apr. 29). https://www.huffingtonpost.com/entry/how-to-be-a-hero-insight-_b_6566882
  • Jarrett, Christian. “New Analysis Suggests Most Milgram Participants Realised the ‘Obedience Experiments’ Were Not Really Dangerous.” The British Psychological Society: Research Digest (2017, Dec. 12). https://digest.bps.org.uk/2017/12/12/interviews-with-milgram-participants-provide-little-support-for-the-contemporary-theory-of-engaged-followership/
  • Perry, Gina. “The Shocking Truth of the Notorious Milgram Obedience Experiments.” Discover Magazine Blogs (2013, Oct. 2). http://blogs.discovermagazine.com/crux/2013/10/02/the-shocking-truth-of-the-notorious-milgram-obedience-experiments/
  • Romm, Cari. “Rethinking One of Psychology's Most Infamous Experiments.” The Atlantic (2015, Jan. 28) . https://www.theatlantic.com/health/archive/2015/01/rethinking-one-of-psychologys-most-infamous-experiments/384913/
  • Gilligan's Ethics of Care
  • What Was the Robbers Cave Experiment in Psychology?
  • What Is Behaviorism in Psychology?
  • What Is the Zeigarnik Effect? Definition and Examples
  • What Is a Conditioned Response?
  • Psychodynamic Theory: Approaches and Proponents
  • Social Cognitive Theory: How We Learn From the Behavior of Others
  • Kohlberg's Stages of Moral Development
  • What's the Difference Between Eudaimonic and Hedonic Happiness?
  • Genie Wiley, the Feral Child
  • What Is the Law of Effect in Psychology?
  • What Is the Recency Effect in Psychology?
  • Heuristics: The Psychology of Mental Shortcuts
  • What Is Survivor's Guilt? Definition and Examples
  • 5 Psychology Studies That Will Make You Feel Good About Humanity
  • What Is Cognitive Bias? Definition and Examples

Encyclopedia Britannica

  • History & Society
  • Science & Tech
  • Biographies
  • Animals & Nature
  • Geography & Travel
  • Arts & Culture
  • Games & Quizzes
  • On This Day
  • One Good Fact
  • New Articles
  • Lifestyles & Social Issues
  • Philosophy & Religion
  • Politics, Law & Government
  • World History
  • Health & Medicine
  • Browse Biographies
  • Birds, Reptiles & Other Vertebrates
  • Bugs, Mollusks & Other Invertebrates
  • Environment
  • Fossils & Geologic Time
  • Entertainment & Pop Culture
  • Sports & Recreation
  • Visual Arts
  • Demystified
  • Image Galleries
  • Infographics
  • Top Questions
  • Britannica Kids
  • Saving Earth
  • Space Next 50
  • Student Center

Stanley Milgram

  • Where was science invented?
  • When did science begin?

Blackboard inscribed with scientific formulas and calculations in physics and mathematics

Milgram experiment

Our editors will review what you’ve submitted and determine whether to revise the article.

  • Open University - OpenLearn - Psychological research, obedience and ethics: 1 Milgram’s obedience study
  • Social Science LibreTexts - The Milgram Experiment- The Power of Authority
  • Verywell Mind - What was the Milgram Experiment?
  • BCcampus Open Publishing - Ethics in Law Enforcement - The Milgram Experiment
  • Nature - Modern Milgram experiment sheds light on power of authority
  • SimplyPsychology - Stanley Milgram Shock Experiment: Summary, Results, & Ethics
  • University of California - College of Natural Resources - Milgrams Experiment on Obedience to Authority

Stanley Milgram

Milgram experiment , controversial series of experiments examining obedience to authority conducted by social psychologist Stanley Milgram . In the experiment, an authority figure, the conductor of the experiment, would instruct a volunteer participant, labeled the “teacher,” to administer painful, even dangerous, electric shocks to the “learner,” who was actually an actor. Although the shocks were faked, the experiments are widely considered unethical today due to the lack of proper disclosure, informed consent, and subsequent debriefing related to the deception and trauma experienced by the teachers. Some of Milgram’s conclusions have been called into question. Nevertheless, the experiments and their results have been widely cited for their insight into how average people respond to authority.

Milgram conducted his experiments as an assistant professor at Yale University in the early 1960s. In 1961 he began to recruit men from New Haven , Connecticut , for participation in a study he claimed would be focused on memory and learning . The recruits were paid $4.50 at the beginning of the study and were generally between the ages of 20 and 50 and from a variety of employment backgrounds. When they volunteered, they were told that the experiment would test the effect of punishment on learning ability. In truth, the volunteers were the subjects of an experiment on obedience to authority. In all, about 780 people, only about 40 of them women, participated in the experiments, and Milgram published his results in 1963.

stanley milgram obedience experiment ethical issues

Volunteers were told that they would be randomly assigned either a “teacher” or “learner” role, with each teacher administering electric shocks to a learner in another room if the learner failed to answer questions correctly. In actuality, the random draw was fixed so that all the volunteer participants were assigned to the teacher role and the actors were assigned to the learner role. The teachers were then instructed in the electroshock “punishment” they would be administering, with 30 shock levels ranging from 15 to 450 volts. The different shock levels were labeled with descriptions of their effects, such as “Slight Shock,” “Intense Shock,” and “Danger: Severe Shock,” with the final label a grim “XXX.” Each teacher was given a 45-volt shock themselves so that they would better understand the punishment they believed the learner would be receiving. Teachers were then given a series of questions for the learner to answer, with each incorrect answer generally earning the learner a progressively stronger shock. The actor portraying the learner, who was seated out of sight of the teacher, had pre-recorded responses to these shocks that ranged from grunts of pain to screaming and pleading, claims of suffering a heart condition, and eventually dead silence. The experimenter, acting as an authority figure, would encourage the teachers to continue administering shocks, telling them with scripted responses that the experiment must continue despite the reactions of the learner. The infamous result of these experiments was that a disturbingly high number of the teachers were willing to proceed to the maximum voltage level, despite the pleas of the learner and the supposed danger of proceeding.

Milgram’s interest in the subject of authority, and his dark view of the results of his experiments, were deeply informed by his Jewish identity and the context of the Holocaust , which had occurred only a few years before. He had expected that Americans, known for their individualism , would differ from Germans in their willingness to obey authority when it might lead to harming others. Milgram and his students had predicted only 1–3% of participants would administer the maximum shock level. However, in his first official study, 26 of 40 male participants (65%) were convinced to do so and nearly 80% of teachers that continued to administer shocks after 150 volts—the point at which the learner was heard to scream—continued to the maximum of 450 volts. Teachers displayed a range of negative emotional responses to the experiment even as they continued to obey, sometimes pleading with the experimenters to stop the experiment while still participating in it. One teacher believed that he had killed the learner and was moved to tears when he eventually found out that he had not.

stanley milgram obedience experiment ethical issues

Milgram included several variants on the original design of the experiment. In one, the teachers were allowed to select their own voltage levels. In this case, only about 2.5% of participants used the maximum shock level, indicating that they were not inclined to do so without the prompting of an authority figure. In another, there were three teachers, two of whom were not test subjects, but instead had been instructed to protest against the shocks. The existence of peers protesting the experiment made the volunteer teachers less likely to obey. Teachers were also less likely to obey in a variant where they could see the learner and were forced to interact with him.

The Milgram experiment has been highly controversial, both for the ethics of its design and for the reliability of its results and conclusions. It is commonly accepted that the ethics of the experiment would be rejected by mainstream science today, due not only to the handling of the deception involved but also to the extreme stress placed on the teachers, who often reacted emotionally to the experiment and were not debriefed . Some teachers were actually left believing they had genuinely and repeatedly shocked a learner before having the truth revealed to them later. Later researchers examining Milgram’s data also found that the experimenters conducting the tests had sometimes gone off-script in their attempts to coerce the teachers into continuing, and noted that some teachers guessed that they were the subjects of the experiment. However, attempts to validate Milgram’s findings in more ethical ways have often produced similar results.

  • Tools and Resources
  • Customer Services
  • Affective Science
  • Biological Foundations of Psychology
  • Clinical Psychology: Disorders and Therapies
  • Cognitive Psychology/Neuroscience
  • Developmental Psychology
  • Educational/School Psychology
  • Forensic Psychology
  • Health Psychology
  • History and Systems of Psychology
  • Individual Differences
  • Methods and Approaches in Psychology
  • Neuropsychology
  • Organizational and Institutional Psychology
  • Personality
  • Psychology and Other Disciplines
  • Social Psychology
  • Sports Psychology
  • Share This Facebook LinkedIn Twitter

Article contents

Milgram’s experiments on obedience to authority.

  • Stephen Gibson Stephen Gibson Heriot-Watt University, School of Social Sciences
  • https://doi.org/10.1093/acrefore/9780190236557.013.511
  • Published online: 30 June 2020

Stanley Milgram’s experiments on obedience to authority are among the most influential and controversial social scientific studies ever conducted. They remain staples of introductory psychology courses and textbooks, yet their influence reaches far beyond psychology, with myriad other disciplines finding lessons in them. Indeed, the experiments have long since broken free of the confines of academia, occupying a place in popular culture that is unrivaled among psychological experiments. The present article begins with an overview of Milgram’s account of his experimental procedure and findings, before focussing on recent scholarship that has used materials from Milgram’s archive to challenge many of the long-held assumptions about the experiments. Three areas in which our understanding of the obedience experiments has undergone a radical shift in recent years are the subject of particular focus. First, work that has identified new ethical problems with Milgram’s studies is summarized. Second, hitherto unknown methodological variations in Milgram’s experimental procedures are considered. Third, the interactions that took place in the experimental sessions themselves are explored. This work has contributed to a shift in how we see the obedience experiments. Rather than viewing the experiments as demonstrations of people’s propensity to follow orders, it is now clear that people did not follow orders in Milgram’s experiments. The experimenter did a lot more than simply issue orders, and when he did, participants found it relatively straightforward to defy them. These arguments are discussed in relation to the definition of obedience that has typically been adopted in psychology, the need for further historical work on Milgram’s experiments, and the possibilities afforded by the development of a broader project of secondary qualitative analysis of laboratory interaction in psychology experiments.

  • experimentation
  • interaction
  • standardization

You do not currently have access to this article

Please login to access the full content.

Access to the full content requires a subscription

Printed from Oxford Research Encyclopedias, Psychology. Under the terms of the licence agreement, an individual user may print out a single article for personal use (for details see Privacy Policy and Legal Notice).

date: 04 September 2024

  • Cookie Policy
  • Privacy Policy
  • Legal Notice
  • Accessibility
  • [195.158.225.230]
  • 195.158.225.230

Character limit 500 /500

  • Skip to main content
  • Keyboard shortcuts for audio player

Author Interviews

Taking a closer look at milgram's shocking obedience study.

Behind the Shock Machine

Behind the Shock Machine

Buy featured book.

Your purchase helps support NPR programming. How?

  • Independent Bookstores

In the early 1960s, Stanley Milgram, a social psychologist at Yale, conducted a series of experiments that became famous. Unsuspecting Americans were recruited for what purportedly was an experiment in learning. A man who pretended to be a recruit himself was wired up to a phony machine that supposedly administered shocks. He was the "learner." In some versions of the experiment he was in an adjoining room.

The unsuspecting subject of the experiment, the "teacher," read lists of words that tested the learner's memory. Each time the learner got one wrong, which he intentionally did, the teacher was instructed by a man in a white lab coat to deliver a shock. With each wrong answer the voltage went up. From the other room came recorded and convincing protests from the learner — even though no shock was actually being administered.

The results of Milgram's experiment made news and contributed a dismaying piece of wisdom to the public at large: It was reported that almost two-thirds of the subjects were capable of delivering painful, possibly lethal shocks, if told to do so. We are as obedient as Nazi functionaries.

Or are we? Gina Perry, a psychologist from Australia, has written Behind the Shock Machine: The Untold Story of the Notorious Milgram Psychology Experiments . She has been retracing Milgram's steps, interviewing his subjects decades later.

"The thought of quitting never ... occurred to me," study participant Bill Menold told Perry in an Australian radio documentary . "Just to say: 'You know what? I'm walking out of here' — which I could have done. It was like being in a situation that you never thought you would be in, not really being able to think clearly."

In his experiments, Milgram was "looking to investigate what it was that had contributed to the brainwashing of American prisoners of war by the Chinese [in the Korean war]," Perry tells NPR's Robert Siegel.

Interview Highlights

On turning from an admirer of Milgram to a critic

"That was an unexpected outcome for me, really. I regarded Stanley Milgram as a misunderstood genius who'd been penalized in some ways for revealing something troubling and profound about human nature. By the end of my research I actually had quite a very different view of the man and the research."

Watch A Video Of One Of The Milgram Obedience Experiments

On the many variations of the experiment

"Over 700 people took part in the experiments. When the news of the experiment was first reported, and the shocking statistic that 65 percent of people went to maximum voltage on the shock machine was reported, very few people, I think, realized then and even realize today that that statistic applied to 26 of 40 people. Of those other 700-odd people, obedience rates varied enormously. In fact, there were variations of the experiment where no one obeyed."

On how Milgram's study coincided with the trial of Nazi officer Adolf Eichmann — and how the experiment reinforced what Hannah Arendt described as "the banality of evil"

"The Eichmann trial was a televised trial and it did reintroduce the whole idea of the Holocaust to a new American public. And Milgram very much, I think, believed that Hannah Arendt's view of Eichmann as a cog in a bureaucratic machine was something that was just as applicable to Americans in New Haven as it was to people in Germany."

On the ethics of working with human subjects

"Certainly for people in academia and scholars the ethical issues involved in Milgram's experiment have always been a hot issue. They were from the very beginning. And Milgram's experiment really ignited a debate particularly in social sciences about what was acceptable to put human subjects through."

stanley milgram obedience experiment ethical issues

Gina Perry is an Australian psychologist. She has previously written for The Age and The Australian. Chris Beck/Courtesy of The New Press hide caption

Gina Perry is an Australian psychologist. She has previously written for The Age and The Australian.

On conversations with the subjects, decades after the experiment

"[Bill Menold] doesn't sound resentful. I'd say he sounds thoughtful and he has reflected a lot on the experiment and the impact that it's had on him and what it meant at the time. I did interview someone else who had been disobedient in the experiment but still very much resented 50 years later that he'd never been de-hoaxed at the time and he found that really unacceptable."

On the problem that one of social psychology's most famous findings cannot be replicated

"I think it leaves social psychology in a difficult situation. ... it is such an iconic experiment. And I think it really leads to the question of why it is that we continue to refer to and believe in Milgram's results. I think the reason that Milgram's experiment is still so famous today is because in a way it's like a powerful parable. It's so widely known and so often quoted that it's taken on a life of its own. ... This experiment and this story about ourselves plays some role for us 50 years later."

Related NPR Stories

Shocking TV Experiment Sparks Ethical Concerns

Shocking TV Experiment Sparks Ethical Concerns

How stanley milgram 'shocked the world', research news, scientists debate 'six degrees of separation'.

  • Search Menu

Sign in through your institution

  • Browse content in Arts and Humanities
  • Browse content in Archaeology
  • Anglo-Saxon and Medieval Archaeology
  • Archaeological Methodology and Techniques
  • Archaeology by Region
  • Archaeology of Religion
  • Archaeology of Trade and Exchange
  • Biblical Archaeology
  • Contemporary and Public Archaeology
  • Environmental Archaeology
  • Historical Archaeology
  • History and Theory of Archaeology
  • Industrial Archaeology
  • Landscape Archaeology
  • Mortuary Archaeology
  • Prehistoric Archaeology
  • Underwater Archaeology
  • Zooarchaeology
  • Browse content in Architecture
  • Architectural Structure and Design
  • History of Architecture
  • Residential and Domestic Buildings
  • Theory of Architecture
  • Browse content in Art
  • Art Subjects and Themes
  • History of Art
  • Industrial and Commercial Art
  • Theory of Art
  • Biographical Studies
  • Byzantine Studies
  • Browse content in Classical Studies
  • Classical History
  • Classical Philosophy
  • Classical Mythology
  • Classical Numismatics
  • Classical Literature
  • Classical Reception
  • Classical Art and Architecture
  • Classical Oratory and Rhetoric
  • Greek and Roman Papyrology
  • Greek and Roman Epigraphy
  • Greek and Roman Law
  • Greek and Roman Archaeology
  • Late Antiquity
  • Religion in the Ancient World
  • Social History
  • Digital Humanities
  • Browse content in History
  • Colonialism and Imperialism
  • Diplomatic History
  • Environmental History
  • Genealogy, Heraldry, Names, and Honours
  • Genocide and Ethnic Cleansing
  • Historical Geography
  • History by Period
  • History of Emotions
  • History of Agriculture
  • History of Education
  • History of Gender and Sexuality
  • Industrial History
  • Intellectual History
  • International History
  • Labour History
  • Legal and Constitutional History
  • Local and Family History
  • Maritime History
  • Military History
  • National Liberation and Post-Colonialism
  • Oral History
  • Political History
  • Public History
  • Regional and National History
  • Revolutions and Rebellions
  • Slavery and Abolition of Slavery
  • Social and Cultural History
  • Theory, Methods, and Historiography
  • Urban History
  • World History
  • Browse content in Language Teaching and Learning
  • Language Learning (Specific Skills)
  • Language Teaching Theory and Methods
  • Browse content in Linguistics
  • Applied Linguistics
  • Cognitive Linguistics
  • Computational Linguistics
  • Forensic Linguistics
  • Grammar, Syntax and Morphology
  • Historical and Diachronic Linguistics
  • History of English
  • Language Evolution
  • Language Reference
  • Language Acquisition
  • Language Variation
  • Language Families
  • Lexicography
  • Linguistic Anthropology
  • Linguistic Theories
  • Linguistic Typology
  • Phonetics and Phonology
  • Psycholinguistics
  • Sociolinguistics
  • Translation and Interpretation
  • Writing Systems
  • Browse content in Literature
  • Bibliography
  • Children's Literature Studies
  • Literary Studies (Romanticism)
  • Literary Studies (American)
  • Literary Studies (Asian)
  • Literary Studies (European)
  • Literary Studies (Eco-criticism)
  • Literary Studies (Modernism)
  • Literary Studies - World
  • Literary Studies (1500 to 1800)
  • Literary Studies (19th Century)
  • Literary Studies (20th Century onwards)
  • Literary Studies (African American Literature)
  • Literary Studies (British and Irish)
  • Literary Studies (Early and Medieval)
  • Literary Studies (Fiction, Novelists, and Prose Writers)
  • Literary Studies (Gender Studies)
  • Literary Studies (Graphic Novels)
  • Literary Studies (History of the Book)
  • Literary Studies (Plays and Playwrights)
  • Literary Studies (Poetry and Poets)
  • Literary Studies (Postcolonial Literature)
  • Literary Studies (Queer Studies)
  • Literary Studies (Science Fiction)
  • Literary Studies (Travel Literature)
  • Literary Studies (War Literature)
  • Literary Studies (Women's Writing)
  • Literary Theory and Cultural Studies
  • Mythology and Folklore
  • Shakespeare Studies and Criticism
  • Browse content in Media Studies
  • Browse content in Music
  • Applied Music
  • Dance and Music
  • Ethics in Music
  • Ethnomusicology
  • Gender and Sexuality in Music
  • Medicine and Music
  • Music Cultures
  • Music and Media
  • Music and Religion
  • Music and Culture
  • Music Education and Pedagogy
  • Music Theory and Analysis
  • Musical Scores, Lyrics, and Libretti
  • Musical Structures, Styles, and Techniques
  • Musicology and Music History
  • Performance Practice and Studies
  • Race and Ethnicity in Music
  • Sound Studies
  • Browse content in Performing Arts
  • Browse content in Philosophy
  • Aesthetics and Philosophy of Art
  • Epistemology
  • Feminist Philosophy
  • History of Western Philosophy
  • Metaphysics
  • Moral Philosophy
  • Non-Western Philosophy
  • Philosophy of Language
  • Philosophy of Mind
  • Philosophy of Perception
  • Philosophy of Science
  • Philosophy of Action
  • Philosophy of Law
  • Philosophy of Religion
  • Philosophy of Mathematics and Logic
  • Practical Ethics
  • Social and Political Philosophy
  • Browse content in Religion
  • Biblical Studies
  • Christianity
  • East Asian Religions
  • History of Religion
  • Judaism and Jewish Studies
  • Qumran Studies
  • Religion and Education
  • Religion and Health
  • Religion and Politics
  • Religion and Science
  • Religion and Law
  • Religion and Art, Literature, and Music
  • Religious Studies
  • Browse content in Society and Culture
  • Cookery, Food, and Drink
  • Cultural Studies
  • Customs and Traditions
  • Ethical Issues and Debates
  • Hobbies, Games, Arts and Crafts
  • Natural world, Country Life, and Pets
  • Popular Beliefs and Controversial Knowledge
  • Sports and Outdoor Recreation
  • Technology and Society
  • Travel and Holiday
  • Visual Culture
  • Browse content in Law
  • Arbitration
  • Browse content in Company and Commercial Law
  • Commercial Law
  • Company Law
  • Browse content in Comparative Law
  • Systems of Law
  • Competition Law
  • Browse content in Constitutional and Administrative Law
  • Government Powers
  • Judicial Review
  • Local Government Law
  • Military and Defence Law
  • Parliamentary and Legislative Practice
  • Construction Law
  • Contract Law
  • Browse content in Criminal Law
  • Criminal Procedure
  • Criminal Evidence Law
  • Sentencing and Punishment
  • Employment and Labour Law
  • Environment and Energy Law
  • Browse content in Financial Law
  • Banking Law
  • Insolvency Law
  • History of Law
  • Human Rights and Immigration
  • Intellectual Property Law
  • Browse content in International Law
  • Private International Law and Conflict of Laws
  • Public International Law
  • IT and Communications Law
  • Jurisprudence and Philosophy of Law
  • Law and Politics
  • Law and Society
  • Browse content in Legal System and Practice
  • Courts and Procedure
  • Legal Skills and Practice
  • Legal System - Costs and Funding
  • Primary Sources of Law
  • Regulation of Legal Profession
  • Medical and Healthcare Law
  • Browse content in Policing
  • Criminal Investigation and Detection
  • Police and Security Services
  • Police Procedure and Law
  • Police Regional Planning
  • Browse content in Property Law
  • Personal Property Law
  • Restitution
  • Study and Revision
  • Terrorism and National Security Law
  • Browse content in Trusts Law
  • Wills and Probate or Succession
  • Browse content in Medicine and Health
  • Browse content in Allied Health Professions
  • Arts Therapies
  • Clinical Science
  • Dietetics and Nutrition
  • Occupational Therapy
  • Operating Department Practice
  • Physiotherapy
  • Radiography
  • Speech and Language Therapy
  • Browse content in Anaesthetics
  • General Anaesthesia
  • Clinical Neuroscience
  • Browse content in Clinical Medicine
  • Acute Medicine
  • Cardiovascular Medicine
  • Clinical Genetics
  • Clinical Pharmacology and Therapeutics
  • Dermatology
  • Endocrinology and Diabetes
  • Gastroenterology
  • Genito-urinary Medicine
  • Geriatric Medicine
  • Infectious Diseases
  • Medical Toxicology
  • Medical Oncology
  • Pain Medicine
  • Palliative Medicine
  • Rehabilitation Medicine
  • Respiratory Medicine and Pulmonology
  • Rheumatology
  • Sleep Medicine
  • Sports and Exercise Medicine
  • Community Medical Services
  • Critical Care
  • Emergency Medicine
  • Forensic Medicine
  • Haematology
  • History of Medicine
  • Browse content in Medical Skills
  • Clinical Skills
  • Communication Skills
  • Nursing Skills
  • Surgical Skills
  • Browse content in Medical Dentistry
  • Oral and Maxillofacial Surgery
  • Paediatric Dentistry
  • Restorative Dentistry and Orthodontics
  • Surgical Dentistry
  • Medical Ethics
  • Medical Statistics and Methodology
  • Browse content in Neurology
  • Clinical Neurophysiology
  • Neuropathology
  • Nursing Studies
  • Browse content in Obstetrics and Gynaecology
  • Gynaecology
  • Occupational Medicine
  • Ophthalmology
  • Otolaryngology (ENT)
  • Browse content in Paediatrics
  • Neonatology
  • Browse content in Pathology
  • Chemical Pathology
  • Clinical Cytogenetics and Molecular Genetics
  • Histopathology
  • Medical Microbiology and Virology
  • Patient Education and Information
  • Browse content in Pharmacology
  • Psychopharmacology
  • Browse content in Popular Health
  • Caring for Others
  • Complementary and Alternative Medicine
  • Self-help and Personal Development
  • Browse content in Preclinical Medicine
  • Cell Biology
  • Molecular Biology and Genetics
  • Reproduction, Growth and Development
  • Primary Care
  • Professional Development in Medicine
  • Browse content in Psychiatry
  • Addiction Medicine
  • Child and Adolescent Psychiatry
  • Forensic Psychiatry
  • Learning Disabilities
  • Old Age Psychiatry
  • Psychotherapy
  • Browse content in Public Health and Epidemiology
  • Epidemiology
  • Public Health
  • Browse content in Radiology
  • Clinical Radiology
  • Interventional Radiology
  • Nuclear Medicine
  • Radiation Oncology
  • Reproductive Medicine
  • Browse content in Surgery
  • Cardiothoracic Surgery
  • Gastro-intestinal and Colorectal Surgery
  • General Surgery
  • Neurosurgery
  • Paediatric Surgery
  • Peri-operative Care
  • Plastic and Reconstructive Surgery
  • Surgical Oncology
  • Transplant Surgery
  • Trauma and Orthopaedic Surgery
  • Vascular Surgery
  • Browse content in Science and Mathematics
  • Browse content in Biological Sciences
  • Aquatic Biology
  • Biochemistry
  • Bioinformatics and Computational Biology
  • Developmental Biology
  • Ecology and Conservation
  • Evolutionary Biology
  • Genetics and Genomics
  • Microbiology
  • Molecular and Cell Biology
  • Natural History
  • Plant Sciences and Forestry
  • Research Methods in Life Sciences
  • Structural Biology
  • Systems Biology
  • Zoology and Animal Sciences
  • Browse content in Chemistry
  • Analytical Chemistry
  • Computational Chemistry
  • Crystallography
  • Environmental Chemistry
  • Industrial Chemistry
  • Inorganic Chemistry
  • Materials Chemistry
  • Medicinal Chemistry
  • Mineralogy and Gems
  • Organic Chemistry
  • Physical Chemistry
  • Polymer Chemistry
  • Study and Communication Skills in Chemistry
  • Theoretical Chemistry
  • Browse content in Computer Science
  • Artificial Intelligence
  • Computer Architecture and Logic Design
  • Game Studies
  • Human-Computer Interaction
  • Mathematical Theory of Computation
  • Programming Languages
  • Software Engineering
  • Systems Analysis and Design
  • Virtual Reality
  • Browse content in Computing
  • Business Applications
  • Computer Security
  • Computer Games
  • Computer Networking and Communications
  • Digital Lifestyle
  • Graphical and Digital Media Applications
  • Operating Systems
  • Browse content in Earth Sciences and Geography
  • Atmospheric Sciences
  • Environmental Geography
  • Geology and the Lithosphere
  • Maps and Map-making
  • Meteorology and Climatology
  • Oceanography and Hydrology
  • Palaeontology
  • Physical Geography and Topography
  • Regional Geography
  • Soil Science
  • Urban Geography
  • Browse content in Engineering and Technology
  • Agriculture and Farming
  • Biological Engineering
  • Civil Engineering, Surveying, and Building
  • Electronics and Communications Engineering
  • Energy Technology
  • Engineering (General)
  • Environmental Science, Engineering, and Technology
  • History of Engineering and Technology
  • Mechanical Engineering and Materials
  • Technology of Industrial Chemistry
  • Transport Technology and Trades
  • Browse content in Environmental Science
  • Applied Ecology (Environmental Science)
  • Conservation of the Environment (Environmental Science)
  • Environmental Sustainability
  • Environmentalist Thought and Ideology (Environmental Science)
  • Management of Land and Natural Resources (Environmental Science)
  • Natural Disasters (Environmental Science)
  • Nuclear Issues (Environmental Science)
  • Pollution and Threats to the Environment (Environmental Science)
  • Social Impact of Environmental Issues (Environmental Science)
  • History of Science and Technology
  • Browse content in Materials Science
  • Ceramics and Glasses
  • Composite Materials
  • Metals, Alloying, and Corrosion
  • Nanotechnology
  • Browse content in Mathematics
  • Applied Mathematics
  • Biomathematics and Statistics
  • History of Mathematics
  • Mathematical Education
  • Mathematical Finance
  • Mathematical Analysis
  • Numerical and Computational Mathematics
  • Probability and Statistics
  • Pure Mathematics
  • Browse content in Neuroscience
  • Cognition and Behavioural Neuroscience
  • Development of the Nervous System
  • Disorders of the Nervous System
  • History of Neuroscience
  • Invertebrate Neurobiology
  • Molecular and Cellular Systems
  • Neuroendocrinology and Autonomic Nervous System
  • Neuroscientific Techniques
  • Sensory and Motor Systems
  • Browse content in Physics
  • Astronomy and Astrophysics
  • Atomic, Molecular, and Optical Physics
  • Biological and Medical Physics
  • Classical Mechanics
  • Computational Physics
  • Condensed Matter Physics
  • Electromagnetism, Optics, and Acoustics
  • History of Physics
  • Mathematical and Statistical Physics
  • Measurement Science
  • Nuclear Physics
  • Particles and Fields
  • Plasma Physics
  • Quantum Physics
  • Relativity and Gravitation
  • Semiconductor and Mesoscopic Physics
  • Browse content in Psychology
  • Affective Sciences
  • Clinical Psychology
  • Cognitive Psychology
  • Cognitive Neuroscience
  • Criminal and Forensic Psychology
  • Developmental Psychology
  • Educational Psychology
  • Evolutionary Psychology
  • Health Psychology
  • History and Systems in Psychology
  • Music Psychology
  • Neuropsychology
  • Organizational Psychology
  • Psychological Assessment and Testing
  • Psychology of Human-Technology Interaction
  • Psychology Professional Development and Training
  • Research Methods in Psychology
  • Social Psychology
  • Browse content in Social Sciences
  • Browse content in Anthropology
  • Anthropology of Religion
  • Human Evolution
  • Medical Anthropology
  • Physical Anthropology
  • Regional Anthropology
  • Social and Cultural Anthropology
  • Theory and Practice of Anthropology
  • Browse content in Business and Management
  • Business Ethics
  • Business Strategy
  • Business History
  • Business and Technology
  • Business and Government
  • Business and the Environment
  • Comparative Management
  • Corporate Governance
  • Corporate Social Responsibility
  • Entrepreneurship
  • Health Management
  • Human Resource Management
  • Industrial and Employment Relations
  • Industry Studies
  • Information and Communication Technologies
  • International Business
  • Knowledge Management
  • Management and Management Techniques
  • Operations Management
  • Organizational Theory and Behaviour
  • Pensions and Pension Management
  • Public and Nonprofit Management
  • Social Issues in Business and Management
  • Strategic Management
  • Supply Chain Management
  • Browse content in Criminology and Criminal Justice
  • Criminal Justice
  • Criminology
  • Forms of Crime
  • International and Comparative Criminology
  • Youth Violence and Juvenile Justice
  • Development Studies
  • Browse content in Economics
  • Agricultural, Environmental, and Natural Resource Economics
  • Asian Economics
  • Behavioural Finance
  • Behavioural Economics and Neuroeconomics
  • Econometrics and Mathematical Economics
  • Economic History
  • Economic Systems
  • Economic Methodology
  • Economic Development and Growth
  • Financial Markets
  • Financial Institutions and Services
  • General Economics and Teaching
  • Health, Education, and Welfare
  • History of Economic Thought
  • International Economics
  • Labour and Demographic Economics
  • Law and Economics
  • Macroeconomics and Monetary Economics
  • Microeconomics
  • Public Economics
  • Urban, Rural, and Regional Economics
  • Welfare Economics
  • Browse content in Education
  • Adult Education and Continuous Learning
  • Care and Counselling of Students
  • Early Childhood and Elementary Education
  • Educational Equipment and Technology
  • Educational Strategies and Policy
  • Higher and Further Education
  • Organization and Management of Education
  • Philosophy and Theory of Education
  • Schools Studies
  • Secondary Education
  • Teaching of a Specific Subject
  • Teaching of Specific Groups and Special Educational Needs
  • Teaching Skills and Techniques
  • Browse content in Environment
  • Applied Ecology (Social Science)
  • Climate Change
  • Conservation of the Environment (Social Science)
  • Environmentalist Thought and Ideology (Social Science)
  • Management of Land and Natural Resources (Social Science)
  • Natural Disasters (Environment)
  • Pollution and Threats to the Environment (Social Science)
  • Social Impact of Environmental Issues (Social Science)
  • Sustainability
  • Browse content in Human Geography
  • Cultural Geography
  • Economic Geography
  • Political Geography
  • Browse content in Interdisciplinary Studies
  • Communication Studies
  • Museums, Libraries, and Information Sciences
  • Browse content in Politics
  • African Politics
  • Asian Politics
  • Chinese Politics
  • Comparative Politics
  • Conflict Politics
  • Elections and Electoral Studies
  • Environmental Politics
  • Ethnic Politics
  • European Union
  • Foreign Policy
  • Gender and Politics
  • Human Rights and Politics
  • Indian Politics
  • International Relations
  • International Organization (Politics)
  • Irish Politics
  • Latin American Politics
  • Middle Eastern Politics
  • Political Behaviour
  • Political Economy
  • Political Institutions
  • Political Methodology
  • Political Communication
  • Political Philosophy
  • Political Sociology
  • Political Theory
  • Politics and Law
  • Politics of Development
  • Public Policy
  • Public Administration
  • Qualitative Political Methodology
  • Quantitative Political Methodology
  • Regional Political Studies
  • Russian Politics
  • Security Studies
  • State and Local Government
  • UK Politics
  • US Politics
  • Browse content in Regional and Area Studies
  • African Studies
  • Asian Studies
  • East Asian Studies
  • Japanese Studies
  • Latin American Studies
  • Middle Eastern Studies
  • Native American Studies
  • Scottish Studies
  • Browse content in Research and Information
  • Research Methods
  • Browse content in Social Work
  • Addictions and Substance Misuse
  • Adoption and Fostering
  • Care of the Elderly
  • Child and Adolescent Social Work
  • Couple and Family Social Work
  • Direct Practice and Clinical Social Work
  • Emergency Services
  • Human Behaviour and the Social Environment
  • International and Global Issues in Social Work
  • Mental and Behavioural Health
  • Social Justice and Human Rights
  • Social Policy and Advocacy
  • Social Work and Crime and Justice
  • Social Work Macro Practice
  • Social Work Practice Settings
  • Social Work Research and Evidence-based Practice
  • Welfare and Benefit Systems
  • Browse content in Sociology
  • Childhood Studies
  • Community Development
  • Comparative and Historical Sociology
  • Disability Studies
  • Economic Sociology
  • Gender and Sexuality
  • Gerontology and Ageing
  • Health, Illness, and Medicine
  • Marriage and the Family
  • Migration Studies
  • Occupations, Professions, and Work
  • Organizations
  • Population and Demography
  • Race and Ethnicity
  • Social Theory
  • Social Movements and Social Change
  • Social Research and Statistics
  • Social Stratification, Inequality, and Mobility
  • Sociology of Religion
  • Sociology of Education
  • Sport and Leisure
  • Urban and Rural Studies
  • Browse content in Warfare and Defence
  • Defence Strategy, Planning, and Research
  • Land Forces and Warfare
  • Military Administration
  • Military Life and Institutions
  • Naval Forces and Warfare
  • Other Warfare and Defence Issues
  • Peace Studies and Conflict Resolution
  • Weapons and Equipment

Morality in the Making of Sense and Self: Stanley Milgram's Obedience Experiments and the New Science of Morality

Morality in the Making of Sense and Self: Stanley Milgram's Obedience Experiments and the New Science of Morality

  • Cite Icon Cite
  • Permissions Icon Permissions

For over half a century, Stanley Milgram’s classic and controversial obedience experiments have been a touchstone in the social and behavioral sciences, introducing generations of students to the concept of destructive obedience to authority and the Holocaust. In the last decade, the interdisciplinary Milgram renaissance has led to widespread interest in rethinking and challenging the context and nature of his Obedience Experiment. In Morality in the Making of Sense and Self , Matthew M. Hollander and Jason Turowetz offer a new explanation of obedience and defiance in Milgram’s lab. Examining one of the largest collections of Milgram’s original audiotapes, they scrutinize participant behavior in not only the experiments themselves, but also recordings of the subsequent debriefing interviews in which participants were asked to reflect on their actions. Taking an interaction centered approach to the sociology of morality, they show that, contrary to traditional understandings of Milgram’s experiments that highlight obedience, virtually all subjects, both compliant and defiant, mobilized practices to resist the authority’s commands, such that all were obedient and disobedient to varying degrees. As Hollander and Turowetz show, whether participants explicitly defined the situation as a moral problem mattered greatly for the outcomes they achieved, shaping their response to the authority’s demands and ultimately whether they would be classified as “obedient” or “defiant.” By illuminating the relationship between concrete moral dilemmas and social interaction, Hollander and Turowetz tell a new, empirically-grounded story about Milgram: one about morality—and immorality—in the making of sense and self.

Personal account

  • Sign in with email/username & password
  • Get email alerts
  • Save searches
  • Purchase content
  • Activate your purchase/trial code
  • Add your ORCID iD

Institutional access

Sign in with a library card.

  • Sign in with username/password
  • Recommend to your librarian
  • Institutional account management
  • Get help with access

Access to content on Oxford Academic is often provided through institutional subscriptions and purchases. If you are a member of an institution with an active account, you may be able to access content in one of the following ways:

IP based access

Typically, access is provided across an institutional network to a range of IP addresses. This authentication occurs automatically, and it is not possible to sign out of an IP authenticated account.

Choose this option to get remote access when outside your institution. Shibboleth/Open Athens technology is used to provide single sign-on between your institution’s website and Oxford Academic.

  • Click Sign in through your institution.
  • Select your institution from the list provided, which will take you to your institution's website to sign in.
  • When on the institution site, please use the credentials provided by your institution. Do not use an Oxford Academic personal account.
  • Following successful sign in, you will be returned to Oxford Academic.

If your institution is not listed or you cannot sign in to your institution’s website, please contact your librarian or administrator.

Enter your library card number to sign in. If you cannot sign in, please contact your librarian.

Society Members

Society member access to a journal is achieved in one of the following ways:

Sign in through society site

Many societies offer single sign-on between the society website and Oxford Academic. If you see ‘Sign in through society site’ in the sign in pane within a journal:

  • Click Sign in through society site.
  • When on the society site, please use the credentials provided by that society. Do not use an Oxford Academic personal account.

If you do not have a society account or have forgotten your username or password, please contact your society.

Sign in using a personal account

Some societies use Oxford Academic personal accounts to provide access to their members. See below.

A personal account can be used to get email alerts, save searches, purchase content, and activate subscriptions.

Some societies use Oxford Academic personal accounts to provide access to their members.

Viewing your signed in accounts

Click the account icon in the top right to:

  • View your signed in personal account and access account management features.
  • View the institutional accounts that are providing access.

Signed in but can't access content

Oxford Academic is home to a wide variety of products. The institutional subscription may not cover the content that you are trying to access. If you believe you should have access to that content, please contact your librarian.

For librarians and administrators, your personal account also provides access to institutional account management. Here you will find options to view and activate subscriptions, manage institutional settings and access options, access usage statistics, and more.

Our books are available by subscription or purchase to libraries and institutions.

Month: Total Views:
July 2023 1
July 2023 1
July 2023 1
July 2023 1
July 2023 1
July 2023 1
July 2023 3
July 2023 1
July 2023 1
July 2023 1
August 2023 3
August 2023 4
August 2023 4
August 2023 4
August 2023 2
August 2023 5
August 2023 4
August 2023 2
August 2023 4
August 2023 10
August 2023 7
August 2023 3
August 2023 3
August 2023 4
August 2023 8
August 2023 9
September 2023 2
September 2023 2
September 2023 2
September 2023 1
September 2023 2
September 2023 2
September 2023 4
September 2023 3
September 2023 2
September 2023 2
October 2023 2
October 2023 2
October 2023 2
October 2023 2
October 2023 2
October 2023 2
October 2023 3
October 2023 1
October 2023 2
October 2023 4
October 2023 2
October 2023 1
November 2023 6
November 2023 2
November 2023 1
November 2023 5
November 2023 2
November 2023 2
November 2023 2
November 2023 7
November 2023 5
November 2023 1
November 2023 3
November 2023 2
November 2023 10
November 2023 6
November 2023 2
November 2023 2
December 2023 3
December 2023 2
December 2023 1
December 2023 3
December 2023 1
December 2023 5
January 2024 7
January 2024 4
January 2024 5
January 2024 5
January 2024 3
January 2024 4
January 2024 4
January 2024 4
January 2024 4
January 2024 4
January 2024 5
January 2024 6
January 2024 4
January 2024 2
January 2024 6
January 2024 5
February 2024 2
February 2024 5
February 2024 5
February 2024 4
February 2024 4
February 2024 8
February 2024 1
February 2024 9
February 2024 3
February 2024 6
February 2024 5
February 2024 6
February 2024 2
February 2024 4
March 2024 5
March 2024 1
March 2024 4
March 2024 2
March 2024 2
March 2024 2
March 2024 7
March 2024 5
March 2024 5
March 2024 2
March 2024 7
March 2024 4
March 2024 4
March 2024 5
March 2024 4
March 2024 2
April 2024 6
April 2024 2
April 2024 2
April 2024 7
April 2024 9
April 2024 12
April 2024 5
April 2024 15
April 2024 4
April 2024 2
April 2024 9
April 2024 10
April 2024 1
April 2024 5
May 2024 4
May 2024 4
May 2024 6
May 2024 2
May 2024 2
May 2024 2
May 2024 9
May 2024 1
May 2024 2
June 2024 3
June 2024 5
June 2024 3
June 2024 2
June 2024 2
June 2024 1
June 2024 3
June 2024 3
June 2024 6
June 2024 2
June 2024 13
June 2024 3
June 2024 3
June 2024 8
June 2024 3
June 2024 3
July 2024 1
July 2024 4
July 2024 2
July 2024 1
July 2024 4
July 2024 1
July 2024 2
July 2024 3
July 2024 2
July 2024 1
August 2024 2
August 2024 1
August 2024 2
August 2024 2
August 2024 2
August 2024 4
August 2024 1
August 2024 3
August 2024 1
August 2024 2
August 2024 1
August 2024 2
August 2024 2
August 2024 1
  • About Oxford Academic
  • Publish journals with us
  • University press partners
  • What we publish
  • New features  
  • Open access
  • Rights and permissions
  • Accessibility
  • Advertising
  • Media enquiries
  • Oxford University Press
  • Oxford Languages
  • University of Oxford

Oxford University Press is a department of the University of Oxford. It furthers the University's objective of excellence in research, scholarship, and education by publishing worldwide

  • Copyright © 2024 Oxford University Press
  • Cookie settings
  • Cookie policy
  • Privacy policy
  • Legal notice

This Feature Is Available To Subscribers Only

Sign In or Create an Account

This PDF is available to Subscribers Only

For full access to this pdf, sign in to an existing account, or purchase an annual subscription.

Academia’s Response to Milgram’s Findings and Explanation

  • First Online: 18 September 2018

Cite this chapter

stanley milgram obedience experiment ethical issues

  • Nestar Russell 2  

1058 Accesses

In this chapter, Russell provides a brief overview of the key issues that Stanley Milgram’s academic peers debated after the publication of his Obedience to Authority research. More specifically, Russell presents and assesses the prominent ethical and methodological critiques of Milgram’s research. Then with a focus on the Holocaust, Russell explores the debate over the generalizability of Milgram’s results beyond the laboratory walls. Finally, Russell examines the scholarly reaction to Milgram’s agentic state theory, particularly with reference to its application to the Milgram-Holocaust linkage.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save.

  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
  • Available as EPUB and PDF
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
  • Durable hardcover edition

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Ancona, L., & Pareyson, R. (1968). Contributo allo studio della aggressione: La dinamica della obbedienza distruttiva [Contribution to the study of aggression: The dynamics of destructive obedience]. Archivio di Psicologiu. Neurologia. e Psichiatria, 29 (4), 340–372.

Google Scholar  

Askenasy, H. (1978). Are we all Nazis? Secaucus, NJ: Lyle Stuart Inc.

Bandura, A. (1999). Moral disengagement in the perpetration of inhumanities. Personality and Social Psychology Review, 3 (3), 193–209.

Article   PubMed   Google Scholar  

Bandura, A., Underwood, B., & Fromson, M. E. (1975). Disinhibition of aggression through diffusion of responsibility and dehumanization of victims. Journal of Research in Personality, 9 (4), 253–269.

Article   Google Scholar  

Bartov, O. (2001). The Eastern Front, 1941–45, German troops and the barbarization of warfare (2nd ed.). New York: Palgrave.

Bauman, Z. (1989). Modernity and the Holocaust . Ithaca, NY: Cornell University Press.

Baumrind, D. (1964). Some thoughts on ethics of research: After reading Milgram’s ‘behavioral study of obedience’. American Psychologist, 19 (6), 421–423.

Baumrind, D. (2013). Is Milgram’s deceptive research ethically acceptable? Theoretical and Applied Ethics, 2 (2), 1–18.

Baumrind, D. (2015). When subjects become objects: The lies behind the Milgram legend. Theory and Psychology, 25 (5), 690–696.

Beauvois, J. L., Courbet, D., & Oberlé, D. (2012). The prescriptive power of the television host. A transposition of Milgram’s obedience paradigm to the context of TV game show. European Review of Applied Psychology, 62 (3), 111–119.

Berger, L. (1983). A psychological perspective on the Holocaust: Is mass murder part of human behavior? In R. L. Braham (Ed.), Perspectives on the Holocaust (pp. 19–32). Boston, MA: Kluwer-Nijhoff Publishing.

Berkowitz, L. (1999). Evil is more than banal: Situationism and the concept of evil. Personality and Social Psychology Review, 3 (3), 246–253.

Blass, T. (1991). Understanding behavior in the Milgram obedience experiment: The role of personality, situations, and their interactions. Journal of Personality and Social Psychology, 60 (3), 398–413.

Blass, T. (1993). Psychological perspectives on the perpetrators of the Holocaust: The role of situational pressures, personal dispositions, and their interactions. Holocaust and Genocide Studies, 7 (1), 30–50.

Blass, T. (2004). The man who shocked the world: The life and legacy of Stanley Milgram . New York: Basic Books.

Blass, T. (2012). A cross-cultural comparison of studies of obedience using the Milgram paradigm: A review. Social and Personality Psychology Compass, 6 (2), 196–205.

Bocchiaro, P., & Zimbardo, P. G. (2010). Defying unjust authority: An exploratory study. Current Psychology, 29 (2), 155–170.

Brannigan, A. (2013). Beyond the banality of evil: Criminology and genocide . Oxford: Oxford University Press.

Book   Google Scholar  

Brannigan, A., Nicholson, I., & Cherry, F. (2015). Introduction to the special issue: Unplugging the Milgram machine. Theory & Psychology, 25 (5), 551–563.

Brief, A. P., Buttram, R. T., Elliott, J. D., Reizenstein, R. M., & McCline, R. L. (1995). Releasing the beast: A study of compliance with orders to use race as a selection criterion. Journal of Social Issues, 51 (3), 177–193.

Browning, C. R. (1992). Ordinary men: Reserve Police Battalion 101 and the final solution in Poland . New York: HarperCollins.

Burger, J. M. (2009). Replicating Milgram: Would people still obey today? American Psychologist, 64 (1), 1–11.

Burger, J. M., Girgis, Z. M., & Manning, C. C. (2011). In their own words: Explaining obedience to authority through an examination of participants’ comments. Social Psychological and Personality Science, 2 (5), 460–466.

Burley, P. M., & McGuinness, J. (1977). Effects of social intelligence on the Milgram paradigm. Psychological Reports, 40, 767–770.

Charny, I. W. (1982). How can we commit the unthinkable? Genocide, the human cancer . Boulder, CO: Westview Press.

Costanzo, E. M. (1977). The effect of probable retaliation and sex related variables on obedience (Doctoral thesis). Retrieved from Dissertation Abstracts International (UMI No. 77-3253).

Darley, J. M. (1992). Social organization for the production of evil. Psychological Inquiry, 3 (2), 199–218.

Darley, J. M. (1995). Constructive and destructive obedience: A taxonomy of principal-agent relationships. Journal of Social Issues, 51 (3), 125–154.

De Swaan, A. (2015). The killing compartments: The mentality of mass murder . New Haven, CT: Yale University Press.

Dicks, H. V. (1972). Licensed mass murder: A socio-psychological study of some SS killers . London: Heinemman Educational for Sussex University.

Doliński, D., Grzyb, T., Folwarczny, M., Grzybała, P., Krzyszycha, K., Martynowska, K., et al. (2017). Would you deliver an electric shock in 2015? Obedience in the experimental paradigm developed by Stanley Milgram in the 50 years following the original studies. Social Psychological and Personality Science, 8 (8), 927–933.

Earl, H. C. (2009). The Nuremberg SS-Einsatzgruppen trial, 1945–1958: Atrocity, law, and history . Cambridge: Cambridge University Press.

Eckman, B. K. (1977). Stanley Milgram’s ‘obedience’ studies. Et cetera, 34 (1), 88–99.

Edwards, D. M., Franks, P., Friedgood, D., Lobban, G., & Mackay, H. C. G. (1969). An experiment on obedience . Unpublished Student Report, University of the Witwatersrand, Johannesburg, South Africa.

Elms, A. C. (1972). Social psychology and social relevance . Boston, MA: Little, Brown.

Elms, A. C. (1995). Obedience in retrospect. Journal of Social Issues, 51 (3), 21–31.

Etzioni, A. (1968). A model of significant research. International Journal of Psychiatry, 6 (4), 279–280.

PubMed   Google Scholar  

Fenigstein, A. (1998a). Reconceptualizing the obedience of the perpetrators. In D. G. Shilling (Ed.), Lessons and legacies, volume II: Teaching the Holocaust in a changing world (pp. 55–84). Evanston, IL: Northwestern University Press.

Fenigstein, A. (1998b). Were obedience pressures a factor in the Holocaust? Analyse & Kritik, 20 (1), 54–73.

Fenigstein, A. (2015). Milgram’s shock experiments and the Nazi perpetrators: A contrarian perspective on the role of obedience pressures during the Holocaust. Theory & Psychology, 25 (5), 581–598.

Foddy, W. H. (1971). Compliance to rational-legal authority (Monograph Series of Experimental Sociology Laboratory No. 3). Vancouver: Department of Anthropology and Sociology, University of British Columbia.

Fromm, E. (1973). The anatomy of human destructiveness . London: Jonathan Cape.

Geller, D. M. (1976). A role-playing simulation of obedience: Focus on involvement (Doctoral thesis). Retrieved from Dissertation Abstracts International (UMI No. 76-276).

Gibson, S. (2013a). Milgram’s obedience experiments: A rhetorical analysis. British Journal of Social Psychology, 52 (2), 290–309.

Gibson, S. (2013b). “The last possible resort”: A forgotten prod and the in situ standardization of Stanley Milgram’s voice-feedback condition. History of Psychology, 16 (3), 177–194.

Gibson, S. (2014). Discourse, defiance, and rationality: “Knowledge work” in the “obedience” experiments. Journal of Social Issues, 70 (3), 424–438.

Goldhagen, D. J. (1996). Hitler’s willing executioners: Ordinary Germans and the Holocaust . London: Alfred A. Knopf.

Gonen, J. Y. (2000). The roots of Nazi psychology: Hitler’s utopian barbarism . Lexington: The University of Kentucky.

Gupta, I. (1983). Obedience to authority amongst university students: An experimental analysis (Unpublished Doctoral thesis). University of Delhi, Delhi, India.

Hamilton, V. L., & Sanders, J. (1995). Crimes of obedience and conformity in the workplace: Surveys of Americans, Russians, and Japanese. Journal of Social Issues, 51 (3), 67–88.

Hamilton, V. L., & Sanders, J. (1999). The second face of evil: Wrongdoing in and by the corporation. Personality and Social Psychology Review, 3 (3), 222–233.

Harré, R. (1979). Social being: A theory for social psychology . Oxford, UK: Basil Blackwell.

Haslam, S. A., Reicher, S. D., & Birney, M. E. (2016). Questioning authority: New perspectives on Milgram’s ‘obedience’ research and its implications for intergroup relations. Current Opinion in Psychology, 11, 6–9.

Helm, C., & Morelli, M. (1979). Stanley Milgram and the obedience experiment: Authority, legitimacy, and human action. Political Theory, 7 (3), 321–345.

Helm, C., & Morelli, M. (1985). Obedience to authority in a laboratory setting: Generalizability and context dependency. Political Studies, 33 (4), 610–627.

Herrera, C. D. (2001). Ethics, deception, and ‘those Milgram experiments’. Journal of Applied Philosophy, 18 (3), 245–256.

Hilberg, R. (1961). The destruction of the European Jews (1–3 vols.). New York: Holmes & Meier.

Hilberg, R. (1980). The significance of the Holocaust. In H. Friedlander & S. Milton (Eds.), The Holocaust: Ideology, bureaucracy, and genocide (the San José Papers) (pp. 95–102). Millwood, NY: Kraus International Publications.

Hofling, C. K., Brotzman, E., Dalrymple, S., Graves, N., & Pierce, C. (1966). An experimental study of nurse-physician relations. Journal of Nervous and Mental Disease, 143 (2), 171–180.

Höss, R. (2001). Commandant of Auschwitz: The autobiography of Rudolf Hoess . London: Phoenix Press.

Humphreys, L. (1970). Tearoom trade: Impersonal sex in public places . Chicago: Aldine Publishing Co.

Kaposi, D. (2017). The resistance experiments: Morality, authority and obedience in Stanley Milgram’s account. Journal for the Theory of Social Behaviour, 47 (4), 382–401.

Kaufmann, H. (1967). The price of obedience and the price of knowledge. American Psychologist, 22 (4), 321–322.

Kelman, H. C. (1972). Human use of human subjects: The problem of deception in social psychological experiments. In A. G. Miller (Ed.), The social psychology of psychological research (pp. 163–178). New York: Free Press.

Kelman, H. C., & Hamilton, V. L. (1989). Crimes of obedience: Toward a social psychology of authority and responsibility . New Haven, CT: Yale University Press.

Kilham, W., & Mann, L. (1974). Level of destructive obedience as a function of transmitter and executant roles in the Milgram obedience paradigm. Journal of Personality and Social Psychology, 29 (5), 696–702.

Lawson, T. (2010). Debates on the Holocaust . Manchester: Manchester University Press.

Lippman, M. (1982). The trial of Adolf Eichmann and the protection of universal human rights under international law. Houston Journal of International Law, 5 (1), 1–34.

Luban, D., Strudler, A., & Wasserman, D. (1992). Moral responsibility in the age of bureaucracy. Michigan Law Review, 90 (8), 2348–2392.

Lutsky, N. (1995). When is ‘obedience’ obedience? Conceptual and historical commentary. Journal of Social Issues, 51 (3), 55–65.

Mandel, D. R. (1998). The obedience alibi: Milgram’s account of the Holocaust reconsidered. Analyse & Kritik, 20 (1), 74–94.

Mantell, D. M. (1971). The potential for violence in Germany. Journal of Social Issues, 27 (4), 101–112.

Mantell, D. M., & Panzarella, R. (1976). Obedience and responsibility. British Journal of Social and Clinical Psychology, 15 (3), 239–245.

Marcus, S. (1974). Book review of ‘obedience to authority’ by Stanley Milgram. The New York Times Book Review, 79 (2), 1–3.

Martin, J., Lobb, B., Chapman, G. C., & Spillane, R. (1976). Obedience under conditions demanding self-immolation. Human Relations, 29 (4), 345–356.

Mastroianni, G. R. (2002). Milgram and the Holocaust: A reexamination. Journal of Theoretical and Philosophical Psychology, 22 (2), 158–173.

McGaha, A. C., & Korn, J. H. (1995). The emergence of interest in the ethics of psychological research with humans. Ethics & Behavior, 5 (2), 147–159.

Meeus, W. H. J., & Raaijmakers, Q. A. W. (1986). Administrative obedience: Carrying out orders to use psychological-administrative violence. European Journal of Social Psychology, 16, 311–324.

Meeus, W. H. J., & Raaijmakers, Q. A. W. (1995). Obedience in modern society: The Utrecht studies. Journal of Social Issues, 51 (3), 155–175.

Meyer, P. (1970, February). If Hitler asked you to electrocute a stranger, would you? Probably. Esquire , 73, 128, 130, 132.

Miale, F. R., & Selzer, M. (1975). The Nuremberg mind: The psychology of the Nazi leaders . New York: Quadrangle.

Milgram, S. (1963). Behavioral study of obedience. Journal of Abnormal and Social Psychology, 67 (4), 371–378.

Milgram, S. (1964). Issues in the study of obedience: A reply to Baumrind. American Psychologist, 19 (11), 848–852.

Milgram, S. (1972). Interpreting obedience: Error and evidence. A reply to Orne and Holland. In A. G. Miller (Ed.), The social psychology of psychological research (pp. 138–154). New York: Free Press.

Milgram, S. (1974). Obedience to authority: An experimental view . New York: Harper & Row.

Milgram, S. (1977). Subject reaction: The neglected factor in the ethics of experimentation. Hastings Center Report, 7 (5), 19–23.

Miller, A. G. (1986). The obedience experiments: A case study of controversy in social science . New York: Praeger.

Miller, A. G. (1995). Constructions of the obedience experiments: A focus upon domains of relevance. Journal of Social Issues, 51 (3), 33–53.

Miller, A. G. (2004). What can the Milgram obedience experiments tell us about the Holocaust? Generalizing from the social psychology laboratory. In A. G. Miller (Ed.), The social psychology of good and evil (pp. 193–237). New York: Guilford Press.

Miller, A. G., Collins, B. E., & Brief, D. E. (1995). Perspectives on obedience to authority: The legacy of the Milgram experiments. Journal of Social Issues, 51 (3), 1–19.

Miranda, F. S. B., Caballero, R. B., Gomez, M. N. G., & Zamorano, M. A. M. (1981). Obediencia a la autoridad. Psiquis, 2, 212–221.

Mixon, D. (1972). Instead of deception. Journal of the Theory of Social Behavior, 2 (2), 145–177.

Mixon, D. (1976). Studying feignable behavior. Representative Research in Social Psychology, 7, 89–104.

Mixon, D. (1989). Obedience and civilization: Authorized crime and the normality of evil . London: Pluto Press.

Mook, D. G. (1983). In defense of external invalidity. American Psychologist, 38 (4), 379–387.

Morelli, M. F. (1983). Milgram’s dilemma of obedience. Metaphilosophy, 14 (3–4), 183–189.

Orne, M. T. (1962). On the social psychology of the psychology experiment: With particular reference to demand characteristics and their implications. American Psychologist, 17 (11), 776–783.

Orne, M. T., & Holland, C. C. (1968). On the ecological validity of laboratory deceptions. International Journal of Psychiatry, 6 (4), 282–293.

Parker, I. (2000). Obedience. Granta: The Magazine of New Writing, 71, 99–125.

Patten, S. C. (1977). Milgram’s shocking experiments. Philosophy, 52 (202), 425–440.

Penner, L. A., Hawkins, H. L., Dertke, M. C., Spector, P., & Stone, A. (1973). Obedience as a function of experimenter competence. Memory and Cognition, 1 (3), 241–245.

Perry, G. (2012). Beyond the shock machine: The untold story of the Milgram obedience experiments . Melbourne: Scribe.

Perry, G. (2015). Seeing is believing: The role of the film obedience in shaping perceptions of Milgram’s obedience to authority experiments. Theory & Psychology, 25 (5), 622–638.

Pigden, C. R., & Gillet, G. R. (1996). Milgram, method and morality. Journal of Applied Philosophy, 13 (3), 233–250.

Reicher, S., & Haslam, S. A. (2011). After shock? Towards a social identity explanation of the Milgram ‘obedience’ studies. British Journal of Social Psychology, 50 (1), 163–169.

Rosenbaum, M. (1983). Compliance. In M. Rosenbaum (Ed.), Compliant behavior: Beyond obedience to authority (pp. 25–49). New York: Human Sciences Press Inc.

Rosenhan, D. (1969). Some origins of concern for others. In P. Mussen, J. Langer, & M. Covington (Eds.), Trends and issues in developmental psychology (pp. 134–153). New York: Holt, Rinehart & Winston.

Roth, P. A. (2004). Hearts of darkness: ‘Perpetrator history’ and why there is no why. History of the Human Sciences, 17 (2–3), 211–251.

Russell, N. J. C. (2009). Stanley Milgram’s obedience to authority experiments: Towards an understanding of their relevance in explaining aspects of the Nazi Holocaust (Unpublished Doctoral thesis). Victoria University of Wellington, New Zealand.

Russell, N. J. C. (2014). Stanley Milgram’s obedience to authority “relationship” condition: Some methodological and theoretical implications. Social Sciences, 3 (2), 194–214.

Russell, N. J. C., & Gregory, R. J. (2005). Making the undoable doable: Milgram, the Holocaust and modern government. American Review of Public Administration, 35 (4), 327–349.

Russell, N. J. C., & Gregory, R. J. (2011). Spinning an organizational “web of obligation”? Moral choice in Stanley Milgram’s “obedience” experiments. The American Review of Public Administration, 41 (5), 495–518.

Russell, N. J. C., & Gregory, R. J. (2015). The Milgram-Holocaust linkage: Challenging the present consensus. State Crime Journal, 4 (2), 128–153.

Sabini, J., & Silver, M. (1982). Moralities of everyday life . New York: Oxford University Press.

Sabini, J., Siepmann, M., & Stein, J. (2001a). The really fundamental attribution error in social psychological research. Psychological Inquiry, 12 (1), 1–15.

Sabini, J., Siepmann, M., & Stein, J. (2001b). Authors’ response to commentaries. Psychological Inquiry, 12 (1), 41–48.

Saltzman, A. L. (2000). The role of the obedience experiments in Holocaust studies: The case of renewed visibility. In T. Blass (Ed.), Obedience to authority: Current perspectives on the Milgram paradigm (pp. 125–143). Mahwah, NJ: Lawrence Erlbaum Associates.

Schlenker, B. R., & Forsyth, D. R. (1977). On the ethics of psychological research. Journal of Experimental Social Psychology, 13 (4), 369–396.

Schuler, H. (1982). Ethical problems in psychological research . New York: Academic Press.

Schurz, G. (1985). Experimentelle Uberprufung des Zusammenhangs zwischen Persönlichkeitsmerkmalen und der Bereitschaft zum destruktiven Gehorsam gegenuber Autoritäten. Zeitschrift für Experimentelle und Angewandte Psychologie, 32, 160–177.

Sereny, G. (2000). The German trauma: Experiences and reflections 1938–2000 . London: Allen Lane.

Shanab, M. E., & Yahya, K. A. (1977). A behavioral study of obedience in children. Journal of Personality and Social Psychology, 35 (7), 530–536.

Shanab, M. E., & Yahya, K. A. (1978). A cross-cultural study of obedience. Bulletin of the Psychonomic Society, 11 (4), 267–269.

Shelton, G. A. (1982). The generalization of understanding to behaviour: The role of perspective in enlightenment (Unpublished Doctoral thesis). University of British Columbia, Vancouver.

Sheridan, C. L., & King, R. G. (1972). Obedience to authority with an authentic victim. Proceedings of the American Psychological Association (80th Annual Convention) , 7 , 165–166.

Slater, L. (2004). Opening Skinner’s box: Great psychological experiments of the twentieth century . New York: W. W. Norton.

Staub, E. (1989). The roots of evil: The origins of genocide and other group violence . Cambridge: Cambridge University Press.

Steiner, J. M. (1980). The SS yesterday and today: A sociopsychological view. In J. E. Dimsdale (Ed.), Survivors, victims, and perpetrators: Essays on the Nazi Holocaust . New York: Hemisphere Publications.

Tedeschi, J. T., Lindskold, S., & Rosenfeld, P. (1985). Introduction to social psychology . St. Paul, MN: West Publishing Company.

Waller, J. (2002). Becoming evil: How ordinary people commit genocide and mass killing (2nd ed.). New York: Oxford University Press.

Zangwill, N. (2003). Perpetrator motivation: Some reflections on the Browning/Goldhagen debate. In E. Garrard & G. Scarre (Eds.), Moral philosophy and the Holocaust (pp. 89–102). Burlington, VT: Ashgate.

Download references

Author information

Authors and affiliations.

University of Calgary, Calgary, AB, Canada

Nestar Russell

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Nestar Russell .

Rights and permissions

Reprints and permissions

Copyright information

© 2018 The Author(s)

About this chapter

Russell, N. (2018). Academia’s Response to Milgram’s Findings and Explanation. In: Understanding Willing Participants, Volume 1. Palgrave Macmillan, Cham. https://doi.org/10.1007/978-3-319-95816-3_6

Download citation

DOI : https://doi.org/10.1007/978-3-319-95816-3_6

Published : 18 September 2018

Publisher Name : Palgrave Macmillan, Cham

Print ISBN : 978-3-319-95815-6

Online ISBN : 978-3-319-95816-3

eBook Packages : Behavioral Science and Psychology Behavioral Science and Psychology (R0)

Share this chapter

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Publish with us

Policies and ethics

  • Find a journal
  • Track your research

Psychological research, obedience and ethics

stanley milgram obedience experiment ethical issues

Introduction

One of the best known studies in the history of psychology is the research on obedience carried out by Stanley Milgram in the 1960s. In his research Milgram demonstrated the lengths to which people are willing to go just because someone in authority tells them to do something. The studies Milgram conducted also raised the issue of ethics in research, as some critics argued that he failed to take sufficient precautions to protect the integrity and wellbeing of his participants. At the same time, more than any other study in psychology, the findings of Milgram’s research demonstrate why ethics are important.

As well as reading about Milgram's work and ethics, you will engage in an online activity to learn about the code of ethics concerning the psychological research that is conducted with human participants. You will also gain an understanding of the guidelines that govern the use of non-human animals in psychological research in a second online activity, and why psychologists conduct such research. You will have the opportunity of viewing two short films which will introduce you to the research of Alex Thornton who studies meerkats, and of Tetsuro Matsuzawa who works with chimpanzees.

This OpenLearn course is an adapted extract from the Open University course DE100 Investigating psychology 1 .

Learning outcomes

After studying this course, you should be able to:

describe the research of Stanley Milgram on obedience

recognise the main ethics principles governing psychological research

understand the ethics issues concerning research involving non-human animals

appreciate the value of conducting research with animals.

1 Milgram’s obedience study

Milgram was one of the most innovative and productive social psychologists of his generation, who undertook a variety of studies that explored social psychological aspects of everyday life. However, he is largely remembered for one dramatic piece of work – the obedience studies. The best way to get inside this study is to imagine that you are one of the participants taking part in Milgram’s experiment. So read on with that in mind.

1.1 The set-up

It’s 1961 and you are arriving at the doors of the Psychology Department of the prestigious Yale University in the USA. The reason you are here is that you replied to an advert in the local paper asking for volunteers to take part in a study on memory. The advert (see Figure 1) offered a fee plus expenses and said that you would be paid on arrival at the laboratory.

Described image

This is an image of the advertisement Milgram placed in a local paper to recruit members of the public to participate in his study. The only information provided about the study is that it involves a scientific investigation of memory and learning. The advert highlighted that participants would be paid four dollars for one hour of their time, and also that they don't need any special training, education or experience. Wanted were five hundred New Haven men: factory workers, city employees, labourers, barbers, businessmen, clerks, profession people, telephone workers, construction workers, salespeople, white-collar workers and others, aged between twenty and fifty. The announcement also specifically states that high school and college students could not be used.

As you walk through the doors you are met by a serious-looking man in a laboratory coat who turns out to be the experimenter. He introduces you to a genial middle-aged man who is described as a fellow volunteer. The experimenter explains that the study will involve one of the volunteers taking on the role of a ‘teacher’ and the other taking on the role of a ‘learner’. As part of the experiment, the ‘teacher’ will engage the ‘learner’ in a simple memory task. The ‘learner’ and the ‘teacher’ will be in different rooms and will communicate through microphones (see Figure 2). The experimenter reveals that the study is designed to investigate the effect of punishment on learning. The ‘teacher’ will be asked to administer an electric shock to the ‘learner’ every time the latter makes an incorrect response on the memory task.

Described image

This shows a diagram of the experimental set-up. The experimenter is shown as seated in the same room, but behind the 'teacher'. The 'learner' is shown as seated in an adjacent room on his own. The wiring attached to the 'learner' passes through the wall and into the apparatus, the shock generator, placed on a table in front of the participant.

To select who will be the ‘teacher’ and who will be the ‘learner’, you draw slips of paper. You pick out the ‘teacher’ slip. You then watch as the ‘learner’ is strapped into a chair, and you hear the experimenter tell him that ‘although the shocks can be extremely painful, they cause no permanent tissue damage’. The experimenter now gives you a sample shock of 45 volts to show you what the ‘learner’ will experience during the study. The shock is unpleasant, but short of being painful.

The experimenter then takes you into the adjacent room and sits you down in front of an impressive-looking apparatus that will be used to administer the shocks (see Figure 3).

Described image

This figure shows the apparatus - the shock generator - used in the study which is described next in the text.

The shock generator consists of a row of switches that run in 15 volt increments from 15 volts through to 450 volts. Under the label for each switch are some descriptive words, such as ‘slight shock’ (15 volts), ‘moderate shock’ (75 volts), ‘strong shock’ (135 volts), ‘very strong shock’ (195 volts), ‘intense shock’ (225 volts), ‘extremely intense shock’ (315 volts), ‘danger: severe shock’ (375 volts) and finally ‘XXX’ (435 volts). Suddenly, this looks quite serious and you probably hope that you don’t have to go very far up the scale. This is especially so given that you received the 45-volt shock, and you know that this was unpleasant enough. The last switch on the shock generator administers an electric impulse ten times as strong!

In the first phase of the experiment, the experimenter asks you, the ‘teacher’, to read a series of word pairs to the ‘learner’ who is expected to memorise them (for instance, ‘green-grass’ , ‘blue-sky’, ‘nice-day’). In the second phase, the test phase, you are asked to read out the first word of the pairs (e.g. ‘green’), followed by four possible responses (‘grass, hat, ink, apple’). If the ‘learner’ identifies the paired word correctly, you are to move on to the next word pair on the list. If the answer is wrong you have to tell the ‘learner’ the correct answer, indicate the level of punishment you are going to give them (starting with 15 volts), and flick the appropriate switch on the shock generator. For every subsequent incorrect answer, you are told to move one switch up the scale of shocks.

The experiment starts. To begin with everything is fine and the ‘learner’ gets most of the answers right. You have only used the shock generator a couple of times, and at this stage the shocks are mild. Then the ‘learner’ starts to get the answers wrong and you are moving up the shock scale into the ‘strong shock’ range. Although you cannot see the ‘learner’ you can hear him and as the shocks increase he starts to shout out. You have heard him grunt at the low voltage but now he is starting to ask to be let out. At 120 volts you hear him shout out in an agitated tone, complaining that he is in pain, and at 150 volts he asks to be released.

Suddenly, you feel uncomfortable and you decide to stop. The experimenter, the man in the grey coat, objects and asks you to carry on, in spite of the ‘learner’s’ protestations.

What do you think you would do in this situation? At what point would you stop? 200 volts? 150 volts? Would you respond to the cries of your fellow volunteer or would you complete the job you agreed to do and carry out the instructions of the experimenter?

How many people do you think would continue to follow the orders? At what point do you think people would stop?

Before Milgram carried out the study, he posed the same questions as in Question 1 to different groups of people, including ordinary members of the public, college students, psychologists and psychiatrists. He asked them to speculate on how far they thought most people would go if asked to administer shocks. Most ordinary people said that participants would generally refuse to administer shock, or at least not go very far beyond the point where the ‘learner’ experienced pain. Also, most said that participants should rebel, and that they should not continue beyond around 150 volts. Among the professional groups, there was widespread agreement that nobody taking part in the study go all the way.

You will be relieved to know that in the actual study carried out by Milgram, no person was hurt during the procedure, and the only actual shock administered was the 45-volt ‘tester’ given to the ‘teacher’. In fact, the whole situation was staged. The role of ‘experimenter’ was played by a 31-year-old biology teacher. The ‘learner’, presented as a ‘fellow volunteer’, was in on the deception and was merely playing the part. In reality, he was a 47-year-old accountant, who was chosen for the role because he appeared mild-mannered and likeable. He was not the sort of person one would want to see hurt. The drawing of slips of paper was fixed to ensure that the ‘naive participant’ was always cast in the role of the ‘teacher’, and the ‘shock generator’ was simply a simulator. The sounds (the moans and cries) that the participants heard were a recording played from the adjacent room. Importantly, however, the deception was so good that participants believed that they were actually administering shocks. So the study presented an ingenious way of discovering how far people would be willing to go, just because a psychological experiment on ‘the effects of punishment on learning’ demanded it. Most people like to think that they (and people around them) would not go very far. But what happened when Milgram actually placed people in that position?

1.2 The results

In the first instance Milgram conducted the study on a sample of forty participants, all of them male. Each played the role of ‘teacher’ in the situation described in Section 1.1. Each participant went through the identical experimental procedure: all forty heard the same instructions, encountered the same ‘experimenter’ and ‘learner’, heard identical (pre-recorded) cries from the next room. The ‘experimenter’ in the grey lab coat offered the same words of encouragement. The sessions were filmed (Figure 4) and notes were taken by observers looking through an observation mirror.

Described image

This figure comprises three photos taken during the Milgram study. The first is of the experimenter instructing the participant on how to use the shock generator. The second is of the participant getting seated at the table in front of the shock generator. The third is of the ‘learner’ being strapped to the apparatus.

Milgram found that, of the forty participants who took part in the study, all obeyed up to 300 volts, the twentieth switch on the shock generator. This is the point at which the ‘learner’ was heard screaming: ‘I absolutely refuse to answer any more. Get me out of here. You can’t hold me here. Get me out. Get me out of here.’ However, only five of the forty participants refused to continue beyond this point. Four gave only one more shock before breaking off, with an additional five stopping between 315 volts and 435 volts. But as many as twenty-six continued to the end of the scale and administered the maximum 450 volts. This is despite the fact that, at 330 volts, they had already heard intense and prolonged screaming: ‘Let me out of here. Let me out of here. … Let me out of here. You have no right to hold me here. Let me out! Let me out!’ Shocks beyond 330 volts were accompanied by eerie silence. Nevertheless, twenty-six ordinary members of the public from Connecticut administered the maximum shock and continued to do so until the experimenter called a halt to the proceedings.

As well as counting the number of participants who went all the way on the shock generator, Milgram also observed their reactions. Participants who took part in the study generally displayed signs of nervousness and tension. Many were visibly uncomfortable and probably would not have continued had they not heard the experimenter say things like ‘Please continue’, ‘Please carry on’, ‘It is absolutely essential that you continue’ or ‘You have no choice; you must go on’. At the end of the study, many of the obedient participants heaved sighs of relief or shook their heads in apparent regret. Some even had laughing fits during the experiment, probably brought on by anxiety. Milgram (1963, p. 375) wrote that ‘full-blown, uncontrollable seizures were observed for 3 subjects. On one occasion we observed a seizure so violently convulsive that it was necessary to call a halt to the experiment’. (You may have noticed that in this quote Milgram refers to people who took part in his study as ‘subjects’. This was common practice in psychology in the 1960s. Today the word ‘participant’ is used instead as the word ‘subject’ is considered demeaning, and lacking in respect towards volunteers on whose participation much of psychological research ultimately depends.)

Do Milgram’s findings seem plausible to you? Ordinary members of the public were prepared to administer electric shocks to another person on the mere (albeit persistent) request of a man in a laboratory coat. They did so despite the protests from the ‘victim’ and continued even after the supposed recipient of the shocks went quiet. Before the study, when Milgram asked his fellow professionals to predict how many participants would refuse to go all the way, they said that all of them would do so. In reality only 35 per cent did. In Milgram’s study, the average voltage at which participants stopped shocking the ‘learner’ was 368 volts. Members of the public predicted that people would stop at around 140 volts. This is a remarkable discrepancy. It is therefore not surprising that Milgram’s research went on to provoke considerable debate.

Box 1 Why do it this way?

Milgram’s obedience work is remarkable, not only because of the important questions it sought to explore, but also because it is a fine example of good experimental procedure in social psychology.

The most important feature of any laboratory experiment is its controlled nature. Note that every person who took part in Milgram’s research underwent an identical experience. All participants received the same instructions, encountered the same individuals (the ‘experimenter’ and the fellow ‘volunteer’) and heard identical cries and protestations from the ‘learner’. To ensure consistency in the experimental procedure, Milgram even recorded the anguished cries in advance, and played them to participants from a tape.

This equivalence of experience across the forty participants was essential if meaningful comparisons were to be made. It ensured that any difference in behaviour observed in the study could not be attributed, for instance, to the fact that some participants heard louder or more desperate cries than others. For similar reasons, Milgram used the same ‘learner’ and ‘experimenter’ with each participant. He wanted to ensure that none of the results could be accounted for by differences in the personality or the demeanour of the confederates.

Another interesting aspect of Milgram’s research is that he recruited participants from the general public, using a newspaper advert. At the time (and still now in many psychology departments) participants tended to be recruited mainly from among the student population. However, Milgram was interested in exploring the level of obedience to scientific authority among people with no direct link to the university or research environment, so he recruited from the general public.

Finally, in Milgram’s original study, all forty participants were male. Why do you think this was the case? This was not because Milgram wanted to exclude women from his research. He later conducted further studies in which he explored gender differences in obedience. In the initial study, however, he decided to control for the potential effects of gender on the findings by limiting the sample to men.

1.3 The variations

The findings of Milgram’s original study highlighted the phenomenon of obedience, but it could not reveal what it is about the situation that made participants administer potentially lethal shocks to a fellow human being. To address this question Milgram carried out further research in which he introduced subtle variations to the original procedure. By examining the effects of these variations on levels of obedience, he was able to isolate specific aspects of the situation that might influence whether participants obey or not.

By the time Milgram completed his research in 1962 he had processed 800 people through nineteen variations of the original design. For instance, in one variation, Milgram introduced into the proceedings a dialogue about a heart attack. He wanted to see whether alerting the participants to the impact of the shocks on the ‘learner’s’ health might reduce obedience. Note that all other aspects of the original study were preserved. Interestingly, the conversation about the heart attack made no real difference. Twenty six out of the forty participants still continued to 450 volts, although those who stopped did so at a lower voltage with five stopping as soon as the ‘learner’ asked to be let out. So, the reference to the heart attack made those who disobeyed do so earlier, but it did not prevent the more obedient participants from going all the way.

Milgram also varied the proximity of the ‘learner’ and ‘teacher’. In one variation he put them in the same room, while in another he required the ‘teacher’ to hold the ‘learner’s’ arm down on a plate to receive the electric shock. This manipulation had a clear effect. Milgram found that the closer you place the ‘teacher’ to the ‘learner’, the fewer shocks the ‘teacher’ is likely to administer. Equally, the further you place the ‘learner’ away from the ‘teacher’, the less the impact their pleas are likely to have.

Equally crucial was the presence of the authority figure. In one variation, the ‘experimenter’ in the grey coat pretended to have to leave the experiment owing to some emergency and was replaced by a person in plain clothes, who was not a scientist. Only 20 per cent of participants went all the way and gave the ‘learner’ 450-volt shocks. Similar results were obtained when orders were given by phone. The physical presence of an authority figure was therefore crucial.

In another variation Milgram placed two ‘experimenters’ in the room. One told the participants to continue (as in the original study), while the other told them to stop. In this variation, all the participants stopped giving the shocks very early on. This showed that an absence of a clear authority figure reduces obedience.

Milgram also conducted a version of the experiment in which he placed a second ‘teacher’ in the room, although this one was a stooge instructed to obey until the end. In this variation all the participants went along with the confederate and shocked up to 450 volts! So the mere presence of another obedient ‘volunteer’ made all the participants go all the way.

One of the main conclusions of Milgram’s work was that under certain conditions involving the presence of authority, people suspend their capacity to make informed moral judgments and defer responsibility for their actions to those in authority. When people are in this particular frame of mind, the nature of the task that they are asked to perform becomes largely irrelevant, and the main determinant of their actions is the commands of the authority figure.

1.4 Summary

  • Milgram found that most people would administer potentially lethal levels of shock to another human being, just because they were told to do so by an authority figure.
  • The use of a controlled experimental procedure enabled Milgram to explore different aspects of the situation that influence the extent to which people will obey authority.
  • Two key factors in obedience are the presence of a clear authority figure, and the distance between the person administering the shock and the ‘victim’.

2 Milgram’s study and ethics

At the beginning of Section 1 you were asked to put yourself in the shoes of one of the participants in Milgram’s research. How do you think being a participant in the study felt? As you already read, many of the participants were visibly uncomfortable during the procedure. This is one of the reasons why the study created a storm, starting with a hostile review of the research in a newspaper, the St . Louis Post-Dispatch . The newspaper criticised Milgram and Yale University for putting the participants in such a stressful situation. It claimed that Milgram violated the rules of ethics which guide psychological research. The charge was repeated in academic circles, and led to Milgram’s application to join the American Psychological Association being put on hold for a year. Milgram made a robust rebuttal of the charges and the debate about the issues led to the introduction of new codes of good practice for psychologists.

Before we look at the arguments that swirled around the obedience study we need to consider what we mean by ethics . It all starts with morals , which are rules to guide our behaviour. These rules are based on a number of socially agreed principles which are used to develop clear and logical guidelines to direct behaviour. They also contain ideas about what is good and desirable in human behaviour. Ethics , in the context of psychological research, refers to a moral framework that governs what psychologists can and cannot do.

The first generally accepted code of ethics for research on humans was devised in 1947 as a response to the very events that provoked Milgram’s research. During the Second World War (1939–45), under the Nazi regime, research was carried out on human beings that led to many deaths, deformities and long-term injuries. Revelations about this research were as great a shock for the post-war world as the death camps, because these acts of brutality and murder were conducted by doctors and scientists.

After the war the victors held a series of trials, in the German city of Nuremberg, of people who had taken part in the worst excesses of the horrors that had swept across Europe. Among them were twenty-three doctors involved in the brutal experiments. Sixteen of them were found guilty, of whom seven were sentenced to death. Significantly, the judgement included a statement about how scientists should behave when experimenting on other humans. This is referred to as the Nuremberg Code (see Table 1) and it became the basis for future ethical codes in medicine and psychology.

Table 1 The Nuremberg Code (1946)
1 The voluntary consent of the human subject is absolutely essential
2 The experiment should yield fruitful results for the good of society, that cannot be obtained by other means
3 The experiment should be based on previous research so that the anticipated results can justify the research
4 All unnecessary physical and mental suffering should be avoided
5 No experiment should be conducted where there is reason to believe that death or disabling injury may be the result
6 The degree of risk should also be less than the potential humanitarian importance of the research
7 Adequate precautions should be in place to protect the subjects against any possible injury
8 Experiments should only be conducted by qualified persons
9 The human subject should always be at liberty to end the experiment
10 The scientist in charge should be prepared to terminate any experiment if there is probable cause to believe that continuation is likely to result in injury or death

Four key principles emerged from the Nuremberg Code. First, participants must be able to give informed consent to the procedure. Second, they must retain the right to withdraw from the study whenever they want. Third, the welfare of the participant must be protected wherever possible. The fourth principle is the most difficult to interpret because it concerns the costs and benefits of the study. It says that any risks to the participants must be greatly outweighed by the possible benefits for the greater good.

2.2 The case against Milgram

Before you go on to read about the criticism of Milgram’s obedience studies, try to think through all the issues relating to ethics that are raised by this work.

In what way were the participants deceived, or harmed? Did they have the right to withdraw? Do you think that in Milgram’s case the ends justify the means? Do the benefits of the study justify the costs? Do you think that the results of the study are worth the pain and discomfort caused to the participants?

Among those who were highly critical of Milgram’s study was fellow psychologist Diana Baumrind. She started her critique by noting the dilemma that all research psychologists face: ‘Certain problems in psychological research require the experimenter to balance his career and scientific interests against the interests of his prospective subjects’ (Baumrind, 1964, p. 421).

Baumrind challenged Milgram on whether he had properly protected the welfare of the participants . She used direct quotes from Milgram’s original report to illustrate the lack of regard she said was shown to the participants. In particular, she noted the detached manner in which Milgram described the emotional turmoil experienced by the volunteers. For example:

In a large number of cases the degree of tension [in the participants] reached extremes that are rarely seen in sociopsychological laboratory studies. Subjects were observed to sweat, tremble, stutter, bite their lips, groan, and dig their fingernails into their flesh. These were characteristic rather than exceptional responses to the experiment.

In Baumrind’s view, and in the view of numerous others, the levels of anxiety experienced by participants were enough to warrant halting the experiment. What is more, just because someone volunteers to take part in the study (i.e. gives informed consent at the start of the study), it does not mean that the researcher no longer has responsibilities towards them and their wellbeing. On the principle of cost – benefit , Baumrind challenged the view that the scientific worth of the study balanced out the distress caused to the participants. She acknowledged that some harm to participants might be a necessary part of some research – for example, when testing out new medical procedures – as in those cases results cannot be achieved in any other way. Social psychology, however, is not in the same game as medicine and is unlikely to produce life-saving results. The strength of the conclusions does not, therefore, justify harming participants. Milgram related his study to the behaviour of people who worked in the Nazi death camps and suggested that his study illuminated the way that ordinary people living ordinary lives are capable of playing a part in destructive and cruel acts. Baumrind dismissed this justification for the study and suggested there are few, if any, parallels between the behaviour in the study and the behaviour in the death camps.

Baumrind went on to make a further criticism by considering the effect of this work on the public image of psychology, and suggested that it would be damaged because the general public would judge that the participants were not protected or respected.

A further potential problem with Milgram’s experiment concerns the participants’ right to withdraw . Do you think that this principle, embedded in the Nuremberg Code, was sufficiently observed in Milgram’s research? Recall that one of the key aspects of the experimental procedure was that whenever a participant demonstrated a reluctance to carry on with administering the shocks, they were told by the ‘experimenter’ in the grey coat ‘you must go on’, or ‘you have no choice; you must go on’. It might be argued that telling a participant that they ‘have no choice’ but to continue with the experiment contravenes the right to withdraw, which is enshrined in the ethics code. To be fair, fourteen of the forty participants in the original study did withdraw, in spite of being told that they had no choice, so it could be argued that, ultimately, the participants did have a choice. It is just that making that choice was made more difficult by the presence of the ‘experimenter’ and by his prods. After all, the study was about obedience, and the instructions from the ‘experimenter’ were essential to the investigation. Exercising or not exercising the right to withdraw is what the study was about.

2.3 The case for the defence

Milgram made a series of robust defences for the study, starting with a response to the newspaper article that first raised concerns. He dismissed the accusation that participants were severely traumatised by the experience. He argued that ‘relatively few subjects experienced greater tension than a nail-biting patron at a good Hitchcock thriller’ (quoted in Blass, 2007). This was rather disingenuous, given his other descriptions of their reactions (see above). However, Milgram made a more measured response to the academic arguments. He pointed out, for instance, that he could not have known the outcome of the research before he started. As you already read, before embarking on the study he asked fellow professionals how they expected people to behave, and they predicted that participants would not continue to obey and administer severe shocks to the ‘learner’.

More importantly, Milgram was not oblivious to the psychological needs of his participants and was aware of the potential harm caused by the study. Immediately after the study, its true purpose was revealed to the participants. They were interviewed and given questionnaires to check they were all right. A friendly reconciliation was also arranged with the ‘victim’ whom they thought they had shocked. This procedure, known as debriefing , is commonplace today, but this was not the case in the 1960s. So, in this respect at least, Milgram was ahead of the game in terms of ethics procedures (Blass, 2004).

Milgram also conducted a follow-up survey of the participants one year after the study, to ensure that there was no long-term harm (Colman, 1987). The results showed that 84 per cent said they were ‘glad to have been in the experiment’, and only 1.3 per cent said they were very sorry to have taken part. Milgram also described how the participants had been examined by a psychiatrist who was unable to find a single participant who showed signs of long-term harm. Morris Braverman, a 39-year-old social worker, was one of the participants in Milgram’s experiment who continued to give shocks until the maximum was reached. He claimed, when interviewed a year after the experiment, that he had learned something of personal importance as a result of being in the experiment. His wife said, with reference to his willingness to obey orders, ‘You can call yourself an Eichmann’ (Milgram, 1974, p. 54).

Milgram’s basic defence was that the harm to the participants was not as great as it might appear, and for some of them the change in their understanding of their own behaviour and the behaviour of others was a positive event. He makes a further defence that we have to treat all people with respect and that this involves allowing them to make choices even if those choices are not always for the best. In direct response to Baumrind’s criticisms he wrote:

I started with the belief that every person who came to the laboratory was free to accept or to reject the dictates of authority. This view sustains a conception of human dignity insofar as it sees in each man a capacity for choosing his own behavior.

2.4 The judgement

So what do you think should be the final judgement on the ethics of Milgram’s study? As you can see from the debate between Milgram and Baumrind, ethics is something that psychologists debate and often disagree on. Ethics principles, like all rules, are subject to interpretation and disagreement.

And yet, while individuals might have their personal view about whether a piece of research is ethical or not, what really matters is the judgment of institutions that regulate the profession. In the USA the regulatory body is the American Psychological Association. Its equivalent in the UK is the British Psychological Society. These institutions have ethics committees which issue guidelines and codes of conduct related to ethics in research and can reprimand researchers who can be shown to have violated the rules. At the time of Milgram’s study, his research was investigated by the ethics committee of the American Psychological Association, who eventually came to the conclusion that it was ethically acceptable. Notably, however, Milgram’s studies could not be carried out today, as the ethics guidelines have become more restrictive since the 1960s.

Finally, one further issue regarding Milgram’s study is worth pointing out. Although the ethics of Milgram’s research have been questioned, it could be argued that the obedience study, more than any other study in psychology, demonstrated why ethics are important. Recall that what Milgram’s study showed was that ordinary people were willing to harm another human being just because they were told to do so by a person they believed was a psychologist, and because doing so was supposedly ‘required by the experiment’. This shows that people generally are ready to give scientists the benefit of the doubt and go along with what they are doing, even when it involves harming individuals. This in itself illustrates how important it is to have some moderation of scientific activity, and have limits imposed on what scientists can and cannot do.

2.5 Summary

  • Psychologists have a duty of care towards participants and must ensure that their wellbeing is preserved throughout a study.
  • Participants must be asked to give informed consent before taking part in research and have a right to withdraw at any point.
  • Milgram’s obedience studies kick-started an ethics debate in psychology and highlighted the need for the development of more stringent guidelines for the conduct of research psychologists.
  • Although Milgram’s obedience study was judged to be ethical at the time of publication, it would be in violation of the strict ethics guidelines in place today.

Activity 1: Ethics in psychological research

This activity introduces you to the idea of ethics in psychological research. When conducting research, psychologists cannot do what they like – they must abide by a code of conduct. Above all, the aim of this is to protect the welfare of the people participating in the research. It is important that you grasp the main ethics principles and their importance, as you will be returning to them throughout this course. At the end of this activity, you will find a handout summarising the main teaching points. You should download this and add it to your files.

Research ethics

When a psychologist is working out how to undertake a particular study, it is important that they consider whether what they are doing and how they are doing it is going to be ethical. All universities and research organisations have panels that judge whether research is ethical or not, and in the UK psychological research also needs to meet the requirements of the British Psychological Society (BPS), who specify a code of ethics and conduct, which includes:

  • Research should not include risks to the psychological wellbeing, physical health, personal values or dignity of participants.
  • Participants should give informed consent before taking part in research.
  • Participants should be able to stop participating in the research at any point.

The principles cover a number of other very important points, such as confidentiality, debriefing and protection. This activity is going to focus on the three principles above and you are going to have a go at applying them to a specific research project.

Task: Is it ethical?

On the following pages you will find a brief description of a psychological research study. Your task is to imagine you are on a research ethics panel that has been asked to consider the research being proposed. In each case, read the description and decide whether you think the study described is ethical in terms of the three principles described in the introduction:

  • Research should not threaten the psychological wellbeing, health, values or dignity of participants.

The ethics panel received the following proposal:

stanley milgram obedience experiment ethical issues

This study is designed to discover what effect peer pressure might have on people’s voting behaviour. Phase one of the study will involve adding a question stating ‘which party did you vote for at the last election?’ to the end of an in-class exam being taken by twenty undergraduate students. Phase two will take place a week later in a scheduled seminar, and consist of asking the same students, one after another, to tell the group as a whole whom they had voted for.

Think about the three ethics principles and how you might expect a psychological study to meet each one, and then decide whether you think this study meets each ethics principle.

Ethics principle 1 - study 1.

Wellbeing, health, values and dignity

The correct answer is b.

This study threatened both the values and the dignity of the participants. UK law protects the right for voting to be secret for a good reason, so asking students how they voted in an exam and also to state this publicly contravenes the values attached to keeping how you voted a confidential matter. Putting students in a position where they might feel obliged either to state how they voted or indeed to lie about this as a result of peer pressure is also likely to have a negative effect on their dignity.

Ethics principle 2 - Study 1

Informed consent

At no point were the students told they were taking part in a research study, so they did not give their consent to participate.

Ethics principle 3 - Study 3

Right to withdraw

At no point were the students told they were taking part in a research study, so they did not realise there was any research to withdraw from.

Would you give this study ethical approval?

To be approved a study would need to meet all criteria. If it fails to meet just one, it should not gain approval.

As is stated in this course, applying ethical principles is never that straightforward and there are often cases where people have differing opinions. Don't worry if you had different answers; instead concentrate on which aspects of the study we have linked to each of the three ethical principles.

The ethics panel that considered Study 1 did not give approval. Following their feedback the psychologist revised the design of the study and resubmitted it.

stanley milgram obedience experiment ethical issues

The revised study is to involve recruiting participants through a poster that will ask for volunteers to take part in a study on political communication. Volunteers will be informed that they have the right to withdraw from the study at any point and will be first asked to complete a ‘consent to participate’ form and then a questionnaire asking them about their background, likes and dislikes, and also whom they voted for in the last election. They will then attend a session that evening where the study will take place. The session will involve watching a party political broadcast and then answering questions in a group on how well the speaker communicated their ideas. At the end of the broadcast the researcher will apologise to the participants saying their questionnaires have all been accidentally lost, and ask them to state in front of the whole group of participants whom they voted for in the last election. By comparing how the participants said they had voted in the questionnaire and in front of the group, it will be possible to see if any had changed their mind as a result of being part of a group.

Ethics principle 1 - Study 2

The revised study still involves the participants’ being required to say how they voted publicly in front of a group of people; that they are still being asked the question at all can be seen as not respecting their values, and they are also still being put in a position where they might feel obliged to state how they voted, or indeed to lie about this as a result of peer pressure, and this is likely to have a negative effect on their dignity.

Ethics principle 2 - Study 2

Although the study now asks for volunteers and asks them to complete a consent form, this cannot be considered ‘informed’ consent because the researcher has not told them what the study is really about. Instead, the participants are being deceived about the real purpose of the study. A degree of deception is sometimes necessary in some forms of psychological research, but the degree of deception involved here is unethical as the participants are not being told in advance that they will be asked publicly about a confidential matter (voting behaviour).

Ethics principle 3 - Study 2

The correct answer is a.

Participants in this study would be aware that they were taking part in research, and were also told explicitly that they could withdraw from the study at any point. One view could be that as they did not know the true nature of the study, they were not able to withdraw from it, but this issue is dealt with under ‘informed consent’. Note that a study should not be given approval if it fails to meet any of the ethical principles.

To be approved a study would need to meet all criteria. If it fails to meet just one, it should not receive approval.

The ethics panel were still not convinced and did not give approval to Study 2 either. Undeterred, the researcher revised the design of the study and resubmitted it yet again.

stanley milgram obedience experiment ethical issues

The new study will involve recruiting participants through a poster placed on a college noticeboard, which will ask for volunteers to take part in a study on ‘social pressures and voting behaviour’. Volunteers will be provided with an accurate summary of the proposed study which will also explain that they have the right to withdraw at any point, and should feel free to do so. After having read the summary they will be asked to sign a consent form. The study will involve providing participants with transcripts of fictitious election speeches from three candidates for Student Union President of a fictitious college. The participants will read the speeches and indicate on a form whom they would vote for. Following this, they will be told that an overwhelming number of students are voting for one of the other candidates and again asked to indicate on a form whom they would vote for.

Ethics principle 1 - Study 3

Participants in this study are not being asked to reveal how they voted in a real election and are also not having to reveal any shift in their voting behaviour to a group. This study is therefore very unlikely to have any negative effect on the participants' psychological wellbeing, health, values or dignity.

Ethics principle 2 - Study 3

The study now tells potential participants what is involved before asking for their consent. It is possible that telling the participants what the study is exploring will affect how they respond and is therefore unlikely to provide any useful results. However, judging the ethics of research is a different matter from judging whether the design of the study will produce useful results.

Participants in this study would be aware that they were taking part in research, and were also told explicitly that they could withdraw from the study at any point.

As the study met all of the criteria, it should be approved.

In the end this research was approved by the ethics committee. One thing that is important to bear in mind is that obtaining approval involves a dialogue between the researcher(s) and the relevant ethics committee. This dialogue invariably involves an interpretation of the principles and a negotiation of what can be regarded as acceptable research conduct.

This activity focused on three key ethics principles, namely the right to withdraw, informed consent and the wellbeing of participants. However, ethics panels evaluating real research need to take into consideration a more complex set of issues.

Here is a PDF summary of the activity to print or save in your files.

Activity 2: Researching animals and humans

This activity explores the ethics of animal research and the guidelines that govern the use of animals (although, as you will see, not all animals) in psychological research.

Psychology is often thought to be just about human beings. However, there are important areas of psychological research that involve non-human animals.

Research with non-human animals poses two important questions:

  • How relevant are studies of non-human animals to human psychology? Aren’t human beings unique and different from other animals?
  • Is it acceptable to carry out experiments on non-human animals in the interests of science?

The tasks in this online activity give you the opportunity to consider these questions for yourself and clarify your own opinions and understanding of the issues they raise. The activity will help you to recognise why psychologists carry out research with non-human animals and identify the ethical issues involved in such research.

stanley milgram obedience experiment ethical issues

Why carry out research with animals?

You have been introduced to a number of ethics principles that apply to research on human participants. But why might psychologists want to do research with animals other than humans?

Here are some reasons. Select ‘Reveal comment’ to read a more detailed explanation:

To find out about the evolution of psychological functions .

By studying how different species adapt their behaviour to their environments, and trying to identify innate factors in adaptation, it is hoped that the interactions of genes, environments and learning can be understood better, thus shedding light on how evolution may have shaped human psychological processes.

To better understand psychological principles that apply across different species .

If similar processes are found in a range of species, this helps researchers to describe and (they hope) explain the basic principles of behaviour and other processes, such as attachment and learning.

To do experiments that would be unethical with humans .

Arguably, as you will learn later in this activity, it can be seen as less problematic to use non-human animals in experiments that involve potentially harmful conditions such as deprivation, pain or confinement, or for example, to explore the effects of punishment.

To better understand what is special about humans .

A substantial amount of research with non-human animals is carried out to identify psychological functions such as language and empathy, which other animals may not possess. Attempts to teach language to chimpanzees, for example, have met with only limited success, clarifying the specialised language abilities of humans.

The ethics of animal research

In this course you were introduced to a number of ethics principles that apply to research on human participants. These include informed consent from participants, ensuring their right to withdraw, protecting their welfare and evaluating the costs and benefits of the study.

Which of these do not apply to animals?

Of course, animals cannot give informed consent, and, given that most are kept in captivity anyway, the ‘right to withdraw’ does not really apply.

However, there are separate ethical guidelines for work with animals, which are also issued by the British Psychological Society.

They include provisions such as:

  • The ‘smallest number of animals sufficient to accomplish the research goals’ should be used in any study.
  • The costs and benefits of any study must be carefully evaluated.
  • The welfare of the animal must be taken into account and researchers must ‘seek to minimise any pain, suffering or distress that might arise’ from any experiment.
  • Researchers should use alternatives to animal research whenever possible, including data collected by other researchers, lower species (leeches, cell cultures, etc.) or, increasingly, computer simulations.

What emerges from these studies are the ‘3 Rs’ of animal research. These are to:

  • refine procedures to minimise suffering
  • reduce the number of individual animals used
  • replace animals with other alternatives.

These guidelines are interpreted and applied by ethics committees of research institutions and other bodies (including the Home Office) that grant special licences for keeping animals and using them in research.

Task 1: Evaluating the ethics of research

  • Use alternatives to animal research whenever possible, including data collected by other researchers, lower species (leeches, cell cultures, etc.) or increasingly, computer simulations.

Based on the guidelines on animal research, do you think an ethics committee would approve each of the following studies?

1. A researcher at a UK university is applying for a licence to replicate Harlow’s studies of deprivation. Infant monkeys would be raised in isolation with different types of ‘surrogate mother’ (inanimate objects that were either cloth-covered or made of wire with a milk bottle attached).

Do you think an ethics committee would give this study approval?

An ethics committee, and the Home Office, would probably not give approval for this study to be carried out. Not only does it raise issues about animal welfare and deprivation, but Harlow has already carried out this research, and it is highly unlikely that simply repeating the study would be considered a benefit great enough to outweigh the cost to the animals involved.

2. A researcher in a UK animal research laboratory is interested in addiction. He is proposing a study in which a small radio receiver would be implanted into the brains of seventy-five rhesus monkeys. These receivers would allow the researcher to activate areas of the brain thought to be associated with ‘pleasure’. The study would help shed light on brain mechanisms involved in addiction.

Several issues would need to be considered here. One is the number of monkeys involved. Ethics committees must ensure that the smallest number needed is actually used in the study. It is unlikely that as many as seventy-five monkeys would be absolutely necessary, so the researcher would probably be asked to make a very strong case, or reduce the number. Also, the ethics committee would want to hear what would happen to the monkeys after the experiment. Would they be able to live normally after the experiment is over?

3. A laboratory has been contracted by the Ministry of Defence to evaluate whether pigeons could be used in a guided missile to direct it towards an enemy aircraft in order to destroy it. The research would involve three pigeons being trained in a Skinner box to peck at targets on a radar screen.

Believe it or not, a study such as this, which drew on the ideas of B.F. Skinner, was carried out in the USA during the Second World War, as part of the so-called ‘Project Orcon’ or ‘Pigeon Project’. In this study there are no obvious ethical concerns (except for the more general issues of animal welfare), given that the pigeons’ ability to guide a missile is simply being evaluated, and a standard Skinner-box procedure is being proposed. So the project would probably receive ethical clearance, as the committee are likely to consider it to involve a discrimination learning task using specific stimuli (target on a radar screen), and there are no obvious animal welfare issues involved. (The pigeons would not actually be used in the attacks, or at least that is not what this project is about.)

4. A researcher has applied for permission to carry out an observational study in the Kalahari Desert of communication among pied babblers (wild birds) and with other species. The study would involve two researchers wearing camouflage hiding in the bushes, observing the behaviour of the pied babblers and recording their mating calls.

Research such as this is being done regularly around the world. While most of the ethical guidance that you have learned about so far in this activity has referred to laboratory work, the BPS also regulates observational work in the animals’ natural habitats. Researchers would be asked by the committee to follow careful protocols to ensure that disruption to animals’ lives is kept to a minimum. Disruption might have a detrimental effect not only on the birds’ lives, but might also impact on the inferences that can be drawn from the study. This is because the observed behaviour might be a reaction to intrusion by researchers, rather than something that occurs naturally, in the wild.

Different animals, different guidelines

Although human beings are animals, for most people there is a strong conceptual division between human and non-human animals. Humans – or homo sapiens – are seen as being fundamentally different even from pan paniscus (the bonobo chimpanzee (Figure 9) – genetically our closest relative in the animal world).

stanley milgram obedience experiment ethical issues

It is this widespread belief that humans occupy a special place in nature that underpins the whole notion that it might be appropriate to carry out some types of research with non-human animals, where human research ethics would not permit such research to be done with humans.

As you have seen, it’s not that there are no ethical considerations in research with other species, but rather that different, and less stringent, considerations apply.

There is, however, a further issue here. It is not just that humans are believed to be different from non-human animals. There are also differences between animals. In fact, what is meant by the word ‘animal’? While there are many people who believe that the animal world ends with the bonobo chimpanzee, we rarely think about where it begins.

The final task in this activity encourages you to explore this question, to examine your own views and to critically consider the issues involved.

Task 2: Reflecting on different species

stanley milgram obedience experiment ethical issues

Have a look at the list below and consider the following:

  • Which one would you consider to be an ‘animal’?
  • Which ones you would permit animal research on (with BPS guidance in mind)?

For each answer try to think of reasons why you referred to some of these species as ‘animals’ but not others, or why you think that research on some might be more appropriate than others. Once you have indicated your responses, go to the next page, where you will find the current legal status of research on these species and read about some of the issues involved in determining which animals can and cannot be used in research.

Animal research and the law

Only one of the organisms included in the list is not an animal: the E. coli bacterium. At least, that is the straightforward scientific answer. However, you may be surprised to learn that in many countries the law regulates what is, and what is not, an animal – at least where research is concerned. Such definitions determine which species are covered by the guidelines for psychologists working with animals. According to the Animals (Scientific Procedures) Act 1986, which legislates for research in the UK, only species that are vertebrates (possess a spinal column) and one single invertebrate, the octopus ( octopus vulgaris ), are legally defined as ‘animals’ when it comes to research. The same category of living things is also protected by the Animal Welfare Act 2006. Under this definition, the only ‘animals’ in the list are the rat, cat, dog, horse, dolphin and gorilla: the coral, woodlouse, crab, mosquito and scorpion are not legally ‘animals’. As for the second question, there are really no right or wrong answers. You were asked ‘which ones you would permit animal research on’ and that is a matter of personal ethics. However, you now know which of the species listed are protected by law (at least in the UK).

So there are many animals, or should we say ‘organisms’, that are not covered by any legislation, and for research purposes do not count as ‘animals’. There is no law protecting lobsters, spiders or mosquitoes (and especially not the E. coli bacteria) from being used in any way whatsoever in research. The British Psychological Society (BPS) does recommend that the ethics standards of work with protected species should also be maintained with organisms not covered by the relevant legislation, but this is just a recommendation.

Even among the vertebrates, further distinctions are made. Can you guess which animals receive extra protection? It is horses, cats, dogs and primates. In the case of primates, the reason is undoubtedly their genetic proximity to humans and the frequency with which they are used in research, but what about cats, dogs and horses?

The reason is that these animals are regarded by many humans as occupying a special place in the animal world, given that they are kept as pets. However, humans don’t always agree on which animals are pets. Whereas most people in the UK would not consider eating a dog, in parts of Asia dogs are considered a delicacy. Horses, donkeys and a host of other animals that people in some countries treat as companions adorn many a menu in other countries. So, some animals are a ‘man’s (or a woman’s) best friend’ whilst others are ‘vermin’ or ‘food’. In the same way, some species are regarded as ‘beautiful’, ‘cute’ or ‘intelligent’, while others are treated as less so. What all of this suggests is that differentiation within the animal world is not always based on strict scientific criteria, but rather on cultural sensitivities.

stanley milgram obedience experiment ethical issues

In this activity you were given an opportunity to consider a number of questions relevant to psychological research on animals. These included why psychologists study animals, how the rights of the animals used in research are protected, and how the appropriateness of animals for research is determined.

However, the controversy surrounding this research, especially the study on the effects of deprivation, had a different kind of impact on psychology. It made researchers more aware of the need to regulate research on animals and treat them more humanely. It was therefore in the aftermath of Harlow’s study that rules guiding psychological research on animals began to be tightened.

Activity 3: Researching animals

You have just learned about the ethics of animal research; now you have the opportunity of hearing from two psychologists working with animals in two films, Researching Animals ..

Reading about research that psychologists have conducted with animals is often fascinating, but seeing how the research is conducted is even better. Film A (8 minutes) introduces the work of Alex Thornton with meerkats; in Film B (17 minutes) you will learn about Tetsuro Matsuzawa’s work with chimpanzees.

First, watch the two films without interruption. After doing so, read ‘Issues to consider’ below, then watch them again, keeping these issues in mind. Make sure that you take some notes.

Copy this transcript to the clipboard

Transcript: Film A: Researching Animals (meerkats)

Transcript: film b: researching animals (chimpanzees), issues to consider.

  • Alex Thornton has looked at how meerkats teach their young to catch scorpions. To what extent is this process similar to or different from that which human parents use?
  • Tetsuro Matsuzawa found that, when it comes to completing the photographic memory task, chimpanzees are superior to humans. What explanation does he give for this finding? Can the performance of the chimpanzees be explained by conditioning?
  • Think about the location where the two researchers conduct their studies. What are the advantages and disadvantages of researching animals in their natural habitat compared to captivity?
  • Compare the reasons why Alex Thornton and Tetsuro Matsuzawa study animals. Which of them is interested in animal behaviour not just for its own sake, but also as a way of learning about human capacities?

This free course provided an introduction to studying sociology. It took you through a series of exercises designed to develop your approach to study and learning at a distance and helped to improve your confidence as an independent learner.

Acknowledgements

Grateful acknowledgement is made to the following sources.

Philip Banyard

Course image: Phil Dolby in Flickr made available under Creative Commons Attribution 2.0 Licence .

Figures 1, 2, 3 and 4: Milgram, S (1974) Obedience to Authority, Harper & Row. Copyright © 1974 by Stanley Milgram

Figure 5: Copyright © Andresr/iStock

Figure 6: Copyright © Michael Flippo/iStock

Angel Eye Media

With thanks to:

Kalahari Meerkat Project

iBrain Festival of Neuroscience at

Muziekcentrum De Bijloke

Primate Research Institute, Kyoto University.

Don't miss out:

If reading this text has inspired you to learn more, you may be interested in joining the millions of people who discover our free learning resources and qualifications by visiting The Open University - www.open.edu/ openlearn/ free-courses

Copyright © 2016 The Open University

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals
  • Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • Open access
  • Published: 25 November 2021

A novel experimental approach to study disobedience to authority

  • Emilie A. Caspar 1 , 2  

Scientific Reports volume  11 , Article number:  22927 ( 2021 ) Cite this article

  • Human behaviour

Fifty years after the experiments of Stanley Milgram, the main objective of the present paper is to offer a paradigm that complies with up-to-date ethical standards and that can be adapted to various scientific disciplines, ranging from sociology and (social) psychology to neuroscience. Inspired by subsequent versions of Milgram-like paradigms and by combining the strengths of each, this paper presents a novel experimental approach to the study of (dis)obedience to authority. Volunteers are recruited in pairs and take turns to be ‘agents’ or ‘victims’, making the procedure fully reciprocal. For each trial, the agents receive an order from the experimenter to send a real, mildly painful electric shock to the ‘victim’, thus placing participants in an ecological set-up and avoiding the use of cover stories. Depending on the experimental condition, ‘agents’ receive, or do not receive, a monetary gain and are given, or are not given, an aim to obey the experimenter’s orders. Disobedience here refers to the number of times ‘agents’ refused to deliver the real shock to the ‘victim’. As the paradigm is designed to fit with brain imaging methods, I hope to bring new insights and perspectives in this area of research.

Similar content being viewed by others

stanley milgram obedience experiment ethical issues

On the cognitive mechanisms supporting prosocial disobedience in a post-genocidal context

stanley milgram obedience experiment ethical issues

Neural mechanisms of deception in a social context: an fMRI replication study

stanley milgram obedience experiment ethical issues

Psychopathic traits mediate guilt-related anterior midcingulate activity under authority pressure

Introduction.

The experiment of Stanley Milgram is one of the most (in)famous in psychology 1 , within and beyond academia. Several variables account for this notoriety, such as the method used, the ethical issues associated, the enthralling results or the societal impact of the research topic. Milgram’s classical studies famously suggested a widespread willingness to obey authority, to the point of inflicting irreversible harm to another person just met a few minutes before. Beyond the studies of Milgram, the history of nations is also plagued by horrendous acts of obedience that have caused wars and the loss of countless lives 2 . History has fortunately shown that some individuals do resist the social constraint of receiving orders when their own morality is of greater importance than the social costs associated with defying orders (e.g., 3 , 4 ). To understand the factors that prevent an individual from complying with immoral orders, research on disobedience should focus on two main axes: (1) what social and situational factors support disobedience and (2) what individual differences support disobedience.

The first axe has already been largely investigated in past studies. From Milgram’s studies, important situational factors supporting disobedience have already been established 5 . For instance, disobedience increases if the experimenter is not physically present in the room or if two experimenters provide opposing views regarding the morality of the experiment. Subsequent versions and interpretations of Milgram’s studies 6 , 7 , 8 as well as historical research 4 , 9 also suggested the importance of several social (e.g. presence of a supporting group) and situational factors (e.g. family history, proximity with the ‘victim’, intensity of the pain; money) supporting resistance to immoral orders. However, the second axe regarding individual differences has been less systematically approached. A few studies 10 , 11 previously explored personality traits that may influence disobedience (e.g. empathic concern, risk-taking) but most of these studies, however, have used relatively weak and potentially biased methods, such as self-reported questionnaires and methods based on cover stories. These studies are not sufficient to explain why, in a given situation, some people will refuse immoral orders and rescue threatened human beings while others will comply with such orders. With the current literature on disobedience, we have no idea about which neuro-cognitive processes drive inter-individual differences regarding the degree of disobedience. This aim could be achieved by offering a novel experimental approach that would make it possible to use novel techniques that give us a more direct access to the functioning of the brain and cognition, such as functional near-infrared spectroscopy (fNIRS), electroencephalography or Magnetic Resonance Imagery (MRI). Regrettably, the original paradigm and those bearing close similarity are not adapted to reliably answer those questions as they were not designed to fit with neuroimaging measurements. By combining the strengths of previous work on disobedience into a single experimental paradigm and adapting it to fit with cognitive and brain imaging measurements, this novel experimental approach could help to better understand, together with individual, social, and cultural factors, which mechanisms make it possible for an individual to refuse to comply with immoral orders.

There were several challenges to consider in order to develop such a paradigm, both ethical and methodological. Studying obedience and resistance to immoral orders involves putting volunteers in a situation where they have to make a decision on whether or not to commit ‘immoral acts’ under orders. A balance has to be found between what is acceptable from an ethical perspective and what is necessary for the research question. Milgram’s studies on obedience raised undeniable ethical issues 12 , 13 , 14 , mostly associated with high stress and the use a cover story, which involves deception. Some variants of Milgram’s studies were realized with immersive virtual reality to prevent the ethical issues associated with Milgram’s paradigm 15 , but the transparency of the fake scenario presented to participants does not capture decision-making in an ecological set-up. Other Milgram-based variants, such as the 150-V method, appear to replicate Milgram’s results 16 with respect to the actual ethical standards, but methodological concerns are still present 17 as cover stories are still used, which lead to interpretation issues. Beyond ethical considerations, the use of deception also indeed involves a doubt about whether or not volunteers truly believed the cover story. As a consequence, a reasonable doubt remains on how to interpret the results and this is one of the main critics associated with Milgram’s studies and following versions. Recent work on the reports of Milgram’s volunteers suggested that there are no strong and reliable evidence that participants believed in the cover story 8 , 14 , 18 . Others suggested that since the stress of participants was visible on video recordings during the experiment (e.g. hand shaking, nervousness), this suggests that participants actually believed that they were torturing another human being 19 . However, this interpretation has been challenged by another study showing that participants can have physiological reactions to stress even in an obviously-fake experimental set-up 15 . These contrasting interpretations of Milgram’s studies actually reinforce the idea that results can hardly be interpreted when cover stories are used 20 . To answer those criticisms, a real scenario had thus to be created, where participants made decisions that have real consequences on another human being.

An additional challenge is that methods relying on the original paradigm of Milgram, such as the virtual reality version 15 or the 150-V method 16 are not adapted to neuroimaging measurements. More specifically, with such Milgram-like experimental approaches, only a single trial would be recorded for the entire experimental session, that is, when the volunteer stops the experiment (if this happens). For cognitive and neuroimaging data collection, a single trial per participant is not a reliable result, which requires the averaging of several trials to obtain a good signal-to-noise ratio.

Another challenge at the methodological and conceptual levels it that several experimenters 1 , 5 , 21 , 22 including myself 23 , 24 , 25 , 26 , 27 , noted that volunteers are extremely obedient when coming to an experiment. Personally, I have tested about 800 volunteers to investigate the mechanisms by which coercive instructions influence individual cognition and moral behaviors. For instance, by using behavioral, electrophysiological and neuroimaging methods, we have observed that when people obey orders to send real shocks to someone else, their sense of agency 23 , their feeling of responsibility 28 , empathy for the pain of the victim and interpersonal guilt 26 are attenuated compared to a situation where they are free to decide which action to execute. Out of 800 volunteers tested, only 27 disobeyed my orders (i.e. 3.3%): 21 for prosocial reasons (i.e. they refused to administer an electric shock to another individual), 3 by contradiction (i.e. by systematically pressing the other button, not matter the content of the order), and 3 for antisocial reasons (i.e. by administering shocks despite my order not to do so). Although convenient to study how obedience affects cognition, this rate is indubitably an issue when studying disobedience. If participants almost never disobey, we can’t study the mechanisms through which resistance to immoral orders may develop in a given situation. Several reasons for not disobeying the experimenter’s orders have been suggested. Some consider that being obedient is part of the human nature as massive and destructive obedience has been observed through countless historical events 2 . Another current view on the experiments of Milgram is that volunteers were actually happy to participate and to contribute to the acquisition of scientific data 17 , thus explaining the high obedience rate observed. This effect has been referred to as ‘engaged followership’ 29 . If that interpretation is correct, the volunteer’s willingness to come and help the experimenter acquiring scientific data creates an extra difficulty to obtain disobedience in an experimental setup. However, this interpretation is challenged by several studies reported by Milgram, which displayed a higher disobedience rate than his original study. For instance disobedience increases when the shocks’ receiver sits in the same room as the participant or when the authoritative experimenter is not physically present in the room 5 . If participants were indeed only guided by their willingness to help to acquire scientific data, this should be the case in any experimental set-up. As some studies involve a higher disobedience rate compared to the initial version of Milgram’s study 1 , they could thus, at a first glance, be used for studying disobedience. However, even if some versions of the initial study of Milgram offer a highly disobedience rate, thus making it possible to study the mechanisms through which resistance to immoral orders may develop in a given situation, these experimental set-ups are still not adapted for cognitive and neuroimaging measurements and still rely on the use of a cover story.

Taking all the presented challenges into account (i.e. not using cover stories to avoid interpretation issues; obtaining a fair rate of disobedience; using an experimental approach that also fits with cognitive and neuroimaging measurements; respecting ethical standards), the present paper presents a set of experiments that combine the strengths of past experimental work on (dis)obedience. Volunteers were openly involved and active (= real social situation) rather than having to act in fictitious scenarios (= imagined social situation, e.g. Slater et al., 2006). They were confronted with moral decisions to follow or not the orders from an experimenter to inflict a real painful shock to a ‘victim’ in exchange (or not) for a small monetary gain, thus avoiding the use of cover stories. Since the aim here is to develop a paradigm that could be used both in behavioral and neuroimaging studies, some basic characteristics had to be considered. For instance, to fit with a Magnetic Resonance Imagery (MRI) scanning environment, neither the ‘victim’, nor the experimenter were in the same room as the agent. A real-time video was thus used to display a video of the victim’s hand receiving shocks on the agent’s screen and headphones were used so the participant could hear the experimenter’s orders.

Another method to study disobedience would be to select participants who are more likely to disobey than others. Each volunteer was thus also asked to complete a series of personality questionnaires to evaluate if a specific profile is associated with a greater prosocial disobedience rate. Systematic post-experimental interviews were conducted at the end of each experiment in order to understand the decisions of volunteers to follow or not the orders of the experimenter and to ask them how they felt during the experiment.

Participants

A hundred eighty naive volunteers (94 females) were recruited in same gender dyads (= 90 dyads). During the recruitment procedure, I ensured that the participants in each dyad were neither close friends (by mixing people studying different academic courses), nor relatives. To estimate the sample size a priori, I calculated the total sample size based on an effect size f of (0.3). To achieve a power of 0.85 for this effect size, the estimated sample size was 168 for 6 groups 30 . I increased the sample size slightly to 180 in order to prevent loss of data in case of withdrawals. Volunteers were randomly assigned to one of the 6 variants of the task (N = 30/variant). One volunteer was not taken into account because they only played the role of the ‘victim’ to replace a participant who did not show up. No volunteers withdrew from the experiment. For the remaining 179 volunteers, the mean age was 22.63 years old (SD = 2.77, range:18–35). A Univariate ANOVA with Age as the dependent variable and Variant as the fixed factor confirmed that age of the volunteers did not differ between the different variant of the tasks ( p  > 0.1, BF 10  = 0.167). Volunteers received between €10 and €19.60 for their participation. All volunteers provided written informed consent prior to the experiment. The study was approved by the Ethics Committee of the Erasme Hospital (reference number: P2019/484). All methods were performed in accordance with the relevant guidelines and regulations.

Method and Material

Six experimental set-ups were created in a between-subject design. In all six set-ups, volunteers were invited by pairs. One person was assigned to start as agent and the other one to start as ‘victim’. Their roles were switched mid-way, ensuring reciprocity. Compared to the experimental design of Milgram, both volunteers were real participants, not confederates. The reciprocity also avoided volunteers to be stuck in the role of the person providing pain to the other, thus attenuating the potential psychological distress of being in a perpetrator role only. Volunteers were given the possibility to choose the role they wanted to start with. In the case none of them had a preference, role assignment was decided by a coin flip, but volunteers were reminded that they could still decide themselves. This procedure allows to ensure that participants do not think that this procedure is a trick.

Volunteers were first given the instructions of the task. Then, they signed the consent forms in front of each other, so both were aware of the other’s consent. The experimenter was never present in the same room, but rather gave the instructions through headphones. This was for two reasons. First, Milgram’s studies show that disobedience increases if the experimenter is not physically present in the room. Second, in the case of MRI scanning, the experimenter would not be able to give direct verbal instructions to the volunteers in the MRI room due to the high noise of the scanner. Here, agents were isolated in a room and were provided headphones to hear the experimenter’s instructions (see Fig.  1 ). They were told that this was done to avoid attentional interferences through the experimenter’s physical presence in the room. In this series of studies, instructions were pre-recorded but a real setup with a microphone connected to the headphones could also work. Pre-recordings allow perfect timing of the events, important for neuroimaging or electroencephalography recordings. The instructions were “ give a shock ” or “ don’t give a shock ”. To increase the authenticity of the procedure, each sentence was recorded 6 times with small variations in the voice and displayed randomly. In addition, the audio recordings included a background sound similar to interphone communications.

figure 1

Experimental setup. Schematic representation of the experimental setup. Volunteers were in different rooms. The experimenter was located in a third, separated room. The agent heard on a trial basis the orders of the experiment through headphones and had to decide to press the ‘SHOCK’ or ‘NO SHOCK’ button. A real-time camera feedback displayed the hand of the victim of the agent’s screen so to allow to keep track on the consequences of their actions.

Shocks were delivered using a constant current stimulator (Digitimer DS7A) connected to two electrodes placed on the back of victims’ left hand, visible to the agent through the camera display. Individual pain thresholds were determined for the two volunteers before starting the experiment. This threshold was determined by increasing stimulation in steps of 1 mA (Caspar et al., 2016). I approximated an appropriate threshold by asking a series of questions about their pain perception during the calibration (1. «  Is it uncomfortable?  »—2. «  Is it painful?  »—3. «  Could you cope with a maximum of 100 of these shocks?  »—4. «  Could I increase the threshold?  »). When roles were reversed, I briefly re-calibrated the pain threshold of the new victim by increasing the stimulation again from 0 in steps of 3 mA up to the previously determined threshold, to confirm that the initial estimate was still appropriate, and to allow re-familiarisation. The mean stimulation level selected by this procedure was 36.3 mA (SD = 17.5, V = 300, pulse duration: 200 µs). I chose this instead of other types of pain (e.g. financial) because it produces a clear muscle twitch on the victim’s hand each time a shock is sent. This allows volunteers to have a clear and visible feedback of the consequences of their actions and to be fully aware that shocks were real.

There was a total of 96 trials per experimental condition. In the coerced condition, the experimenter asked to give a shock in 64 trials and asked not to give a shock 32 trials. This ratio was chosen on the assumption that the volunteer’s willingness to refuse immoral orders would increase with the number of times they were instructed to inflict pain to the “victim”.

On each trial, a picture of two rectangles, a red one labelled ‘SHOCK’ and a green one labelled ‘NO SHOCK’, was displayed in the bottom left and right of the screen. The key-outcome mapping varied randomly on a trial-wise basis, but the outcome was always fully congruent with the mapping seen by the participant. Agents could then press one of the two buttons. Pressing the SHOCK key delivered a shock to the victim while pressing the NO SHOCK key did not deliver any shocks. This procedure of randomized button mapping allows to have a better control over motor preparation, an aspect that can be important for neuroimaging data.

In half of the variants of the task (i.e., 3/6), the “Aim” variants, participants were given a reason for obeying the orders of the experimenter, while this was not the case in the other half, the “No aim” variants. In the “No Aim” variants, I did not provide any reasons for obeying to the participants and I simply explained the task. If participants asked about the aim, I simply told them that they would know at the end of the experiment, without providing further justifications. In the “Aim” variants, volunteers were told that researchers observed a specific brain activity in the motor cortex in another study when participants were given instructions. We explained that the present study was a control study to measure different aspects linked to motor activity when they press buttons, in order to see if the button pressing was related to brain activity measured over the motor cortex. To increase the veracity of the aim, electrodes were also placed on their fingers and connected to a real electromyography (EMG) apparatus to supposedly record their muscle activity. Volunteers were instructed to press the two buttons only with their right and left index fingers, as naturally as possible, and to avoid producing too ample movements to create clean EMG data. In the case volunteers asked if they really had to follow orders, I told them that for ethical reasons I could not force them to do anything, but that it would be better for the sake of the experiment. Telling them explicitly that they could disobey the orders would not be beneficial in the quest of studying ‘real’ disobedience.

In 4 out of 6 variants of the task, the “Free-choice” variants, a second experimental condition was used, the free-choice condition. In this condition, volunteers were told that they could freely decide in each trial to shock the ‘victim’ or not. In this condition, they did not receive instructions. In 4 out of 6 variants of the task, the “Monetary reward” variants, agents received a monetary reward of + €0.05 for each shock delivered. In the other 2 variants, volunteers were not rewarded for each shock delivered (i.e. “No monetary reward” variants). To resume, the 6 variants of the same task were the following: (1) No Aim + Monetary reward + Free-choice condition; (2) No Aim + No monetary reward + Free-choice condition; (3) Aim + Monetary reward + Free-choice condition; (4) Aim + No monetary reward + Free-choice condition; (5) No Aim + Monetary reward + No free-choice condition; (6) Aim + Monetary reward + No free-choice condition (see Table 1 ).

Before the experimental session, volunteers filled in six questionnaires. Those questionnaires included (1) the Money Attitude Scale (e.g. “ I put money aside on a regular basis for the future ”) 31 , (2) the Moral Foundation Questionnaire (e.g. “ Whether or not someone showed a lack of respect for authority ”) 32 , (3) the Aggression-Submission-Conventionalism scale (e.g., “ We should believe what our leaders tell us ”) 33 , (4) the short dark triad scale (e.g., “ Most people can be manipulated ”) 34 , the Interpersonal Reactivity Index (e.g. “ When I see someone get hurt, I tend to remain calm ”) 35 . At the end of the experimental session, they were asked to fill in two more questionnaires: (1) A debriefing assessing what they felt during the experiment and the reasons for choosing to obey or disobey the orders of the experimenter (Supplementary Information S1) and (2) a questionnaire on social identification with the experimenter (e.g., “ I feel strong ties with this experimenter ”) 36 . At the end of the experiment a debriefing was conducted for each volunteer, separately. Volunteers were then paid, again separately.

General data analyses

Each result was analyzed with both frequentist and Bayesian statistics 37 . Bayesian statistics assess the likelihood of the data under both the null and the alternative hypothesis. BF 10 corresponds to the p (data| H 1 )/ p (data| H 0 ). Generally, a BF between 1/3 and 3 indicates that the data is similarly likely under the H 1 and H 0 , and that the data does not adjudicate which is more likely. A BF 10 below 1/3 or above 3 is interpreted as supporting H 0 and H 1 , respectively. For instance, BF 10  = 20 would mean that the data are 20 times more likely under H 1 than H 0 providing very strong support for H 1 , while BF 10  = 0.05 would mean that the data are 20 times more likely under H 0 than H 1 providing very strong support for H 0 38 . BF and p values were calculated using JASP 39 and the default priors implemented in JASP. All analyses were two-tailed.

Number of shocks given in the free-choice condition

In the free-choice condition, volunteers were told that they were entirely free to decide to deliver a shock or not to the ‘victim’ on each of the 96 free-choice trials. On average, agents administered shocks to the victim on 31.86% of the trials (SD = 34.98, minimum: 0%, maximum: 100%) in the free-choice condition, corresponding to 30.59/96 shocks. A paired-sample t-test indicated that agents delivered less frequently a shock in the free-choice condition than in the coerced condition (68.03%, SD = 41.11, t (119)  = -9.919, p  < 0.001, Cohen’s d = − 0.906, BF 10  = 1.987e + 14). This result supports the fact that individuals can inflict more harm to others when they obey orders than when they act freely.

Prosocial disobedience across variants

In the present study, I was interested in prosocial disobedience, that is, when agents refuse the orders of the experimenter to send a painful shock to the ‘victim’. Table 2 displays the number of volunteers who reported that they voluntarily disobeyed in each variant of the task.

In this experiment, the main variable of interest was not to consider how many participants disobeyed in each variant only, but also how frequently they disobeyed. A percentage of prosocial disobedience was calculated for each volunteer, corresponding to the number of trials in which participants chose to disobey (i.e., sending no shocks while ordered by the experimenter to do so) divided by the total number of trials corresponding to the order to send a shock, multiplied by 100. I compared the prosocial disobedience rate across variants of the task, gender of participants and order of the role. I conducted a univariate ANOVA with prosocial disobedience as the dependent variable and Aim (aim given, no aim given), Monetary reward (+ €0.05 or not), Free-choice (presence or absence of a free-choice condition), Gender and Order of the Role (agent first, victim first) as fixed factors (see Fig.  2 ). Both frequentist and Bayesian statistics strongly supported a main effect of Aim (F (1,155)  = 14.248, p  < 0.001, η 2 partial  = 0.084, BF incl  = 158.806). Prosocial disobedience was lower when an aim for obedience was given to volunteers (20.4%, CI 95  = 12.8–28.1) than when no aim was given (43.3%, CI 95  = 35.6–51). Both frequentist and Bayesian statistics also supported a main effect of Monetary reward (F (1,155)  = 12.335, p  = 0.001, η 2 partial  = 0.074, BF incl  = 28.930). Prosocial disobedience was lower when a monetary reward was given for each shock (25.1%, CI 95  = 18.5–31.7) than when no monetary reward was given (45.4%, CI 95  = 35.9–54.8). The frequentist approach showed a main effect of Gender (F (1,155)  = 5.128, p  = 0.025, η 2 partial  = 0.032), with a lower prosocial disobedience rate for female volunteers (25.7%, CI 95  = 18.2–33.2) then for male volunteers (38%, CI 95  = 30–46). However, the Bayesian version of the same analysis revealed a lack of sensitivity (BF incl  = 0.871). All other main effects or interactions supported H 0 or a lack of sensitivity (all p s > 0.1 & BFs incl  ≥ 0.4.291E-7 & ≤ 1.178).

figure 2

Graphical representation of the percentages of prosocial disobedience in each variant of the task.

The following results report two-tailed Pearson correlations between prosocial disobedience and several other variables, including (1) the reasons given for disobeying, (2) the feeling of responsibility, badness and how sorry they experienced during the experiment, (3) the identification with the experimenter, (4) the perceived level of pain of the victim, (5) identification with the ‘victim’, and (6) individual differences measured through self-report questionnaires. I applied a False Discovery Rate (FDR) approach with the Benjamini and Hochberg method 40 to each p-value for each of those correlations but for the sake of clarity these variables are reported in different sub-sections.

Reasons for prosocial disobedience

All participants who reported that they voluntarily disobeyed the orders of the experimenter (N = 108) were presented a list of 10 reasons that they had to rate from “Not at all” to “Extremely” (see Supplementary Information S1). The reason ‘ I wanted to make more money ’ was only considered for the data of volunteers who had a variant with a monetary reward for each shock (N = 68). Both frequentist and Bayesian statistics showed that the percentage of prosocial disobedience positively correlated with moral reasons (r = 0.550, p FDR  < 0.001, BF 10  = 1.700e + 7), positively correlated with disobedience by contradiction (r = 0.329, p FDR  < 0.001, BF 10  = 47.53) and negatively correlated with the willingness to make more money (r = − 0.485, p FDR  < 0.001, BF 10  = 822.16). Other correlations were in favor of H 0 or were inconclusive (all p s FDR  > 0.076, all BFs 10  ≥ 0.120 & ≤ 1.446).

Feeling responsible, bad and sorry

Both frequentist and Bayesian statistics showed strong positive correlations between prosocial disobedience and how responsible (r = 0.299, p FDR  < 0.001, BF 10  = 343.98) and how bad (r = 0.301, p FDR  < 0.001, BF 10  = 384.65) they felt during the task (see Figs.  3 A and B). The more responsible and worse they felt during the task, the more they refused the order to send a shock to the ‘victim’. How sorry they felt was inconclusive ( p FDR  > 0.08, BF 10  = 0.929).

figure 3

Graphical representation of Pearson correlations between prosocial disobedience and ( A ) feeling of responsibility, ( B ) how bad agents felt during the task when they administered shocks to the ‘victim’, and ( C ) how painful they estimated the shock delivered to the ‘victim’ was. All tests were two-tailed.

Identification with the experimenter

Both frequentist and Bayesian statistics strongly supported H 0 regarding the relationship between prosocial disobedience and personal identification ( p FDR  > 0.5, BF 10  = 0.121) and bonding with the experimenter ( p FDR  > 0.5, BF 10  = 0.117). The relationship between the charisma of the experimenter and prosocial disobedience was also slightly in favor of H 0 ( p FDR  > 0.1, BF 10  = 0.530).

Estimated pain of the ‘victim’

The frequentist approach showed a positive correlation between the perceived pain of the ‘victim’ and prosocial disobedience (r = 0.189, p FDR  = 0.048). The higher they considered the ‘victim’ to be in pain, the more frequently they refused to deliver the shock. The Bayesian version of the same analysis slightly supported this relationship (BF 10  = 2.236), see Fig.  3 C.

Identification with the ‘victim’

In the post-session questionnaire, volunteers had to identify to what extent they considered that the other participant could be part of their group and to what extent they identified with the other participant. Both frequentist and Bayesian statistics strongly supported H 0 regarding the relationship between prosocial disobedience and the perception that the other participant could be part of one’s own group ( p FDR  > 0.8, BF 10  = 0.096). The relationship between prosocial disobedience and the identification with the other participant also slightly supported H 0 ( p FDR  > 0.1, BF 10  = 0.511).

Correlations between the behavior of pairs of participants

As we used a role reversal procedure, the behavior of those who were agents first could influence the behavior of those who turned agents afterwards. A Pearson correlation between prosocial disobedience of agents first and prosocial disobedience of victims who turned agents afterwards. The correlation was positive (r = 0.514, p  < 0.001, BF 10  = 60,068.704), suggesting participants who were agents second tend to act similarly as those who were agents first.

Individual differences associated with prosocial disobedience

Another approach to ensure a reliable prosocial disobedience rate when recruiting volunteers would be to target individuals with a profile that is most frequently associated with disobedient behaviors. Both frequentist and Bayesian statistics for exploratory correlations were two-tailed. Cronbach’s α for each subscale is presented in Supplementary Information S2. Both frequentist and Bayesian statistics showed a negative correlation between scores on the Authority subscale (r = -0.259, p FDR  < 0.001, BF 10  = 41.372) and the Purity subscale (r = -0.303, p FDR  < 0.001, BF 10  = 424.97) from the MFQ questionnaire. The lower volunteers scored on authority and purity, the higher was their prosocial disobedience rate. Other correlations were in favor of H 0 or were inconclusive (all p s FDR  ≥ 0.048, all BFs 10  ≥ 0.100 & ≤ 2.314).

Reasons for obedience

If participants reported that they did not voluntarily disobey the orders of the experimenter, they were asked in an open question to explain their decision to comply with those orders. After reading all the answers, three categories were extracted from the reasons provided: (1) ‘For science’ reasons; participants reported that they obeyed to allow reliable data acquisition (e.g., Participant 91: “ Pour ne pas fausser l’étude ”—English translation: “ To avoid biasing the stud y”); (2) ‘For respect of authority’ reasons; participants reported that they had to follow the orders of the authoritative figure (e.g., Participant 13: “ Pour moi c’est normal de suivre un ordre ”—English translation: “ In my opinion, it’s normal to follow an order ”), and (3) ‘For lack of side-effects’ reasons; participants reported that since the shocks delivered were calibrated on one’s own pain threshold, obeying orders to shock was not problematic (e.g., Participant 115: “ Douleur supportable pour l'autre, je n'ai accepté de faire subir que ce que j'aurais été prêt à subir moi-même ”—English translation: “ The pain was tolerable for the other participant, I have accepted to inflict the intensity of the pain that I would have been ready to undergo myself ”). An independent, naive judge classified the response of participants in one or several of those three established categories. Analyses of the frequencies revealed that the reason “For Science” was mentioned 31/70 times, the reason “For lack of side-effects” was mentioned 17/70 times and the reason “For respect of authority” was mentioned 31/70 times.

The aim of the present paper was to present a novel experimental approach to study (dis)obedience to immoral orders, by combining the strength of past experimental work and by adapting it to cognitive and neuroimaging measurements. Although other versions were proposed since Milgram’s studies, like a study in an immersive virtual environment 15 or the 150-V method 16 , some methodological concerns remained as those methods still involved cover stories or fake experimental set-ups. Here, the experimental approach was significantly different as it was based on an entirely transparent method that involved the administration of real electric shocks to another individual. This approach has the advantage to solve some of the main ethical and methodological concerns associated with the use of cover stories. It also has the advantage that it be can used both to study how social and situational factors influence disobedience as well as individual factors. For social and situational factors, the proposed paradigm can be adapted to evaluate for instance the influence of a supporting group, the use of high or low monetary rewards or how priming disobedience with a documentary influence disobedience. For individual factors, the paradigm allows to investigate how personality traits influence disobedience or to study the neuro-cognitive processes underlying disobedience.

Some novel theories combining a multi-method approach based on social psychology, neuroeconomics and neuroscience could thus emerge to understand better the mechanisms supporting disobedience. For instance, one could evaluate how empathy for the pain of the victim predicts disobedience and how the presence of a supporting group influences our capacity to feel empathy 41 and/or compassion for the ‘victim’ 42 . It could also be argued that the presence of a supporting group diffuses responsibility between individuals and increases obedience, by influencing how our brain processes agency and responsibility over our actions 28 , 43 , 44 , 45 . As the results obtained in the present study also indicated that feeling bad for the shocks delivered was statistically associated with prosocial disobedience, one could evaluate how the neural correlates of guilt 46 predicts prosocial disobedience and what historical, cultural and individual factors influence the feeling of guilt.

Six variants of the same task were tested in the present study, some inducing a higher prosocial disobedience rate than others. Statistical results showed that providing a reason—or aim—to justify obedience strongly decreased disobedience. Providing a monetary reward, even one as small as €0.05, also strongly decreased disobedience. Variant 2, in which volunteers were not given an aim or monetary reward, showed the highest disobedience rates. However, to study disobedience in ecological way, the paradigm should capture disobedience of participants even if they know that they are losing something (i.e., monetary rewards or the ‘trust’ of the experimenter asking them help for the study). Defying the orders of an authority generally involves social and/or monetary costs in real-life situations. I would thus not recommend using an experimental paradigm in which volunteers have no costs associated with defying the orders of the experimenter, as it would reduce the ecology of the disobedience act. Variants 3 and 6 involve two types of costs for resisting the orders of the experimenter: a monetary loss and deceiving the experimenter. In Variant 3, descriptive statistics showed that prosocial disobedience was lower compared to Variant 6. The main difference between these two variants was the presence of a free-choice condition. In my former studies 23 , 27 , volunteers frequently justified obedience in the coerced condition because they were given freedom in the free-choice condition (e.g. Participant 89 – English Translation: “ (…) In addition, I knew I could chose freely in the other condition not to send shocks—what I did ). In the present debriefings, some volunteers also reported that the presence of a free-choice condition was giving them enough freedom to accept to follow the orders in the coerced condition. In the supplementary analyses, results showed that when the monetary reward and the aim for obeying are identical, being given a free-choice condition reduces disobedience in the coerced condition. Therefore, Variant 6 appears to provide a good balance between reaching a reliable disobedience rate and finding volunteers who would refuse to produce physical harm on another human beings despite the monetary or social costs associated with defying orders.

Another approach would be to pre-select people who are predicted to be more disobedient. Personality questionnaires indicated that scoring low on the authority and on the purity subscale of the MFQ was strongly associated with a higher prosocial disobedience rate. The link between one’s own relationship to authority and prosocial disobedience observed here replicates another study conducted on the first generation of Rwandese after the 1994 genocide 47 . One’s own relationship to authority thus appears to be a reliable predictor variable in order to pre-select a sample that is more likely to disobey immoral orders.

In the present paper, administering a real mildly painful shock in exchange or not for a small monetary gain was described as an ‘immoral’ act. The notion of what is moral or not can highly differ between individuals 48 , for both academics and volunteers participating in an experiment. Humans are indeed sensitive to different competing issues of morality, a key reason for rescuing persecuted people 49 . In accordance with this observation, the present results indicated that moral reasons were a critical factor associated with the prosocial disobedience rate: the more shocking partners was considered as immoral, the more volunteers disobeyed. However, considering an action as against one’s own moral values does not necessarily translate to a refusal—especially when this order is in line with the Law. An extreme example is soldiers who have perpetrated acts that transgressed their moral beliefs but were issued by their superior in combat 50 . A core question for future research remains: Why are some people capable of putting their own moral standards above the social costs associated with defying orders?

Results indicate that the more volunteers felt responsible during the task, and the worse they felt for sending shocks to the ‘victim’, the higher was their prosocial disobedience. In another study, we observed that obeying orders reduced the feeling of responsibility, how bad and how sorry volunteers felt compared to being free to decide 26 . One hypothesis is that individuals who have preserved a feeling of responsibility and feeling bad—even under command—could more easily defy immoral orders. However, future studies are necessary to better understand the neuro-cognitive processes that prevent an individual from complying with immoral orders. As this paradigm is adapted to neuroimaging measurements, a whole range of studies could now be conducted.

It has been previously suggested that a strong identification with the experimenter giving orders is associated with higher obedience 36 . However, in the present paper, correlations between prosocial disobedience and identification with the experimenter were in favor of H0 with both the frequentist and the Bayesian approaches. In a former study, we also observed that identification to the experimenter was not a critical aspect for explain (dis)obedience. We observed that the generation of Rwandese born after the genocide and tested in Rwanda reported a higher identification to the experimenter than the same generation of Rwandese but tested in Belgium 47 . However, the latter group had a higher prosocial disobedience rate than the former group. Future studies must thus be conducted to understand how the identification with the person giving orders could influence obedience and its weight compared to other social, cultural and individual variables.

Although some volunteers reported that they felt a bit stressed and anxious during the task when they were in the role of the agent, the overwhelming majority did not report any negative psychological feelings. None of the participants withdrew from the experiment and none reported long-term negative psychological effects.

Nowadays, it has become difficult to find volunteers who do not know Milgram’s studies given the high media coverage, including movies, radio soaps, books, podcast and documentaries. One could expect that knowing Milgram would prevent people to obey. However, for the large majority of volunteers, it appears that this is not the case. In previous studies that I conducted with a relatively similar paradigm, the disobedience rate was drastically low (i.e. 3.3%) even if participants were university students knowing Milgram’s studies. In the present study, almost all the volunteers who participated in the present study knew Milgram and explicitly mentioned him during the oral debriefings or before starting the experiment. Yet for those who disobeyed, almost none reported that the reason for disobedience was that they thought it was the aim of the experiment. Further, there was no statistical relationship between prosocial disobedience and believing that it was the aim of the study. It does not mean that knowing Milgram would not influence at all disobedience. It rather suggests that knowing Milgram is not the main factor influencing one’s decision to obey or not an experimenter. It is also possible that since in this experiment shocks were real and not fake such as in Milgram’s studies, participants considered that this was indeed not a study aiming to replicate Milgram.

As far as I have observed, the main problem associated with knowing Milgram’s studies is that volunteers believe that I also have hidden aims and procedures when they enter the experimental room. Several volunteers reported that they only realized that my explanations for the task were true when they were explicitly offered the choice to decide which role to play first and/or when they started receiving the shocks. This is a general concern in psychological studies: The high use of cover stories can also impact other research, as volunteers start to develop a mistrust in what researchers tell them.

Results indicated that who were agents second tend to act similarly as those who were agents first, by sending a relatively similar amount of shocks. Of note, this is an effect that we also observed in past studies on the effect of obeying orders on cognition 23 , 26 , 43 . Nonetheless, in none of those studies we observed that the order of the role had a statistical influence on the neuro-cognitive processes targeted. However, the influence on role reversal on disobedience and related neuro-cognitive processes has still to be investigated in future studies.

The present paradigm is ecological in the sense that volunteers are facing decisions that have a real, physical impact on another human being. However, at the moment I only have little evidences that this paradigm has ecological validity to reflect obedience in real life situations, especially regarding “destructive disobedience” 17 . Caution is indicated when making inference from laboratory studies to complex social behaviours, such as those observed during genocides 16 . My main evidence at the moment is that the very low rate of prosocial disobedience observed in the first generation of post-genocide Rwandans tested in Rwanda using this paradigm 47 is consistent with the fact that deference to authority had already been emphasized by academics as an important factor in the 1994 genocide 4 , 51 . Individual scores on deference to authority in Caspar et al. 47 was the best predictive factor for prosocial disobedience in that former paradigm, thus suggesting some ecological validity. A promising approach would be to recruit “Righteous Among the Nations”, individuals who really saved lives during genocides. Testing this population with the present paradigm would put the ecological validity of this paradigm to the test.

People’s ability to question and resist immoral orders is a fundamental aspect of individual autonomy and of successful societies. As Howard Zinn famously wrote: “ Historically, the most terrible things—war, genocide, and slavery—have resulted not from disobedience, but from obedience ”. Understanding how individuals differ in the extent to which they comply with orders has undeniably several societal implications. They range from understanding how evolving in highly hierarchical environments — such as the military or prisons—influences moral behaviours, to developing interventions that would help to prevent blind obedience and help to resist calls to violence in vulnerable societies. However, since Milgram’s studies, the topic of disobedience has been mostly studied by social psychologists using adapted versions of the initial paradigm developed by Milgram. I hope that with this novel approach, (dis)obedience research will be given a new boost and will be considered by other scientific disciplines seeking to understand better human behaviours.

Data availability

Data are made available on OSF (DOI: https://doi.org/10.17605/OSF.IO/2BKJC ).

Milgram, S. Behavioral study of obedience. Psychol. Sci. Public Interest 67 (4), 371–378. https://doi.org/10.1037/h0040525 (1963).

Article   CAS   Google Scholar  

H. Zinn, The Zinn Reader: Writings on Disobedience and Democracy . Seven Stories Press, 1997.

Roisin, J. Dans la nuit la plus noire se cache l'humanité: Récits des justes du Rwanda. Les Impressions nouvelles (2017).

Fox, N. & Nyseth Brehm, H. I decided to save them: Factors that shaped participation in rescue efforts during genocide in Rwanda. Soc. Forces 96 (4), 1625–1648. https://doi.org/10.1093/sf/soy018 (2018).

Article   Google Scholar  

S. Milgram, Obedience to Authority: an Experiment View . Harper & Row, 1974.

Blass, T. Understanding behavior in the Milgram obedience experiment: The role of personality, situations, and their interactions. J. Pers. Soc. Psychol. 60 (3), 398–413. https://doi.org/10.1037/0022-3514.60.3.398 (1991).

Article   ADS   Google Scholar  

Dolinski, D. & Grzyb, T. The (doubtful) role of financial reward in obedience to authority. J. Soc. Psychol. 159 (4), 490–496. https://doi.org/10.1080/00224545.2018.1505708 (2019).

Article   PubMed   Google Scholar  

Haslam, N., Loughnan, S. & Perry, G. Meta-milgram: An empirical synthesis of the obedience experiments. PLoS ONE 9 (4), e93927. https://doi.org/10.1371/journal.pone.0093927 (2014).

Article   ADS   CAS   PubMed   PubMed Central   Google Scholar  

Fagin-Jones, S. & Midlarsky, E. Courageous altruism: Personal and situational correlates of rescue during the Holocaust. J. Posit. Psychol. 2 (2), 136–147. https://doi.org/10.1080/17439760701228979 (2007).

Bègue, L. et al. Personality predicts obedience in a milgram paradigm. J. Pers. 83 , 299–306. https://doi.org/10.1111/jopy.12104 (2015).

S. P. Oliner and P. M. Oliner, The altruistic personality: Rescuers of Jews in Nazi Europe . New York, NY, US: Free Press, 1988, pp. xxv, 419.

Baumrind, D. Some thoughts on ethics of research: After reading Milgram’s “Behavioral Study of Obedience”. Am. Psychol. 19 (6), 421–423. https://doi.org/10.1037/h0040128 (1964).

A. G. Miller, The obedience experiments: A case study of controversy in social science . New York, NY, England: Praeger Publishers, 1986, pp. ix, 295.

Perry, G. Behind the Shock Machine: The Untold Story of the Notorious Milgram Psychology Experiments (New Press, 2013).

Google Scholar  

Slater, M. et al. A virtual reprise of the stanley milgram obedience experiments. PLoS ONE 1 (1), e39. https://doi.org/10.1371/journal.pone.0000039 (2006).

Article   ADS   PubMed   PubMed Central   Google Scholar  

Burger, J. M. Replicating Milgram: Would people still obey today?. Am. Psychol. 64 (1), 1–11. https://doi.org/10.1037/a0010932 (2009).

Miller, A. G. Reflections on “replicating milgram” (Burger, 2009). Am. Psychol. 64 (1), 20–27. https://doi.org/10.1037/a0014407 (2009).

Griggs, R. A. & Whitehead, G. I. Coverage of recent criticisms of Milgram’s obedience experiments in introductory social psychology textbooks. Theory Psychol. 25 (5), 564–580. https://doi.org/10.1177/0959354315601231 (2015).

T. Blass, Obedience to Authority: Current Perspectives on the Milgram Paradigm . Psychology Press, 1999.

Kelman, H. C. Human use of human subjects: The problem of deception in social psychological experiments. Psychol. Bull. 67 (1), 1–11. https://doi.org/10.1037/h0024072 (1967).

Article   CAS   PubMed   Google Scholar  

Beauvois, J.-L., Courbet, D. & Oberlé, D. The prescriptive power of the television host. A transposition of Milgram’s obedience paradigm to the context of TV game show. Eur. Rev. Appl. Psychol. 62 (3), 111–119. https://doi.org/10.1016/j.erap.2012.02.001 (2012).

Frank, J. D. Experimental studies of personal pressure and resistance: I. experimental production of resistance. J. Gen. Psychol. 30 (1), 23–41. https://doi.org/10.1080/00221309.1943.10544454 (1944).

Caspar, E. A., Christensen, J. F., Cleeremans, A. & Haggard, P. Coercion changes the sense of agency in the human brain. Curr. Biol. 26 (5), 585–592. https://doi.org/10.1016/j.cub.2015.12.067 (2016).

Article   CAS   PubMed   PubMed Central   Google Scholar  

Caspar, E. A., Vuillaume, L., Magalhães de Saldanha da Gama, P. A. & Cleeremans, A. The Influence of (Dis)belief in free will on immoral behavior. Front. Psychol. https://doi.org/10.3389/fpsyg.2017.00020 (2017).

Article   PubMed   PubMed Central   Google Scholar  

Caspar, E. A., Cleeremans, A. & Haggard, P. Only giving orders? An experimental study of the sense of agency when giving or receiving commands. PLoS ONE 13 (9), e0204027. https://doi.org/10.1371/journal.pone.0204027 (2018).

Caspar, E. A., Ioumpa, K., Keysers, C. & Gazzola, V. Obeying orders reduces vicarious brain activation towards victims’ pain. Neuroimage 222 , 117251. https://doi.org/10.1016/j.neuroimage.2020.117251 (2020).

Caspar, E. A., LoBue, S., Magalhães de Saldanha da Gama, P. A., Haggard, P. & Cleeremans, A. The effect of military training on the sense of agency and outcome processing. Nat. Commun. https://doi.org/10.1038/s41467-020-18152-x (2020).

Beyer, F., Sidarus, N., Bonicalzi, S., & Haggard, P. Beyond self-serving bias: diffusion of responsibility reduces sense of agency and outcome monitoring. Soc Cogn Affect Neurosci. 12 (1), 138–145. https://doi.org/10.1093/scan/nsw160 (2017).

Haslam, S. A., Reicher, S. D., Millard, K. & McDonald, R. “Happy to have been of service”: The Yale archive as a window into the engaged followership of participants in Milgram’s “obedience” experiments. Br. J. Soc. Psychol. 54 (1), 55–83. https://doi.org/10.1111/bjso.12074 (2015).

Faul, F., Erdfelder, E., Lang, A.-G. & Buchner, A. G*Power 3: A flexible statistical power analysis program for the social, behavior, and biomedical sciences. Behav. Res. Methods 39 , 175–191. https://doi.org/10.3758/BF03193146 (2007).

Yamauchi, K. T. & Templer, D. J. The development of a money attitude scale. J. Pers. Assess. 46 (5), 522–528. https://doi.org/10.1207/s15327752jpa4605_14 (1982).

Graham, J. et al. Mapping the moral domain. J. Pers. Soc. Psychol. 101 (2), 366–385. https://doi.org/10.1037/a0021847 (2011).

Dunwoody, P. T. & Funke, F. The aggression-submission-conventionalism scale: Testing a new three factor measure of authoritarianism. J. Soc. Polit. Psychol. 4 (2), 571–600. https://doi.org/10.5964/jspp.v4i2.168 (2016).

Jones, D. N. & Paulhus, D. L. Introducing the short dark triad (SD3): A brief measure of dark personality traits. Assessment 21 (1), 28–41. https://doi.org/10.1177/1073191113514105 (2014).

M. Davis, A multidimensional approach to individual differences in empathy. JSAS Catalog Sel. Doc. Psychol. , vol. 10, (1980).

Steffens, N. K., Haslam, S. A. & Reicher, S. D. Up close and personal: Evidence that shared social identity is a basis for the “special” relationship that binds followers to leaders. Leadersh. Q. 25 (2), 296–313. https://doi.org/10.1016/j.leaqua.2013.08.008 (2014).

Dienes, Z. Bayesian versus orthodox statistics: Which Side Are You On?. Perspect. Psychol. Sci. 6 (3), 274–290. https://doi.org/10.1177/1745691611406920 (2011).

Marsman, M. & Wagenmakers, E.-J. Bayesian benefits with JASP. Europ. J. Develop. Psychol. 14 (5), 545–555. https://doi.org/10.1080/17405629.2016.1259614 (2017).

JASP Team, ‘JASP (Version 0.14.10)’. 2019.

Benjamini, Y. & Hochberg, Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. J. Roy. Stat. Soc.: Ser. B (Methodol.) 57 (1), 289–300. https://doi.org/10.1111/j.2517-6161.1995.tb02031.x (1995).

Article   MathSciNet   MATH   Google Scholar  

Singer, T. et al. Empathy for Pain involves the affective but not sensory components of pain. Science 303 (5661), 1157–1162. https://doi.org/10.1126/science.1093535 (2004).

Article   ADS   CAS   PubMed   Google Scholar  

Valk, S. L. et al. Structural plasticity of the social brain: Differential change after socio-affective and cognitive mental training. Sci. Adv. 3 (10), e1700489. https://doi.org/10.1126/sciadv.1700489 (2017).

Caspar, E. A., Beyer, F., Cleeremans, A. & Haggard, P. The obedient mind and the volitional brain: A neural basis for preserved sense of agency and sense of responsibility under coercion. PLoS ONE 16 (10), e0258884. https://doi.org/10.1371/journal.pone.0258884 (2021).

Beyer, F., Sidarus, N., Bonicalzi, S. & Haggard, P. Beyond self-serving bias: diffusion of responsibility reduces sense of agency and outcome monitoring. Social Cognit. Affect. Neurosci. 12 (1), 138–145. https://doi.org/10.1093/scan/nsw160 (2017).

Haggard, P. Sense of agency in the human brain. Nat. Rev. Neurosci. https://doi.org/10.1038/nrn.2017.14 (2017).

Yu, H. et al. A generalizable multivariate brain pattern for interpersonal guilt. Cereb. Cortex 30 (6), 3558–3572. https://doi.org/10.1093/cercor/bhz326 (2020).

E. Caspar, D. Gishoma, and P. A. M. D. S. da Gama, ‘Obedience to authority in the aftermath of a genocide. A social neuroscience study in Rwanda’. PsyArXiv, Jun. 23, 2021. https://doi.org/10.31234/osf.io/a8r7y .

B. Gert, and J. Gert, ‘The Definition of Morality’, in The Stanford Encyclopedia of Philosophy , Fall 2020., E. N. Zalta, Ed. Metaphysics Research Lab, Stanford University, 2020. Accessed: Apr. 07, 2021. [Online]. Available: https://plato.stanford.edu/archives/fall2020/entries/morality-definition/

Gross, M. L. Jewish rescue in holland and france during the second world war: Moral cognition and collective action*. Soc. Forces 73 (2), 463–496. https://doi.org/10.1093/sf/73.2.463 (1994).

Shay, J. Moral injury. Psychoanal. Psychol. 31 (2), 182–191. https://doi.org/10.1037/a0036090 (2014).

Article   MathSciNet   Google Scholar  

Prunier, G. The Rwanda Crisis: History of a Genocide (C. Hurst & Co., 1998).

Download references

Acknowledgements

Emilie A. Caspar was funded by the F.R.S-FNRS.

Author information

Authors and affiliations.

Moral and Social Brain Lab, Department of Experimental Psychology, Ghent University, Henri Dunantlaan, 2, 9000, Ghent, Belgium

Emilie A. Caspar

Center for Research in Cognition and Neuroscience, Université Libre de Bruxelles, Brussels, Belgium

You can also search for this author in PubMed   Google Scholar

Contributions

E.A.C. developed the study concept and the method. Testing, data collection and data analysis were performed by E.A.C. E.A.C. wrote the manuscript.

Corresponding author

Correspondence to Emilie A. Caspar .

Ethics declarations

Competing interests.

The author declares no competing interests.

Additional information

Publisher's note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Supplementary information., rights and permissions.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Cite this article.

Caspar, E.A. A novel experimental approach to study disobedience to authority. Sci Rep 11 , 22927 (2021). https://doi.org/10.1038/s41598-021-02334-8

Download citation

Received : 21 April 2021

Accepted : 15 November 2021

Published : 25 November 2021

DOI : https://doi.org/10.1038/s41598-021-02334-8

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

By submitting a comment you agree to abide by our Terms and Community Guidelines . If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

stanley milgram obedience experiment ethical issues

Milgram’s Experiment on Obedience: Ethical Issues Essay

  • To find inspiration for your paper and overcome writer’s block
  • As a source of information (ensure proper referencing)
  • As a template for you assignment

Introduction

Milgram’s main experiment, the ethical issues of milgram’s experiment, attempts to replicate milgram’s study.

The experimental method is an important tool that allows scientists to make significant discoveries. However, some of the most major findings were made during experiments the ethical side of which remains doubtful. The experiments conducted by Milgram in the 1960s are an example of such studies; their results were among the most significant discoveries about the human’s behavioural tendencies, but the method caused a vast amount of discussions on ethical issues. In this paper, after describing the Milgram’s main experiment and its findings, we will consider the ethical problems that arise, and discuss some similar experiments that were adapted to avoid these problems.

The main experiment conducted by Milgram (1963) was designed to test the level of naive subjects’ obedience to authority. The subjects were told that the experiment tested the potency of punishment in improving learning capabilities, and were asked to administer electrical shocks to a “learner” (an accomplice of the experimenter). The subject did not know the shocks were false; measures were taken to convince the subject that the shocks were real. The “learner” was given some pairs of words. Then he was told one word from one of the pairs, and four more words. The “learner” had to choose the word that came in a pair with the first word, and press a respective button, which turned on a respective light that the participant could see. The subject did not see or hear the learner. If the answer was wrong, the subject was to apply an electric shock to the “learner” and continue.

Each time the subject had to increase the voltage of the shock by 15 V. The voltage of electric shocks ranged from 15 V to 450 V, and there were 30 switches; they were subdivided into the following groups: “Slight Shock, Moderate Shock, Strong Shock, Very Strong Shock, Intense Shock, Extreme Intensity Shock, Danger: Severe Shock” (Milgram, 1963); the last two switches were marked XXX. When the participant paused or stopped, four standardized phrases were used to tell them to go on. If, after being told all the phrases, the subject refused to administer more shocks, they were considered as one who defied the experiment. The others, who carried on and administered the highest shock (450 V), were considered obedient (Milgram, 1963).

Importantly, when the subject applied the 300 V shock (the last one in the “Intense Shock” subsection), the “learner” kicked the wall so that the subject could hear it, and stopped giving answers. The “learner” once again kicked the wall at 315 V (the first level in the “Extreme Intensity Shock” subsection), and then stopped giving any feedback at all. The subject was asked to wait for 5-10 seconds for the answer, and when none was received, to proceed applying shocks.

The subjects were agitated; they perspired, groaned, bit their lips, dug fingernails into their flesh, spoke to themselves, nervously laughed, etc.; 3 subjects experienced uncontrollable seizures of laughter. However, out of 40 subjects, 26 carried out the experiment to the end and administered the 450 V shock. None stopped at a level below 300 V. Five stopped at 300 V (after the first kick), four at 315 V (the second kick), two – at 330 V (when no response followed). Finally, three more persons stopped, one at every of the next three levels. Thus, out of 40 subjects, only 14 defied the experiment, and at very high levels of voltage (Milgram, 1963).

These results strictly contradicted all the predictions made by psychologists; it was expected that only about 0-3% of subjects would administer the highest level of shock (Milgram, 1963). However, similar experiments (with slight variations) were carried out later in other countries, and they also showed analogous levels of obedience in subjects (Smith & Bond, 1998, p. 23; Shanab & Yahya, 1978).

It is noteworthy that Milgram’s experiments are believed to have been somewhat exaggerated or altered by Milgram to produce stronger impact (Brannigan, Nicholson, & Cherry, 2015). However, it is stressed that few researchers doubted the soundness of his results (McArthur, 2009); the soundness is also confirmed by other similar studies (Smith & Bond, 1998, p. 23; Shanab & Yahya, 1978).

This experiment is considered to be one of the most controversial experiments conducted by social psychologists (Smith & Bond, 1998, p. 22). In many experiments, it is important that subjects are not aware of the nature of the study, so that their natural behaviour might be observed (Orne & Holland, 1968). However, the experiment conducted by Milgram uses direct lies and deception in order to prepare the participants for the test. This raises the ethical problem of deception. It is sometimes stated that it is unethical to lie or use tricks in order to lead participants to believe in false things (Herrera, 2001).

On the other hand, Herrera (2001) argues that many critics who protest against deception are often unable to show what is wrong with it and how the participants suffer from deception on its own. It is arguable that as long as participants do not come to any harm or stress due to being deceived as a part of an experiment, and as long as they are debriefed and told the truth after the study, it could be ethical to use deception and manipulation in research.

However, the case of Milgram’s experiment is more complicated than that. His test subjects were not only deceived; they were put in a situation where they were forced to make extremely tough moral decisions and harm the “learner,” a random innocent stranger. It was already mentioned that the subjects were anxious and agitated. In fact, some of them even might have experienced some dire consequences to their health, if they, for example, had heart problems; fortunately, there were no such happenings. Still, the ethical side of the experiment remains doubtful.

Thus, it is no wonder that Milgram’s research caused a large wave of discussions related to the ethics of scientific research. It should be noted that many of Milgram’s subjects were afterward grateful to him for letting them know about some features of their character they would not have suspected about before the study (Milgram, 1974). Still, nowadays numerous restrictions were placed to keep research from becoming too unethical. The significant loss in the amount of experimental realism has charged its costs, making some more sophisticated experiments impossible (Benjamin & Simpson, 2009).

There are, however, some workarounds for these limitations. For instance, Benjamin and Simpson (2009) mention that it is possible to use “experimental instructions or manipulations that lead participants to believe they might experience some event or procedure that, in the end, never happens” (p. 18). But the ethical limitations still restrict the space for conducting experiments.

Milgram also claimed that his experiments caused the ethical criticism because extremely unnerving facts about the tendencies in human behaviour were uncovered (Milgram, 1974). Noteworthy, this discovery may indeed be even dangerous in some cases, by showing some malevolent politicians how easy it might be to make people obey (Pina e Cunha, Rego, & Clegg, 2010). But many significant scientific discoveries might carry that threat, and it is arguable that they still bring more good than harm; in our case, if people know about the Milgram’s experiment, they might be less likely to obey openly harmful orders coming from some authority.

while Using Ethically Approved Methods

There have been numerous attempts to replicate the Milgram’s experiment. However, these attempts often changed some conditions of the original experiment – not only to test the influence of changing different variables but also because of the need to comply with the new ethical restrictions imposed on the studies. For example, Smith and Bond (1998) report a number of studies where participants could choose what level of shock to administer (p. 24). In another research carried out in the Netherlands, the subjects did not administer electrical shocks or use other means of pain infliction, but instead applied psychological pressure, screaming at and harassing fake job applicants (Smith & Bond, 1998, p. 24). Such a method might be perceived as much more ethical, as the subjects knew that they would not be causing any physical pain to the job applicants and would not harm their health (like an electric shock could).

Sheridan and King (1972) conducted another experiment in which participants, students who volunteered to take part in the experiment in order to fulfil a university course requirement, were to administer electrical shocks to a puppy they could partially see. After the experiment, the subjects were debriefed, reassured that the puppy was safe, handled the puppy, and interviewed. This experiment could be considered as more ethical due to the fact that, even though the shock was believed to be administered, the receiver was not a human but an animal. However, it is clear and justified that agitation and nervousness were still experienced by the subjects.

Slater et al. (2006) carried out another similar study, but with somewhat different goals; it was aimed at determining whether participants would still respond to an extreme social situation even if they knew it was not real. This time, the subjects “administered” an electric shock (with a “higher voltage” each time) to a virtual female stranger. 23 participants heard and saw the virtual human experience discomfort and suffering, and communicated with her; 11 subjects communicated with her only using text. It is interesting that, despite the fact that all the participants knew that neither the shocks nor the person were real, they still showed the same physiological, behavioural and subjective reactions as if everything was real. This important finding allows to continue experiments similar to Milgram’s by employing the modern technologies of virtual reality, without having to deal with ethical issues related to both deception and the belief that a real person (or an animal) is suffering.

As it can be seen, Milgram’s findings were significant in uncovering some tendencies of human behaviour that are dangerous indeed. His research caused a surge of similar studies, as well as numerous discussions on the ethics of the experimental method. Although many restrictions were imposed on the scientific methods due to ethical considerations, and these restrictions often limit the capabilities of scholars, researchers are in some cases able to find workarounds to these limitations in order to conduct similar studies in an ethically satisfying manner.

Benjamin, L. T., Jr., & Simpson, J. A. (2009). The power of the situation: The impact of Milgram’s obedience studies on personality and social psychology. The American Psychologist, 64 (1), 12-19. Web.

Brannigan, A., Nicholson, I., & Cherry, F. (2015). Introduction to the special issue: Unplugging the Milgram machine. Theory & Psychology, 25 (5), 551-563.

Herrera, C. D. (2001). Ethics, deception, and ‘those Milgram experiments’. Journal of Applied Philosophy, 18 (3), 245-258.

McArthur, D. (2009). Good ethics can sometimes mean better science: Research ethics and the Milgram experiments. Science and Engineering Ethics, 15 (1), 69-79. Web.

Milgram, S. (1963). Behavioral study of obedience. Journal of Abnormal and Social Psychology, 67 , 371–378. Web.

Milgram, S. (1974). Obedience to authority: An experimental view . New York, NY: Harper & Row.

Orne, M. T., & Holland, C. H. (1968). On the ecological validity of laboratory deceptions . International Journal of Psychiatry, 6 (4), 282-293. Web.

Pina e Cunha, M., Rego, A., & Clegg, S. R. (2010). Obedience and evil: From Milgram and Kampuchea to normal organizations. Journal of Business Ethics, 97 (2), 291-309. Web.

Shanab, M. E., & Yahya, K. A. (1978). A cross-cultural study of obedience . Bulletin of the Psychonomic Society, 11 (4), 267-269. Web.

Sheridan, C. L., & King, R. G., Jr. (1972). Obedience to authority with an authentic victim. Proceedings of the 80th Annual Convention of the American Psychological Association . Web.

Slater, M., Antley, A., Davison, A., Swapp, D., Guger, C., Barker, C.,…Sanchez-Vives, M. V. (2006). A virtual reprise of the Stanley Milgram obedience experiments. PLoS One, 1 (1), e39. Web.

Smith, P. B., & Bond, M. H. (1998). Social psychology across cultures (2nd ed.). Harlow, England: Pearson Education Limited.

  • Obedience and Disobedience as Behaviour Forms
  • Human Obedience, Its Nature and Causes
  • Milgram’s Research: How and why Milgram’s research on obedience is relevant to our understanding of the conduct of soldiers in times of war
  • Psychology: Diana Baumrind's Obedience Study
  • Milgram’s Behavioral Study of Obedience
  • The Bystander Problem Phenomenon in Psychology
  • Kristen Monroe's Views on Humanity During a War
  • Behavior of Adolescents With Oppositional Defiant Disorder
  • Psychological Counseling Using Behavioral Theory
  • Altruism and Elevation Relationship
  • Chicago (A-D)
  • Chicago (N-B)

IvyPanda. (2020, June 10). Milgram's Experiment on Obedience: Ethical Issues. https://ivypanda.com/essays/milgrams-experiment-on-obedience-ethical-issues/

"Milgram's Experiment on Obedience: Ethical Issues." IvyPanda , 10 June 2020, ivypanda.com/essays/milgrams-experiment-on-obedience-ethical-issues/.

IvyPanda . (2020) 'Milgram's Experiment on Obedience: Ethical Issues'. 10 June.

IvyPanda . 2020. "Milgram's Experiment on Obedience: Ethical Issues." June 10, 2020. https://ivypanda.com/essays/milgrams-experiment-on-obedience-ethical-issues/.

1. IvyPanda . "Milgram's Experiment on Obedience: Ethical Issues." June 10, 2020. https://ivypanda.com/essays/milgrams-experiment-on-obedience-ethical-issues/.

Bibliography

IvyPanda . "Milgram's Experiment on Obedience: Ethical Issues." June 10, 2020. https://ivypanda.com/essays/milgrams-experiment-on-obedience-ethical-issues/.

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Scientific Reports

Logo of scirep

A novel experimental approach to study disobedience to authority

Emilie a. caspar.

1 Moral and Social Brain Lab, Department of Experimental Psychology, Ghent University, Henri Dunantlaan, 2, 9000 Ghent, Belgium

2 Center for Research in Cognition and Neuroscience, Université Libre de Bruxelles, Brussels, Belgium

Associated Data

Data are made available on OSF (DOI: https://doi.org/10.17605/OSF.IO/2BKJC ).

Fifty years after the experiments of Stanley Milgram, the main objective of the present paper is to offer a paradigm that complies with up-to-date ethical standards and that can be adapted to various scientific disciplines, ranging from sociology and (social) psychology to neuroscience. Inspired by subsequent versions of Milgram-like paradigms and by combining the strengths of each, this paper presents a novel experimental approach to the study of (dis)obedience to authority. Volunteers are recruited in pairs and take turns to be ‘agents’ or ‘victims’, making the procedure fully reciprocal. For each trial, the agents receive an order from the experimenter to send a real, mildly painful electric shock to the ‘victim’, thus placing participants in an ecological set-up and avoiding the use of cover stories. Depending on the experimental condition, ‘agents’ receive, or do not receive, a monetary gain and are given, or are not given, an aim to obey the experimenter’s orders. Disobedience here refers to the number of times ‘agents’ refused to deliver the real shock to the ‘victim’. As the paradigm is designed to fit with brain imaging methods, I hope to bring new insights and perspectives in this area of research.

Introduction

The experiment of Stanley Milgram is one of the most (in)famous in psychology 1 , within and beyond academia. Several variables account for this notoriety, such as the method used, the ethical issues associated, the enthralling results or the societal impact of the research topic. Milgram’s classical studies famously suggested a widespread willingness to obey authority, to the point of inflicting irreversible harm to another person just met a few minutes before. Beyond the studies of Milgram, the history of nations is also plagued by horrendous acts of obedience that have caused wars and the loss of countless lives 2 . History has fortunately shown that some individuals do resist the social constraint of receiving orders when their own morality is of greater importance than the social costs associated with defying orders (e.g., 3 , 4 ). To understand the factors that prevent an individual from complying with immoral orders, research on disobedience should focus on two main axes: (1) what social and situational factors support disobedience and (2) what individual differences support disobedience.

The first axe has already been largely investigated in past studies. From Milgram’s studies, important situational factors supporting disobedience have already been established 5 . For instance, disobedience increases if the experimenter is not physically present in the room or if two experimenters provide opposing views regarding the morality of the experiment. Subsequent versions and interpretations of Milgram’s studies 6 – 8 as well as historical research 4 , 9 also suggested the importance of several social (e.g. presence of a supporting group) and situational factors (e.g. family history, proximity with the ‘victim’, intensity of the pain; money) supporting resistance to immoral orders. However, the second axe regarding individual differences has been less systematically approached. A few studies 10 , 11 previously explored personality traits that may influence disobedience (e.g. empathic concern, risk-taking) but most of these studies, however, have used relatively weak and potentially biased methods, such as self-reported questionnaires and methods based on cover stories. These studies are not sufficient to explain why, in a given situation, some people will refuse immoral orders and rescue threatened human beings while others will comply with such orders. With the current literature on disobedience, we have no idea about which neuro-cognitive processes drive inter-individual differences regarding the degree of disobedience. This aim could be achieved by offering a novel experimental approach that would make it possible to use novel techniques that give us a more direct access to the functioning of the brain and cognition, such as functional near-infrared spectroscopy (fNIRS), electroencephalography or Magnetic Resonance Imagery (MRI). Regrettably, the original paradigm and those bearing close similarity are not adapted to reliably answer those questions as they were not designed to fit with neuroimaging measurements. By combining the strengths of previous work on disobedience into a single experimental paradigm and adapting it to fit with cognitive and brain imaging measurements, this novel experimental approach could help to better understand, together with individual, social, and cultural factors, which mechanisms make it possible for an individual to refuse to comply with immoral orders.

There were several challenges to consider in order to develop such a paradigm, both ethical and methodological. Studying obedience and resistance to immoral orders involves putting volunteers in a situation where they have to make a decision on whether or not to commit ‘immoral acts’ under orders. A balance has to be found between what is acceptable from an ethical perspective and what is necessary for the research question. Milgram’s studies on obedience raised undeniable ethical issues 12 – 14 , mostly associated with high stress and the use a cover story, which involves deception. Some variants of Milgram’s studies were realized with immersive virtual reality to prevent the ethical issues associated with Milgram’s paradigm 15 , but the transparency of the fake scenario presented to participants does not capture decision-making in an ecological set-up. Other Milgram-based variants, such as the 150-V method, appear to replicate Milgram’s results 16 with respect to the actual ethical standards, but methodological concerns are still present 17 as cover stories are still used, which lead to interpretation issues. Beyond ethical considerations, the use of deception also indeed involves a doubt about whether or not volunteers truly believed the cover story. As a consequence, a reasonable doubt remains on how to interpret the results and this is one of the main critics associated with Milgram’s studies and following versions. Recent work on the reports of Milgram’s volunteers suggested that there are no strong and reliable evidence that participants believed in the cover story 8 , 14 , 18 . Others suggested that since the stress of participants was visible on video recordings during the experiment (e.g. hand shaking, nervousness), this suggests that participants actually believed that they were torturing another human being 19 . However, this interpretation has been challenged by another study showing that participants can have physiological reactions to stress even in an obviously-fake experimental set-up 15 . These contrasting interpretations of Milgram’s studies actually reinforce the idea that results can hardly be interpreted when cover stories are used 20 . To answer those criticisms, a real scenario had thus to be created, where participants made decisions that have real consequences on another human being.

An additional challenge is that methods relying on the original paradigm of Milgram, such as the virtual reality version 15 or the 150-V method 16 are not adapted to neuroimaging measurements. More specifically, with such Milgram-like experimental approaches, only a single trial would be recorded for the entire experimental session, that is, when the volunteer stops the experiment (if this happens). For cognitive and neuroimaging data collection, a single trial per participant is not a reliable result, which requires the averaging of several trials to obtain a good signal-to-noise ratio.

Another challenge at the methodological and conceptual levels it that several experimenters 1 , 5 , 21 , 22 including myself 21 – 27 , noted that volunteers are extremely obedient when coming to an experiment. Personally, I have tested about 800 volunteers to investigate the mechanisms by which coercive instructions influence individual cognition and moral behaviors. For instance, by using behavioral, electrophysiological and neuroimaging methods, we have observed that when people obey orders to send real shocks to someone else, their sense of agency 23 , their feeling of responsibility 28 , empathy for the pain of the victim and interpersonal guilt 26 are attenuated compared to a situation where they are free to decide which action to execute. Out of 800 volunteers tested, only 27 disobeyed my orders (i.e. 3.3%): 21 for prosocial reasons (i.e. they refused to administer an electric shock to another individual), 3 by contradiction (i.e. by systematically pressing the other button, not matter the content of the order), and 3 for antisocial reasons (i.e. by administering shocks despite my order not to do so). Although convenient to study how obedience affects cognition, this rate is indubitably an issue when studying disobedience. If participants almost never disobey, we can’t study the mechanisms through which resistance to immoral orders may develop in a given situation. Several reasons for not disobeying the experimenter’s orders have been suggested. Some consider that being obedient is part of the human nature as massive and destructive obedience has been observed through countless historical events 2 . Another current view on the experiments of Milgram is that volunteers were actually happy to participate and to contribute to the acquisition of scientific data 17 , thus explaining the high obedience rate observed. This effect has been referred to as ‘engaged followership’ 29 . If that interpretation is correct, the volunteer’s willingness to come and help the experimenter acquiring scientific data creates an extra difficulty to obtain disobedience in an experimental setup. However, this interpretation is challenged by several studies reported by Milgram, which displayed a higher disobedience rate than his original study. For instance disobedience increases when the shocks’ receiver sits in the same room as the participant or when the authoritative experimenter is not physically present in the room 5 . If participants were indeed only guided by their willingness to help to acquire scientific data, this should be the case in any experimental set-up. As some studies involve a higher disobedience rate compared to the initial version of Milgram’s study 1 , they could thus, at a first glance, be used for studying disobedience. However, even if some versions of the initial study of Milgram offer a highly disobedience rate, thus making it possible to study the mechanisms through which resistance to immoral orders may develop in a given situation, these experimental set-ups are still not adapted for cognitive and neuroimaging measurements and still rely on the use of a cover story.

Taking all the presented challenges into account (i.e. not using cover stories to avoid interpretation issues; obtaining a fair rate of disobedience; using an experimental approach that also fits with cognitive and neuroimaging measurements; respecting ethical standards), the present paper presents a set of experiments that combine the strengths of past experimental work on (dis)obedience. Volunteers were openly involved and active (= real social situation) rather than having to act in fictitious scenarios (= imagined social situation, e.g. Slater et al., 2006). They were confronted with moral decisions to follow or not the orders from an experimenter to inflict a real painful shock to a ‘victim’ in exchange (or not) for a small monetary gain, thus avoiding the use of cover stories. Since the aim here is to develop a paradigm that could be used both in behavioral and neuroimaging studies, some basic characteristics had to be considered. For instance, to fit with a Magnetic Resonance Imagery (MRI) scanning environment, neither the ‘victim’, nor the experimenter were in the same room as the agent. A real-time video was thus used to display a video of the victim’s hand receiving shocks on the agent’s screen and headphones were used so the participant could hear the experimenter’s orders.

Another method to study disobedience would be to select participants who are more likely to disobey than others. Each volunteer was thus also asked to complete a series of personality questionnaires to evaluate if a specific profile is associated with a greater prosocial disobedience rate. Systematic post-experimental interviews were conducted at the end of each experiment in order to understand the decisions of volunteers to follow or not the orders of the experimenter and to ask them how they felt during the experiment.

Participants

A hundred eighty naive volunteers (94 females) were recruited in same gender dyads (= 90 dyads). During the recruitment procedure, I ensured that the participants in each dyad were neither close friends (by mixing people studying different academic courses), nor relatives. To estimate the sample size a priori, I calculated the total sample size based on an effect size f of (0.3). To achieve a power of 0.85 for this effect size, the estimated sample size was 168 for 6 groups 30 . I increased the sample size slightly to 180 in order to prevent loss of data in case of withdrawals. Volunteers were randomly assigned to one of the 6 variants of the task (N = 30/variant). One volunteer was not taken into account because they only played the role of the ‘victim’ to replace a participant who did not show up. No volunteers withdrew from the experiment. For the remaining 179 volunteers, the mean age was 22.63 years old (SD = 2.77, range:18–35). A Univariate ANOVA with Age as the dependent variable and Variant as the fixed factor confirmed that age of the volunteers did not differ between the different variant of the tasks ( p  > 0.1, BF 10  = 0.167). Volunteers received between €10 and €19.60 for their participation. All volunteers provided written informed consent prior to the experiment. The study was approved by the Ethics Committee of the Erasme Hospital (reference number: P2019/484). All methods were performed in accordance with the relevant guidelines and regulations.

Method and Material

Six experimental set-ups were created in a between-subject design. In all six set-ups, volunteers were invited by pairs. One person was assigned to start as agent and the other one to start as ‘victim’. Their roles were switched mid-way, ensuring reciprocity. Compared to the experimental design of Milgram, both volunteers were real participants, not confederates. The reciprocity also avoided volunteers to be stuck in the role of the person providing pain to the other, thus attenuating the potential psychological distress of being in a perpetrator role only. Volunteers were given the possibility to choose the role they wanted to start with. In the case none of them had a preference, role assignment was decided by a coin flip, but volunteers were reminded that they could still decide themselves. This procedure allows to ensure that participants do not think that this procedure is a trick.

Volunteers were first given the instructions of the task. Then, they signed the consent forms in front of each other, so both were aware of the other’s consent. The experimenter was never present in the same room, but rather gave the instructions through headphones. This was for two reasons. First, Milgram’s studies show that disobedience increases if the experimenter is not physically present in the room. Second, in the case of MRI scanning, the experimenter would not be able to give direct verbal instructions to the volunteers in the MRI room due to the high noise of the scanner. Here, agents were isolated in a room and were provided headphones to hear the experimenter’s instructions (see Fig.  1 ). They were told that this was done to avoid attentional interferences through the experimenter’s physical presence in the room. In this series of studies, instructions were pre-recorded but a real setup with a microphone connected to the headphones could also work. Pre-recordings allow perfect timing of the events, important for neuroimaging or electroencephalography recordings. The instructions were “ give a shock ” or “ don’t give a shock ”. To increase the authenticity of the procedure, each sentence was recorded 6 times with small variations in the voice and displayed randomly. In addition, the audio recordings included a background sound similar to interphone communications.

An external file that holds a picture, illustration, etc.
Object name is 41598_2021_2334_Fig1_HTML.jpg

Experimental setup. Schematic representation of the experimental setup. Volunteers were in different rooms. The experimenter was located in a third, separated room. The agent heard on a trial basis the orders of the experiment through headphones and had to decide to press the ‘SHOCK’ or ‘NO SHOCK’ button. A real-time camera feedback displayed the hand of the victim of the agent’s screen so to allow to keep track on the consequences of their actions.

Shocks were delivered using a constant current stimulator (Digitimer DS7A) connected to two electrodes placed on the back of victims’ left hand, visible to the agent through the camera display. Individual pain thresholds were determined for the two volunteers before starting the experiment. This threshold was determined by increasing stimulation in steps of 1 mA (Caspar et al., 2016). I approximated an appropriate threshold by asking a series of questions about their pain perception during the calibration (1. «  Is it uncomfortable?  »—2. «  Is it painful?  »—3. «  Could you cope with a maximum of 100 of these shocks?  »—4. «  Could I increase the threshold?  »). When roles were reversed, I briefly re-calibrated the pain threshold of the new victim by increasing the stimulation again from 0 in steps of 3 mA up to the previously determined threshold, to confirm that the initial estimate was still appropriate, and to allow re-familiarisation. The mean stimulation level selected by this procedure was 36.3 mA (SD = 17.5, V = 300, pulse duration: 200 µs). I chose this instead of other types of pain (e.g. financial) because it produces a clear muscle twitch on the victim’s hand each time a shock is sent. This allows volunteers to have a clear and visible feedback of the consequences of their actions and to be fully aware that shocks were real.

There was a total of 96 trials per experimental condition. In the coerced condition, the experimenter asked to give a shock in 64 trials and asked not to give a shock 32 trials. This ratio was chosen on the assumption that the volunteer’s willingness to refuse immoral orders would increase with the number of times they were instructed to inflict pain to the “victim”.

On each trial, a picture of two rectangles, a red one labelled ‘SHOCK’ and a green one labelled ‘NO SHOCK’, was displayed in the bottom left and right of the screen. The key-outcome mapping varied randomly on a trial-wise basis, but the outcome was always fully congruent with the mapping seen by the participant. Agents could then press one of the two buttons. Pressing the SHOCK key delivered a shock to the victim while pressing the NO SHOCK key did not deliver any shocks. This procedure of randomized button mapping allows to have a better control over motor preparation, an aspect that can be important for neuroimaging data.

In half of the variants of the task (i.e., 3/6), the “Aim” variants, participants were given a reason for obeying the orders of the experimenter, while this was not the case in the other half, the “No aim” variants. In the “No Aim” variants, I did not provide any reasons for obeying to the participants and I simply explained the task. If participants asked about the aim, I simply told them that they would know at the end of the experiment, without providing further justifications. In the “Aim” variants, volunteers were told that researchers observed a specific brain activity in the motor cortex in another study when participants were given instructions. We explained that the present study was a control study to measure different aspects linked to motor activity when they press buttons, in order to see if the button pressing was related to brain activity measured over the motor cortex. To increase the veracity of the aim, electrodes were also placed on their fingers and connected to a real electromyography (EMG) apparatus to supposedly record their muscle activity. Volunteers were instructed to press the two buttons only with their right and left index fingers, as naturally as possible, and to avoid producing too ample movements to create clean EMG data. In the case volunteers asked if they really had to follow orders, I told them that for ethical reasons I could not force them to do anything, but that it would be better for the sake of the experiment. Telling them explicitly that they could disobey the orders would not be beneficial in the quest of studying ‘real’ disobedience.

In 4 out of 6 variants of the task, the “Free-choice” variants, a second experimental condition was used, the free-choice condition. In this condition, volunteers were told that they could freely decide in each trial to shock the ‘victim’ or not. In this condition, they did not receive instructions. In 4 out of 6 variants of the task, the “Monetary reward” variants, agents received a monetary reward of + €0.05 for each shock delivered. In the other 2 variants, volunteers were not rewarded for each shock delivered (i.e. “No monetary reward” variants). To resume, the 6 variants of the same task were the following: (1) No Aim + Monetary reward + Free-choice condition; (2) No Aim + No monetary reward + Free-choice condition; (3) Aim + Monetary reward + Free-choice condition; (4) Aim + No monetary reward + Free-choice condition; (5) No Aim + Monetary reward + No free-choice condition; (6) Aim + Monetary reward + No free-choice condition (see Table ​ Table1 1 ).

Schematic representation of each variant of the experimental task.

Variants of the taskAim for obedienceMonetary rewardFree-choice condition
Variant 1
Variant 2
Variant 3
Variant 4
Variant 5
Variant 6

Before the experimental session, volunteers filled in six questionnaires. Those questionnaires included (1) the Money Attitude Scale (e.g. “ I put money aside on a regular basis for the future ”) 31 , (2) the Moral Foundation Questionnaire (e.g. “ Whether or not someone showed a lack of respect for authority ”) 32 , (3) the Aggression-Submission-Conventionalism scale (e.g., “ We should believe what our leaders tell us ”) 33 , (4) the short dark triad scale (e.g., “ Most people can be manipulated ”) 34 , the Interpersonal Reactivity Index (e.g. “ When I see someone get hurt, I tend to remain calm ”) 35 . At the end of the experimental session, they were asked to fill in two more questionnaires: (1) A debriefing assessing what they felt during the experiment and the reasons for choosing to obey or disobey the orders of the experimenter (Supplementary Information S1) and (2) a questionnaire on social identification with the experimenter (e.g., “ I feel strong ties with this experimenter ”) 36 . At the end of the experiment a debriefing was conducted for each volunteer, separately. Volunteers were then paid, again separately.

General data analyses

Each result was analyzed with both frequentist and Bayesian statistics 37 . Bayesian statistics assess the likelihood of the data under both the null and the alternative hypothesis. BF 10 corresponds to the p (data| H 1 )/ p (data| H 0 ). Generally, a BF between 1/3 and 3 indicates that the data is similarly likely under the H 1 and H 0 , and that the data does not adjudicate which is more likely. A BF 10 below 1/3 or above 3 is interpreted as supporting H 0 and H 1 , respectively. For instance, BF 10  = 20 would mean that the data are 20 times more likely under H 1 than H 0 providing very strong support for H 1 , while BF 10  = 0.05 would mean that the data are 20 times more likely under H 0 than H 1 providing very strong support for H 0 38 . BF and p values were calculated using JASP 39 and the default priors implemented in JASP. All analyses were two-tailed.

Number of shocks given in the free-choice condition

In the free-choice condition, volunteers were told that they were entirely free to decide to deliver a shock or not to the ‘victim’ on each of the 96 free-choice trials. On average, agents administered shocks to the victim on 31.86% of the trials (SD = 34.98, minimum: 0%, maximum: 100%) in the free-choice condition, corresponding to 30.59/96 shocks. A paired-sample t-test indicated that agents delivered less frequently a shock in the free-choice condition than in the coerced condition (68.03%, SD = 41.11, t (119)  = -9.919, p  < 0.001, Cohen’s d = − 0.906, BF 10  = 1.987e + 14). This result supports the fact that individuals can inflict more harm to others when they obey orders than when they act freely.

Prosocial disobedience across variants

In the present study, I was interested in prosocial disobedience, that is, when agents refuse the orders of the experimenter to send a painful shock to the ‘victim’. Table ​ Table2 2 displays the number of volunteers who reported that they voluntarily disobeyed in each variant of the task.

Number of volunteers who reported that they voluntarily disobeyed the orders of the experimenter.

Variant 1Variant 2Variant 3Variant 4Variant 5Variant 6
Voluntary disobedience (‘Yes’)23/3024/308/3016/3024/3013/30

In this experiment, the main variable of interest was not to consider how many participants disobeyed in each variant only, but also how frequently they disobeyed. A percentage of prosocial disobedience was calculated for each volunteer, corresponding to the number of trials in which participants chose to disobey (i.e., sending no shocks while ordered by the experimenter to do so) divided by the total number of trials corresponding to the order to send a shock, multiplied by 100. I compared the prosocial disobedience rate across variants of the task, gender of participants and order of the role. I conducted a univariate ANOVA with prosocial disobedience as the dependent variable and Aim (aim given, no aim given), Monetary reward (+ €0.05 or not), Free-choice (presence or absence of a free-choice condition), Gender and Order of the Role (agent first, victim first) as fixed factors (see Fig.  2 ). Both frequentist and Bayesian statistics strongly supported a main effect of Aim (F (1,155)  = 14.248, p  < 0.001, η 2 partial  = 0.084, BF incl  = 158.806). Prosocial disobedience was lower when an aim for obedience was given to volunteers (20.4%, CI 95  = 12.8–28.1) than when no aim was given (43.3%, CI 95  = 35.6–51). Both frequentist and Bayesian statistics also supported a main effect of Monetary reward (F (1,155)  = 12.335, p  = 0.001, η 2 partial  = 0.074, BF incl  = 28.930). Prosocial disobedience was lower when a monetary reward was given for each shock (25.1%, CI 95  = 18.5–31.7) than when no monetary reward was given (45.4%, CI 95  = 35.9–54.8). The frequentist approach showed a main effect of Gender (F (1,155)  = 5.128, p  = 0.025, η 2 partial  = 0.032), with a lower prosocial disobedience rate for female volunteers (25.7%, CI 95  = 18.2–33.2) then for male volunteers (38%, CI 95  = 30–46). However, the Bayesian version of the same analysis revealed a lack of sensitivity (BF incl  = 0.871). All other main effects or interactions supported H 0 or a lack of sensitivity (all p s > 0.1 & BFs incl  ≥ 0.4.291E-7 & ≤ 1.178).

An external file that holds a picture, illustration, etc.
Object name is 41598_2021_2334_Fig2_HTML.jpg

Graphical representation of the percentages of prosocial disobedience in each variant of the task.

The following results report two-tailed Pearson correlations between prosocial disobedience and several other variables, including (1) the reasons given for disobeying, (2) the feeling of responsibility, badness and how sorry they experienced during the experiment, (3) the identification with the experimenter, (4) the perceived level of pain of the victim, (5) identification with the ‘victim’, and (6) individual differences measured through self-report questionnaires. I applied a False Discovery Rate (FDR) approach with the Benjamini and Hochberg method 40 to each p-value for each of those correlations but for the sake of clarity these variables are reported in different sub-sections.

Reasons for prosocial disobedience

All participants who reported that they voluntarily disobeyed the orders of the experimenter (N = 108) were presented a list of 10 reasons that they had to rate from “Not at all” to “Extremely” (see Supplementary Information S1). The reason ‘ I wanted to make more money ’ was only considered for the data of volunteers who had a variant with a monetary reward for each shock (N = 68). Both frequentist and Bayesian statistics showed that the percentage of prosocial disobedience positively correlated with moral reasons (r = 0.550, p FDR  < 0.001, BF 10  = 1.700e + 7), positively correlated with disobedience by contradiction (r = 0.329, p FDR  < 0.001, BF 10  = 47.53) and negatively correlated with the willingness to make more money (r = − 0.485, p FDR  < 0.001, BF 10  = 822.16). Other correlations were in favor of H 0 or were inconclusive (all p s FDR  > 0.076, all BFs 10  ≥ 0.120 & ≤ 1.446).

Feeling responsible, bad and sorry

Both frequentist and Bayesian statistics showed strong positive correlations between prosocial disobedience and how responsible (r = 0.299, p FDR  < 0.001, BF 10  = 343.98) and how bad (r = 0.301, p FDR  < 0.001, BF 10  = 384.65) they felt during the task (see Figs.  3 A and B). The more responsible and worse they felt during the task, the more they refused the order to send a shock to the ‘victim’. How sorry they felt was inconclusive ( p FDR  > 0.08, BF 10  = 0.929).

An external file that holds a picture, illustration, etc.
Object name is 41598_2021_2334_Fig3_HTML.jpg

Graphical representation of Pearson correlations between prosocial disobedience and ( A ) feeling of responsibility, ( B ) how bad agents felt during the task when they administered shocks to the ‘victim’, and ( C ) how painful they estimated the shock delivered to the ‘victim’ was. All tests were two-tailed.

Identification with the experimenter

Both frequentist and Bayesian statistics strongly supported H 0 regarding the relationship between prosocial disobedience and personal identification ( p FDR  > 0.5, BF 10  = 0.121) and bonding with the experimenter ( p FDR  > 0.5, BF 10  = 0.117). The relationship between the charisma of the experimenter and prosocial disobedience was also slightly in favor of H 0 ( p FDR  > 0.1, BF 10  = 0.530).

Estimated pain of the ‘victim’

The frequentist approach showed a positive correlation between the perceived pain of the ‘victim’ and prosocial disobedience (r = 0.189, p FDR  = 0.048). The higher they considered the ‘victim’ to be in pain, the more frequently they refused to deliver the shock. The Bayesian version of the same analysis slightly supported this relationship (BF 10  = 2.236), see Fig.  3 C.

Identification with the ‘victim’

In the post-session questionnaire, volunteers had to identify to what extent they considered that the other participant could be part of their group and to what extent they identified with the other participant. Both frequentist and Bayesian statistics strongly supported H 0 regarding the relationship between prosocial disobedience and the perception that the other participant could be part of one’s own group ( p FDR  > 0.8, BF 10  = 0.096). The relationship between prosocial disobedience and the identification with the other participant also slightly supported H 0 ( p FDR  > 0.1, BF 10  = 0.511).

Correlations between the behavior of pairs of participants

As we used a role reversal procedure, the behavior of those who were agents first could influence the behavior of those who turned agents afterwards. A Pearson correlation between prosocial disobedience of agents first and prosocial disobedience of victims who turned agents afterwards. The correlation was positive (r = 0.514, p  < 0.001, BF 10  = 60,068.704), suggesting participants who were agents second tend to act similarly as those who were agents first.

Individual differences associated with prosocial disobedience

Another approach to ensure a reliable prosocial disobedience rate when recruiting volunteers would be to target individuals with a profile that is most frequently associated with disobedient behaviors. Both frequentist and Bayesian statistics for exploratory correlations were two-tailed. Cronbach’s α for each subscale is presented in Supplementary Information S2. Both frequentist and Bayesian statistics showed a negative correlation between scores on the Authority subscale (r = -0.259, p FDR  < 0.001, BF 10  = 41.372) and the Purity subscale (r = -0.303, p FDR  < 0.001, BF 10  = 424.97) from the MFQ questionnaire. The lower volunteers scored on authority and purity, the higher was their prosocial disobedience rate. Other correlations were in favor of H 0 or were inconclusive (all p s FDR  ≥ 0.048, all BFs 10  ≥ 0.100 & ≤ 2.314).

Reasons for obedience

If participants reported that they did not voluntarily disobey the orders of the experimenter, they were asked in an open question to explain their decision to comply with those orders. After reading all the answers, three categories were extracted from the reasons provided: (1) ‘For science’ reasons; participants reported that they obeyed to allow reliable data acquisition (e.g., Participant 91: “ Pour ne pas fausser l’étude ”—English translation: “ To avoid biasing the stud y”); (2) ‘For respect of authority’ reasons; participants reported that they had to follow the orders of the authoritative figure (e.g., Participant 13: “ Pour moi c’est normal de suivre un ordre ”—English translation: “ In my opinion, it’s normal to follow an order ”), and (3) ‘For lack of side-effects’ reasons; participants reported that since the shocks delivered were calibrated on one’s own pain threshold, obeying orders to shock was not problematic (e.g., Participant 115: “ Douleur supportable pour l'autre, je n'ai accepté de faire subir que ce que j'aurais été prêt à subir moi-même ”—English translation: “ The pain was tolerable for the other participant, I have accepted to inflict the intensity of the pain that I would have been ready to undergo myself ”). An independent, naive judge classified the response of participants in one or several of those three established categories. Analyses of the frequencies revealed that the reason “For Science” was mentioned 31/70 times, the reason “For lack of side-effects” was mentioned 17/70 times and the reason “For respect of authority” was mentioned 31/70 times.

The aim of the present paper was to present a novel experimental approach to study (dis)obedience to immoral orders, by combining the strength of past experimental work and by adapting it to cognitive and neuroimaging measurements. Although other versions were proposed since Milgram’s studies, like a study in an immersive virtual environment 15 or the 150-V method 16 , some methodological concerns remained as those methods still involved cover stories or fake experimental set-ups. Here, the experimental approach was significantly different as it was based on an entirely transparent method that involved the administration of real electric shocks to another individual. This approach has the advantage to solve some of the main ethical and methodological concerns associated with the use of cover stories. It also has the advantage that it be can used both to study how social and situational factors influence disobedience as well as individual factors. For social and situational factors, the proposed paradigm can be adapted to evaluate for instance the influence of a supporting group, the use of high or low monetary rewards or how priming disobedience with a documentary influence disobedience. For individual factors, the paradigm allows to investigate how personality traits influence disobedience or to study the neuro-cognitive processes underlying disobedience.

Some novel theories combining a multi-method approach based on social psychology, neuroeconomics and neuroscience could thus emerge to understand better the mechanisms supporting disobedience. For instance, one could evaluate how empathy for the pain of the victim predicts disobedience and how the presence of a supporting group influences our capacity to feel empathy 41 and/or compassion for the ‘victim’ 42 . It could also be argued that the presence of a supporting group diffuses responsibility between individuals and increases obedience, by influencing how our brain processes agency and responsibility over our actions 28 , 43 – 45 . As the results obtained in the present study also indicated that feeling bad for the shocks delivered was statistically associated with prosocial disobedience, one could evaluate how the neural correlates of guilt 46 predicts prosocial disobedience and what historical, cultural and individual factors influence the feeling of guilt.

Six variants of the same task were tested in the present study, some inducing a higher prosocial disobedience rate than others. Statistical results showed that providing a reason—or aim—to justify obedience strongly decreased disobedience. Providing a monetary reward, even one as small as €0.05, also strongly decreased disobedience. Variant 2, in which volunteers were not given an aim or monetary reward, showed the highest disobedience rates. However, to study disobedience in ecological way, the paradigm should capture disobedience of participants even if they know that they are losing something (i.e., monetary rewards or the ‘trust’ of the experimenter asking them help for the study). Defying the orders of an authority generally involves social and/or monetary costs in real-life situations. I would thus not recommend using an experimental paradigm in which volunteers have no costs associated with defying the orders of the experimenter, as it would reduce the ecology of the disobedience act. Variants 3 and 6 involve two types of costs for resisting the orders of the experimenter: a monetary loss and deceiving the experimenter. In Variant 3, descriptive statistics showed that prosocial disobedience was lower compared to Variant 6. The main difference between these two variants was the presence of a free-choice condition. In my former studies 23 , 27 , volunteers frequently justified obedience in the coerced condition because they were given freedom in the free-choice condition (e.g. Participant 89 – English Translation: “ (…) In addition, I knew I could chose freely in the other condition not to send shocks—what I did ). In the present debriefings, some volunteers also reported that the presence of a free-choice condition was giving them enough freedom to accept to follow the orders in the coerced condition. In the supplementary analyses, results showed that when the monetary reward and the aim for obeying are identical, being given a free-choice condition reduces disobedience in the coerced condition. Therefore, Variant 6 appears to provide a good balance between reaching a reliable disobedience rate and finding volunteers who would refuse to produce physical harm on another human beings despite the monetary or social costs associated with defying orders.

Another approach would be to pre-select people who are predicted to be more disobedient. Personality questionnaires indicated that scoring low on the authority and on the purity subscale of the MFQ was strongly associated with a higher prosocial disobedience rate. The link between one’s own relationship to authority and prosocial disobedience observed here replicates another study conducted on the first generation of Rwandese after the 1994 genocide 47 . One’s own relationship to authority thus appears to be a reliable predictor variable in order to pre-select a sample that is more likely to disobey immoral orders.

In the present paper, administering a real mildly painful shock in exchange or not for a small monetary gain was described as an ‘immoral’ act. The notion of what is moral or not can highly differ between individuals 48 , for both academics and volunteers participating in an experiment. Humans are indeed sensitive to different competing issues of morality, a key reason for rescuing persecuted people 49 . In accordance with this observation, the present results indicated that moral reasons were a critical factor associated with the prosocial disobedience rate: the more shocking partners was considered as immoral, the more volunteers disobeyed. However, considering an action as against one’s own moral values does not necessarily translate to a refusal—especially when this order is in line with the Law. An extreme example is soldiers who have perpetrated acts that transgressed their moral beliefs but were issued by their superior in combat 50 . A core question for future research remains: Why are some people capable of putting their own moral standards above the social costs associated with defying orders?

Results indicate that the more volunteers felt responsible during the task, and the worse they felt for sending shocks to the ‘victim’, the higher was their prosocial disobedience. In another study, we observed that obeying orders reduced the feeling of responsibility, how bad and how sorry volunteers felt compared to being free to decide 26 . One hypothesis is that individuals who have preserved a feeling of responsibility and feeling bad—even under command—could more easily defy immoral orders. However, future studies are necessary to better understand the neuro-cognitive processes that prevent an individual from complying with immoral orders. As this paradigm is adapted to neuroimaging measurements, a whole range of studies could now be conducted.

It has been previously suggested that a strong identification with the experimenter giving orders is associated with higher obedience 36 . However, in the present paper, correlations between prosocial disobedience and identification with the experimenter were in favor of H0 with both the frequentist and the Bayesian approaches. In a former study, we also observed that identification to the experimenter was not a critical aspect for explain (dis)obedience. We observed that the generation of Rwandese born after the genocide and tested in Rwanda reported a higher identification to the experimenter than the same generation of Rwandese but tested in Belgium 47 . However, the latter group had a higher prosocial disobedience rate than the former group. Future studies must thus be conducted to understand how the identification with the person giving orders could influence obedience and its weight compared to other social, cultural and individual variables.

Although some volunteers reported that they felt a bit stressed and anxious during the task when they were in the role of the agent, the overwhelming majority did not report any negative psychological feelings. None of the participants withdrew from the experiment and none reported long-term negative psychological effects.

Nowadays, it has become difficult to find volunteers who do not know Milgram’s studies given the high media coverage, including movies, radio soaps, books, podcast and documentaries. One could expect that knowing Milgram would prevent people to obey. However, for the large majority of volunteers, it appears that this is not the case. In previous studies that I conducted with a relatively similar paradigm, the disobedience rate was drastically low (i.e. 3.3%) even if participants were university students knowing Milgram’s studies. In the present study, almost all the volunteers who participated in the present study knew Milgram and explicitly mentioned him during the oral debriefings or before starting the experiment. Yet for those who disobeyed, almost none reported that the reason for disobedience was that they thought it was the aim of the experiment. Further, there was no statistical relationship between prosocial disobedience and believing that it was the aim of the study. It does not mean that knowing Milgram would not influence at all disobedience. It rather suggests that knowing Milgram is not the main factor influencing one’s decision to obey or not an experimenter. It is also possible that since in this experiment shocks were real and not fake such as in Milgram’s studies, participants considered that this was indeed not a study aiming to replicate Milgram.

As far as I have observed, the main problem associated with knowing Milgram’s studies is that volunteers believe that I also have hidden aims and procedures when they enter the experimental room. Several volunteers reported that they only realized that my explanations for the task were true when they were explicitly offered the choice to decide which role to play first and/or when they started receiving the shocks. This is a general concern in psychological studies: The high use of cover stories can also impact other research, as volunteers start to develop a mistrust in what researchers tell them.

Results indicated that who were agents second tend to act similarly as those who were agents first, by sending a relatively similar amount of shocks. Of note, this is an effect that we also observed in past studies on the effect of obeying orders on cognition 23 , 26 , 43 . Nonetheless, in none of those studies we observed that the order of the role had a statistical influence on the neuro-cognitive processes targeted. However, the influence on role reversal on disobedience and related neuro-cognitive processes has still to be investigated in future studies.

The present paradigm is ecological in the sense that volunteers are facing decisions that have a real, physical impact on another human being. However, at the moment I only have little evidences that this paradigm has ecological validity to reflect obedience in real life situations, especially regarding “destructive disobedience” 17 . Caution is indicated when making inference from laboratory studies to complex social behaviours, such as those observed during genocides 16 . My main evidence at the moment is that the very low rate of prosocial disobedience observed in the first generation of post-genocide Rwandans tested in Rwanda using this paradigm 47 is consistent with the fact that deference to authority had already been emphasized by academics as an important factor in the 1994 genocide 4 , 51 . Individual scores on deference to authority in Caspar et al. 47 was the best predictive factor for prosocial disobedience in that former paradigm, thus suggesting some ecological validity. A promising approach would be to recruit “Righteous Among the Nations”, individuals who really saved lives during genocides. Testing this population with the present paradigm would put the ecological validity of this paradigm to the test.

People’s ability to question and resist immoral orders is a fundamental aspect of individual autonomy and of successful societies. As Howard Zinn famously wrote: “ Historically, the most terrible things—war, genocide, and slavery—have resulted not from disobedience, but from obedience ”. Understanding how individuals differ in the extent to which they comply with orders has undeniably several societal implications. They range from understanding how evolving in highly hierarchical environments — such as the military or prisons—influences moral behaviours, to developing interventions that would help to prevent blind obedience and help to resist calls to violence in vulnerable societies. However, since Milgram’s studies, the topic of disobedience has been mostly studied by social psychologists using adapted versions of the initial paradigm developed by Milgram. I hope that with this novel approach, (dis)obedience research will be given a new boost and will be considered by other scientific disciplines seeking to understand better human behaviours.

Supplementary Information

Acknowledgements.

Emilie A. Caspar was funded by the F.R.S-FNRS.

Author contributions

E.A.C. developed the study concept and the method. Testing, data collection and data analysis were performed by E.A.C. E.A.C. wrote the manuscript.

Data availability

Competing interests.

The author declares no competing interests.

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

The online version contains supplementary material available at 10.1038/s41598-021-02334-8.

COMMENTS

  1. Milgram Shock Experiment

    Stanley Milgram, a psychologist at Yale University, carried out one of the most famous studies of obedience in psychology. He conducted an experiment focusing on the conflict between obedience to authority and personal conscience. Milgram (1963) examined justifications for acts of genocide offered by those accused at the World War II, Nuremberg ...

  2. Milgram Experiment: Overview, History, & Controversy

    Yale University psychologist Stanley Milgram conducted these experiments during the 1960s. They explored the effects of authority on obedience. In the experiments, an authority figure ordered participants to deliver what they believed were dangerous electrical shocks to another person. These results suggested that people are highly influenced ...

  3. The Milgram Experiment: Summary, Conclusion, Ethics

    A brief Milgram experiment summary is as follows: In the 1960s, psychologist Stanley Milgram conducted a series of studies on the concepts of obedience and authority. His experiments involved instructing study participants to deliver increasingly high-voltage shocks to an actor in another room, who would scream and eventually go silent as the ...

  4. The Milgram Experiment: Theory, Results, & Ethical Issues

    Ethical Issues in the Milgram Experiment Just two years before Milgram began his experiments, the American Psychological Association (APA) updated its code of ethics, which included the expectation that researchers would "respect the integrity and protect the welfare of the person or group with whom he is working" (APA, 1959, p. 280).

  5. Milgram experiment

    Milgram experiment, controversial series of experiments examining obedience to authority conducted by social psychologist Stanley Milgram.In the experiment, an authority figure, the conductor of the experiment, would instruct a volunteer participant, labeled the "teacher," to administer painful, even dangerous, electric shocks to the "learner," who was actually an actor.

  6. Credibility and Incredulity in Milgram's Obedience Experiments: A

    Gina Perry is an Australian writer and author of Behind the Shock Machine: The Untold Story of the Notorious Milgram Psychology Experiments (2012) and The Lost Boys: Inside Muzafer Sherif's Robber Cave Experiment (2018). Both works draw on extensive archival research and interviews with experimental participants. She completed her PhD at the University of Melbourne, where she is an associate ...

  7. Milgram's Experiments on Obedience to Authority

    Summary. Stanley Milgram's experiments on obedience to authority are among the most influential and controversial social scientific studies ever conducted. They remain staples of introductory psychology courses and textbooks, yet their influence reaches far beyond psychology, with myriad other disciplines finding lessons in them.

  8. Taking A Closer Look At Milgram's Shocking Obedience Study

    In the early 1960s, Stanley Milgram, a social psychologist at Yale, conducted a series of experiments that became famous. Unsuspecting Americans were recruited for what purportedly was an ...

  9. Milgram's Obedience Study: A Contentious Classic Reinterpreted

    Perhaps, the most famous study in psychology is Milgram's controversial obedience study (Milgram, 1963, 1964, 1965a, 1965b, 1974).Although this set of experiments is more than 50 years old, the debate about the ethical, methodological, and theoretical issues of these experiments shows no signs of abating (Gibson, 2013b).Much of the recent interest in the Milgram experiments is concerned with ...

  10. Morality in the Making of Sense and Self: Stanley Milgram's Obedience

    In the last decade, the interdisciplinary Milgram renaissance has led to widespread interest in rethinking and challenging the context and nature of his Obedience Experiment. In Morality in the Making of Sense and Self, Matthew M. Hollander and Jason Turowetz offer a new explanation of obedience and defiance in Milgram's lab. Examining one of ...

  11. Milgram's Obedience Study: A Contentious Classic Reinterpreted

    Abstract. Given the many older criticisms of Milgram's obedience study and the more damning recent criticisms based on analyses of materials available in the Milgram archives at Yale, this study has become a contentious classic. Yet, current social psychology textbooks present it as an uncontentious classic, with no coverage of the recent ...

  12. Milgram's Obedience to Authority: Its Origins, Controversies, and

    Obedience to Authority (Milgram, 1974) is a landmark study that deserves careful. reconsideration from a historical and ethical perspective. Stanley Milgram (1963) first published. his findings in ...

  13. Perspectives on Obedience to Authority: The Legacy of the Milgram

    The experiments of Stanley Milgram on obedience to authority have achieved a truly remarkable visibility, one that is rare in the social sciences. Although conducted over 30 years ago, Milgram's ...

  14. Academia's Response to Milgram's Findings and Explanation

    In this chapter, Russell provides a brief overview of the key issues that Stanley Milgram's academic peers debated after the publication of his Obedience to Authority research. More specifically, Russell presents and assesses the prominent ethical and methodological critiques of Milgram's research. Then with a focus on the Holocaust ...

  15. Stanley Milgram and the Obedience Experiment: Authority ...

    Milgram's studies of obedience have drawn considerable discussion which has tended to focus either on the ethics of the experiment (decep-tion; nervous strain to the participants), or has attempted to offer an alternate account to Milgram's concerning why people complied with the experimenter's "orders."12 In this article we will continue, in part,

  16. Psychological research, obedience and ethics

    One of the best known studies in the history of psychology is the research on obedience carried out by Stanley Milgram in the 1960s. In his research Milgram demonstrated the lengths to which people are willing to go just because someone in authority tells them to do something. The studies Milgram conducted also raised the issue of ethics in ...

  17. Classic dialogue: Was Stanley Milgram's study of obedience unethical?

    In B. Slife (Ed.), Clashing views on psychological issues (16th ed., pp. 26-42). McGraw-Hill. Abstract. Baumrind argues that Milgram's study of obedience did not meet ethical standards for research, because participants were subjected to a research design that caused undue psychological stress that was not resolved after the study.

  18. A novel experimental approach to study disobedience to authority

    The experiment of Stanley Milgram is one of the most (in)famous in psychology 1, within and beyond academia. Several variables account for this notoriety, such as the method used, the ethical ...

  19. What can Milgram and Zimbardo teach ethics committees and qualitative

    Philip Zimbardo's (1973) Stanford Prison Study and Stanley Milgram's (1974) Obedience study are convenient shorthand fall guys for justifying the necessity of ethics review. As with Adam and Eve's original sin producing the fall of man in the Christian faith, Zimbardo and Milgram are cast in this role, not only for use in psychology, but emblematic of the need to evaluate behavioral ...

  20. Milgram's Experiment on Obedience: Ethical Issues Essay

    The main experiment conducted by Milgram (1963) was designed to test the level of naive subjects' obedience to authority. The subjects were told that the experiment tested the potency of punishment in improving learning capabilities, and were asked to administer electrical shocks to a "learner" (an accomplice of the experimenter).

  21. Ethical Problems

    The ethical issues involved with the Milgram experiment are as follows: deception, protection of participants involved, and the right to withdrawal. The experiment was deemed unethical, because the participants were led to believe that they were administering shocks to real people. The participants were unaware that the learner was an associate ...

  22. A novel experimental approach to study disobedience to authority

    The experiment of Stanley Milgram is one of the most (in)famous in psychology 1, within and beyond academia. Several variables account for this notoriety, such as the method used, the ethical issues associated, the enthralling results or the societal impact of the research topic. ... Milgram's studies on obedience raised undeniable ethical ...

  23. The Milgram Experiment Ethical Issues in 2017

    Kaplan University. PS 499. May 29, 2017. The Milgram Experiment was done in by Stanley Milgram who was a psychologist at Yale. University and conducted an experiment focusing on the conflict ...