Jump to content

Brain-wash: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
→‎Video media: Correcting the Batman reference, the wrong Robin was used.
Line 223: Line 223:
* In Operation: D.A.T.E. from ''[[Codename: Kids Next Door]]'', the [[Delightful Children from Down the Lane]] brainwash people taking a picture into delightful zombies using a camera replaced with a delightfulization ray
* In Operation: D.A.T.E. from ''[[Codename: Kids Next Door]]'', the [[Delightful Children from Down the Lane]] brainwash people taking a picture into delightful zombies using a camera replaced with a delightfulization ray
* In ''[[Avatar: The Last Airbender]]'', the Dai Li and Long Feng use brainwashing to stop people from talking about the war with the Fire Nation.
* In ''[[Avatar: The Last Airbender]]'', the Dai Li and Long Feng use brainwashing to stop people from talking about the war with the Fire Nation.
* In the film ''[[Batman Beyond: Return of the Joker]]'', [[the Joker]] brainwashes [[Jason Todd]] into becoming his own demented Joker Junior.
* In the film ''[[Batman Beyond: Return of the Joker]]'', [[the Joker]] brainwashes [[Tim Drake]] into becoming his own demented Joker Junior.


== In video games ==
== In video games ==

Revision as of 19:06, 3 December 2007

Brainwashing (also known as thought reform or as re-education) consists of any effort aimed at instilling certain attitudes and beliefs in a person — sometimes unwelcome beliefs in conflict with the person's prior beliefs and knowledge.[1]

The American Psychological Association (APA) has neither officially accepted the concept nor officially rejected it (see below), but brainwashing has received more attention from the APA in recent years.[2]

Terminology

Origins of the term "brainwashing"

The term brainwashing first appeared in the English language relatively recently. Author John Marks writes that a journalist later revealed to have worked undercover for the Central Intelligence Agency (CIA)[3] first coined the term in 1950.

Earlier forms of coercive persuasion occurred during the Inquisition and in the course of show trials against "enemies of the state" in the Soviet Union, etc.; but no specific term emerged until the methodologies of these earlier movements became systematized during the early decades of the People's Republic of China for use in struggles against internal class enemies and foreign invaders. Until that time, presentations of the phenomenon described only concrete specific techniques.

The term xǐ năo (洗腦, the Chinese term literally translated as "to wash the brain") originally referred to methodologies of coercive persuasion used in the "reconstruction" (改造 gǎi zào) of the so-called feudal (封建 fēng jiàn) thought-patterns of Chinese citizens raised under pre-revolutionary régimes; the term punned on the Taoist custom of "cleansing/washing the heart" (洗心 xǐ xīn) prior to conducting certain ceremonies or entering certain holy places, and in Chinese, the word "心" xīn also refers to the soul or the mind, contrasting with the brain. The term first came into use in the United States in the 1950s during the Korean War (1950-1953) to describe those same methods as applied by the Chinese communists to attempt deep and permanent behavioral changes in foreign prisoners, and especially during the Korean War to disrupt the ability of captured United Nations troops to effectively organize and resist their imprisonment.

The word brainwashing consequently came into use in the United States of America to explain why, unlike in earlier wars, a relatively high percentage of American GIs defected to the enemy side after becoming prisoners of war in Korea. Later analysis determined that some of the primary methodologies employed on them during their imprisonment included sleep-deprivation and other intense psychological manipulations designed to break down the autonomy of individuals. American alarm at the new phenomenon of substantial numbers of U.S. troops switching their allegiance to support foreign Communists lessened after the repatriation of prisoners, when it emerged that few of them retained allegiance to the Marxist and "anti-American" doctrines inculcated during their incarcerations. When rigid control of information ceased and the former prisoners' "natural" cultural methods of reality testing could resume functioning, the superimposed values and judgments rapidly decreased.

Although the use of brainwashing on United Nations prisoners during the Korean War produced some propaganda-benefits to the forces opposing the United Nations, its main utility to the Chinese lay in the fact that it significantly increased the maximum number of prisoners that one guard could control, thus freeing other Chinese soldiers to go to the battlefield[citation needed].

After the Korean War the term "brainwashing" came to apply to other methods of coercive persuasion and even to the effective use of ordinary propaganda and indoctrination. Formal discourses of the Chinese Communist Party came to prefer the more clinical-sounding term sī xǐang gǎi zào 思想改造 ("thought reform"). Metaphorical uses of "brainwashing" extended as far as the memes of fashion-following.

Some people have come to use the terms "brainwashing" or "mind control" to explain the otherwise intuitively puzzling success of some fast-acting episodes of religious conversion or of recruitment of inductees into groups known variously as new religious movements or as cults. [4]

Compare the article on social influence for generic sociological/psychological approaches to the exercise of power.

See also:

One of the first published uses of the term thought reform occurred in the title of the book by Robert Jay Lifton (a professor of psychology and psychiatry at John Jay College and at the Graduate Center of the City University of New York): Thought Reform and the Psychology of Totalism: A Study of 'Brainwashing' in China (1961). (Lifton also testified on behavioral-change methodologies at the 1976 trial of Patty Hearst.) In his book Lifton used the term "thought reform" as a synonym for "brainwashing", though he preferred the first term. The elements of thought reform as published in that book sometimes serve as a basis for cult checklists, and read as follows:[5]

Benjamin Zablocki sees brainwashing as a "term for a concept that stands for a form of influence manifested in a deliberately and systematically applied traumatizing and obedience-producing process of ideological resocializations" and states this same concept historically also bore the names "thought reform" and "coercive persuasion".

Political brainwashing

Studies of the Korean War (1950-1953)

The Communist Party of China used the phrase "xǐ nǎo" ("wash brain", 洗脑) to describe its methods of persuading into orthodoxy those members who did not conform to the Party message. The phrase played on xǐ xīn (洗心"wash heart"), a monition — found in many Daoist temples — which exhorted the faithful to cleanse their hearts of impure desires before entering.

In September 1950, the Miami Daily News published an article by Edward Hunter (1902-1978) titled "'Brain-Washing' Tactics Force Chinese into Ranks of Communist Party". It contained the first printed use of the English-language term "brainwashing", which quickly became a stock phrase in Cold War headlines. Hunter, a CIA propaganda-operator[6] who worked under-cover as a journalist, turned out a steady stream of books and articles on the theme. An additional article by Hunter on the same subject appeared in New Leader magazine in 1951. In 1953 Allen Welsh Dulles, the CIA director at that time, explained that "the brain under [Communist influence] becomes a phonograph playing a disc put on its spindle by an outside genius over which it has no control."

In his 1956 book Brain-Washing: The Story of Men Who Defied It (Pyramid Books), Edward Hunter described "a system of befogging the brain so a person can be seduced into acceptance of what otherwise would be abhorrent to him". According to Hunter, the process became so destructive of physical and mental health that many of his interviewees had not fully recovered after several years of freedom from Chinese captivity.

Later, two studies[citation needed] of the Korean War defections by Robert Lifton and Edgar Schein concluded that brainwashing had a transient effect when used on prisoners-of-war. Lifton and Schein found that the Chinese did not engage in any systematic re-education of prisoners, but generally used their techniques of coercive persuasion to disrupt the ability of the prisoners to organize to maintain their morale and to try to escape. The Chinese did, however, succeed in getting some of the prisoners to make anti-American statements by placing the prisoners under harsh conditions of physical and social deprivation and disruption, and then by offering them more comfortable situations such as better sleeping quarters, quality food, warmer clothes or blankets. Nevertheless, the psychiatrists noted that even these measures of coercion proved quite ineffective at changing basic attitudes for most people. In essence, the prisoners did not actually adopt Communist beliefs. Rather, many of them behaved as though they did in order to avoid the plausible threat of extreme physical abuse. Moreover, the few prisoners influenced by Communist indoctrination apparently succumbed as a result of the confluence of the coercive persuasion, and of the motives and personality characteristics of the prisoners that already existed before imprisonment. In particular, individuals with very rigid systems of belief tended to snap and realign, whereas individuals with more flexible systems of belief tended to bend under pressure and then restore themselves after the removal of external pressures.

Working individually, Lifton and Schein discussed coercive persuasion in their published analyses of the treatment of Korean War POWs. They defined coercive persuasion as a mixture of social, psychological and physical pressures applied to produce changes in an individual's beliefs, attitudes, and behaviors. Lifton and Schein both concluded that such coercive persuasion can succeed in the presence of a physical element of confinement, "forcing the individual into a situation in which he must, in order to survive physically and psychologically, expose himself to persuasive attempts". They also concluded that such coercive persuasion succeeded only on a minority of POWs, and that the end-result of such coercion remained very unstable, as most of the individuals reverted to their previous condition soon after they left the coercive environment.

The use of coercive persuasion techniques in China

Following the armistice that interrupted hostilities in the Korean War, a large group of intelligence-officers, psychiatrists, and psychologists received assignments to debrief United Nations soldiers in the process of repatriation. The government of the United States wanted to understand the unprecedented level of collaboration, the breakdown of trust among prisoners, and other such indications that the Chinese were doing something new and effective in their handling of prisoners of war. Formal studies in academic journals began to appear in the mid-1950s, as well as some first-person reports from former prisoners. In 1961, two specialists in the field published books which synthesized these studies for the non-specialists concerned with issues of national security and social policy. Edgar H. Schein wrote on Coercive Persuasion, and Robert J. Lifton wrote on Thought Control and the Psychology of Totalism. Both books focussed primarily on the techniques called "xǐ nǎo" or, more formally "sī xiǎng gǎi zào" (reconstructing or remodeling thought). The following discussion largely builds on their studies.

Although American attention came to bear on thought reconstruction or brainwashing as one result of the Korean War (1950 - 1953), the techniques had operated on ordinary Chinese citizens after the establishment of the People's Republic of China (PRC) in October 1949. The PRC had refined and extended techniques earlier used in the Soviet Union to prepare prisoners for show-trials, and they in turn had learned much from the Inquisition[citation needed]. In the Chinese context, these techniques had multiple goals that went far beyond the simple control of subjects in the prison camps of North Korea. They aimed to produce confessions, to convince the accused that they had indeed perpetrated anti-social acts, to make them feel guilty of these crimes against the state, to make them desirous of a fundamental change in outlook toward the institutions of the new communist society, and, finally, to actually accomplish these desired changes in the recipients of the brainwashing/thought-reform. To that end, brainwashers desired techniques that would break down the psychic integrity of the individual with regard to information processing, with regard to information retained in the mind, and with regard to values. Chosen techniques included: dehumanizing of individuals by keeping them in filth, sleep deprivation, partial sensory deprivation, psychological harassment, inculcation of guilt, group social pressure, etc. The ultimate goal that drove these extreme efforts consisted of the transformation of an individual with a "feudal" or capitalist mindset into a "right-thinking" member of the new social system, or, in other words, to transform what the state regarded as a criminal mind into what the state could regard as a non-criminal mind.

The methods of thought-control proved extremely useful when deployed for gaining the compliance of prisoners-of-war. Key elements in their success included tight control of the information available to the individual and tight control over the behavior of the individual. When, after repatriation, close control of information ceased and reality testing could resume, former prisoners fairly quickly regained a close approximation of their original picture of the world and of the societies from which they had come. Furthermore, prisoners subject to thought-control often had simply behaved in ways that pleased their captors, without changing their fundamental beliefs. So the fear of brainwashed sleeper agents, such as that dramatized in the novel and the films The Manchurian Candidate, never materialized.

Terrible though the process frequently seemed to individuals imprisoned by the Chinese Communist Party, these attempts at extreme coercive persuasion ended with a reassuring result: they showed that the human mind has enormous ability to adapt to stress (not a recognized term in common use with reference to psychology in the early 1950s) and also a powerful homeostatic capacity. John Clifford, S.J. gives an account of one man's adamant resistance to brainwashing in In the Presence of My Enemies[7] that substantiates the picture drawn from studies of large groups reported by Lifton and Schein. Allyn and Adele Rickett wrote a more penitent account of their imprisonment (Allyn Rickett had by his own admission broken PRC laws against espionage) in "Prisoners of the Liberation",[8] but it too details techniques such as the “struggle groups” described in other accounts. Between these opposite reactions to attempts by the state to reform them, experience showed that most people would change under pressure and would change back following the removal of that pressure.[original research?] Interestingly, some individuals derived benefit from these coercive procedures due to the fact that the interactions, perhaps as an unintended side effect,[original research?] actually promoted insight into dysfunctional behaviors that the subjects then abandoned.[citation needed]

Criticism of claims of political brainwashing

According to research[citation needed] and forensic psychologist Dick Anthony, the CIA invented the concept of "brainwashing" as a propaganda strategy to undercut communist claims that American POWs in Korean communist camps had voluntarily expressed sympathy for communism. Anthony stated that definitive research demonstrated that fear and duress, not brainwashing, caused western POWs to collaborate. He argued that the books of Edward Hunter (a secret CIA "psychological warfare specialist" passing as a journalist) pushed the CIA brainwashing-theory onto the general public. He further asserted that for twenty years, starting in the early 1950s, the CIA and the Defense Department conducted secret research (notably including Project MKULTRA) in an attempt to develop practical brainwashing techniques (possibly to counteract the brainwashing efforts of the Chinese), and that their attempt failed.

Brainwashing in the context of new religious movements and cults

Frequent disputes regarding brainwashing take place in discussion of cults and of new religious movements (NRMs). The controversy about the existence of cultic brainwashing has become one of the most polarizing issues among cult-followers, academic researchers of cults, and cult-critics. Parties disagree about the existence of a social process attempting coercive influence, and also disagree about the existence of the social outcome — that people become influenced against their will.

The issue gets even more complicated due to the existence of several definitions of the term "brainwashing" (some of them almost strawman-caricature metaphors of the original Korean War era concept[9] ) and through the introduction of the similarly controversial concept of "mind control" in the 1990s. (In some usages "mind control" and "brainwashing" serve as exact synonyms; other usages differentiate the two terms.) Additionally, some authors refer to brainwashing as a recruitment method (Barker) while others refer to brainwashing as a method of retaining existing members (Kent 1997; Zablocki 2001).

Theories on brainwashing have also become the subject of discussions in legal courts, where experts have had to pronounce their views before juries in simpler terms than those used in academic publications and where the issue became presented in rather black-and-white terms in order to make a point in a case. The media have taken up some such cases — including their black and white colorings.

In 1984, the British sociologist Eileen Barker wrote in her book The Making of a Moonie: Choice or Brainwashing? (based on her first-hand studies of British members of the Unification Church) that she had found no extraordinary persuasion techniques used to recruit or retain members.

Charlotte Allen reported that "[i]n his article in Nova Religio, Zablocki was worried less about those academics who may stretch the brainwashing concept than about those, like Bromley, who reject it altogether. And in advancing his case, he took a hard look at such scholars’ intentions and tactics. (His title is deliberately provocative: 'The Blacklisting of a Concept: The Strange History of the Brainwashing Conjecture in the Sociology of Religion.')"[10] In his book Combatting Cult Mind Control American psychologist Steven Hassan describes the extraordinary persuasion technique that (in his opinion) members of the Unification Church used to accomplish his own recruitment and retention.

Philip Zimbardo writes that "[m]ind control is the process by which individual or collective freedom of choice and action is compromised by agents or agencies that modify or distort perception, motivation, affect, cognition and/or behavioral outcomes. It is neither magical nor mystical, but a process that involves a set of basic social psychological principles."(Zimbardo, 2002)

The APA, DIMPAC, and theories of brainwashing

In the early 1980s some U.S. mental-health professionals became prominent figures due to their involvement as expert witnesses in court-cases involving new religious movements. In their testimony they presented certain theories involving brainwashing, mind control, or coercive persuasion as concepts generally accepted within the scientific community. The American Psychological Association (APA) in 1983 asked Margaret Singer, one of the leading proponents of coercive persuasion theories, to chair a taskforce called the APA taskforce on Deceptive and Indirect Techniques of Persuasion and Control (DIMPAC) to investigate whether brainwashing or "coercive persuasion" did indeed play a role in recruitment by such movements. Before the taskforce had submitted its final report, the APA submitted on February 10, 1987 an amicus curiæ brief in an ongoing case. The brief stated that:

[t]he methodology of Drs. Singer and Benson has been repudiated by the scientific community, that the hypotheses advanced by Singer were little more than uninformed speculation, based on skewed data and that "[t]he coercive persuasion theory ... is not a meaningful scientific concept. [...] The theories of Drs. Singer and Benson are not new to the scientific community. After searching scrutiny, the scientific community has repudiated the assumptions, methodologies, and conclusions of Drs. Singer and Benson. The validity of the claim that, absent physical force or threats, "systematic manipulation of the social influences" can coercively deprive individuals of free will lacks any empirical foundation and has never been confirmed by other research. The specific methods by which Drs. Singer and Benson have arrived at their conclusions have also been rejected by all serious scholars in the field.[11]

The brief characterized the theory of brainwashing as not scientifically proven and suggested the hypothesis that cult recruitment techniques might prove coercive for certain sub-groups, while not affecting others coercively. On March 24, 1987, the APA filed a motion to withdraw its signature from this brief, as it considered the conclusion premature, in view of the ongoing work of the DIMPAC taskforce.[12] The amicus as such remained, as only the APA withdraw the signature, but not the co-signed scholars (including Jeffrey Hadden, Eileen Barker, David Bromley and J. Gordon Melton). On May 11, 1987, the APA Board of Social and Ethical Responsibility for Psychology (BSERP) rejected the DIMPAC report because the brainwashing theory espoused "lacks the scientific rigor and evenhanded critical approach necessary for APA imprimatur", and concluded "Finally, after much consideration, BSERP does not believe that we have sufficient information available to guide us in taking a position on this issue."

With the rejection-memo came two letters from external advisers to the APA who reviewed the report. One of the letters, from Professor Benjamin Beit-Hallahmi of the University of Haifa, stated amongst other comments that "lacking psychological theory, the report resorts to sensationalism in the style of certain tabloids" and that "the term 'brainwashing' is not a recognized theoretical concept, and is just a sensationalist 'explanation' more suitable to 'cultists' and revival preachers. It should not be used by psychologists, since it does not explain anything". Professor Beit-Hallahmi asked that the report not be made public. The second letter, from Professor of Psychology Jeffrey D. Fisher, Ph.D., said that the report "[...] seems to be unscientific in tone, and biased in nature. It draws conclusions, which in many cases do not mesh well with the evidence presented. At times, the reasoning seems flawed to the point of being almost ridiculous. In fact, the report sometimes seems to be characterized by the use of deceptive, indirect techniques of persuasion and control — the very thing it is investigating".[13]

When the APA's BSERP rejected her findings, Singer sued the APA in 1992 for "defamation, frauds, aiding and abetting and conspiracy"; and lost in 1994.[14]

Some scholars sympathetic to NRMs have since interpreted these events to imply that the APA had rejected the brainwashing theories and that Singer's ideas lacked scientific support (for example: Introvigne, 1998[citation needed]; Bromley and Hadden in their 1993 Handbook of Cults and Sects in America.)

Zablocki (1997) and Amitrani (2001) cite APA boards and scholars on the subject and conclude that the APA has made no unanimous decision regarding this issue. They also write that Margaret Singer, despite the rejection of the DIMPAC report, continued her work and retained respect in the psychological community, which they corroborate by mentioning that in the 1987 edition of the peer-reviewed Merck's Manual, Margaret Singer wrote the article "Group Psychodynamics and Cults" (Singer, 1987).

Benjamin Zablocki, professor of sociology and one of the reviewers of the rejected DIMPAC report, wrote in 1997:

"Many people have been misled about the true position of the APA and the ASA with regard to brainwashing. Like so many other theories in the behavioral sciences, the jury is still out on this one. The APA and the ASA acknowledge that some scholars believe that brainwashing exists but others believe that it does not exist. The ASA and the APA acknowledge that nobody is currently in a position to make a Solomonic decision as to which group is right and which group is wrong. Instead they urge scholars to do further research to throw more light on this matter. I think this is a reasonable position to take."[citation needed]

APA Division 36 (then "Psychologists Interested in Religious Issues", today "Psychology of Religion") in its 1990 annual convention approved the following resolution:

"The Executive Committee of the Division of Psychologists Interested in Religious Issues supports the conclusion that, at this time, there is no consensus that sufficient psychological research exists to scientifically equate undue non-physical persuasion (otherwise known as "coercive persuasion", "mind control", or "brainwashing") with techniques of influence as typically practiced by one or more religious groups. Further, the Executive Committee invites those with research on this topic to submit proposals to present their work at Divisional programs." (PIRI Executive Committee Adopts Position on Non-Physical Persuasion Winter, 1991, in Amitrano and Di Marzio, 2001)

In 2002, APA's then president, Philip Zimbardo wrote in Psychology Monitor:

"A body of social science evidence shows that when systematically practiced by state-sanctioned police, military or destructive cults, mind control can induce false confessions, create converts who willingly torture or kill "invented enemies," engage indoctrinated members to work tirelessly, give up their money—and even their lives—for "the cause." (Zimbardo, 2002)

Other viewpoints

Two months after her kidnap in 1974, Patty Hearst, an American newspaper heiress, participated in a bank-robbery with her kidnappers. In her trial, the defense postulated a concerted brainwashing program as central. Despite this claim, the court convicted her of bank-robbery.

In the 1990 U.S. v. Fishman Case, Steven Fishman offered a "brainwashing" defense to charges of embezzlement. Margaret Singer and Richard Ofshe would have appeared as expert witnesses for him. The court disallowed the introduction of Singer and Ofshe's testimony:[15]

"The evidence before the Court, which is detailed below, shows that neither the APA nor the ASA has endorsed the views of Dr. Singer and Dr. Ofshe on thought reform ... At best, the evidence establishes that psychiatrists, psychologists and sociologists disagree as to whether or not there is agreement regarding the Singer-Ofshe thesis. The Court therefore excludes defendants' proffered testimony (U.S. vs. Fishman, 1989)."

Social scientists who study new religious movements, such as Jeffrey K. Hadden (see References), understand the general proposition that religious groups can have considerable influence over their members, and that that influence may have come about through deception and indoctrination. Indeed, many sociologists observe that "influence" occurs ubiquitously in human cultures, and some argue that the influence exerted in "cults" or new religious movements does not differ greatly from the influence present in practically every domain of human action and of human endeavor.

The Association of World Academics for Religious Education states that "... without the legitimating umbrella of brainwashing ideology, deprogramming — the practice of kidnapping members of NRMs and destroying their religious faith — cannot be justified, either legally or morally."[citation needed]

F.A.C.T.net states that "Forced deprogramming was sometimes successful and sometimes unsuccessful, but is not considered an acceptable, legal, or ethical method of rescuing a person from a cult."[16]

The American Civil Liberties Union (ACLU) published a statement in 1977 related to brainwashing and mind control. In this statement the ACLU opposed certain methods "depriving people of the free exercise of religion". The ACLU also rejected (under certain conditions) the idea that claims of the use of "brainwashing" or of "mind control" should overcome the free exercise of religion. (See quote)

In the 1960s, after coming into contact with new religious movements (NRMs, a subset of which have gained the popular designation of "cults"), some young people suddenly adopted faiths, beliefs, and behavior that differed markedly from their previous lifestyles and seemed at variance with their upbringings. In some cases, these people neglected or even broke contact with their families. All of these changes appeared very strange and upsetting to their families. To explain these phenomena, some postulated brainwashing on the part of new religious movements. Observers quoted practices such as isolating recruits from their family and friends (inviting them to an end-of-term camp after university for example), arranging a sleep-deprivation program (3 a.m. prayer meetings) and exposing them to loud and repetitive chanting. Another alleged technique of religious brainwashing involved love bombing rather than torture.

James T. (Jim) Richardson, a Professor of Sociology and Judicial Studies at the University of Nevada, states that if the NRMs had access to powerful brainwashing techniques, one would expect that NRMs would have high growth-rates, while in fact most have not had notable success in recruitment, most adherents participate for only a short time, and such groups have limited success in retaining members. Langone has rejected this claim, comparing the figures of various movements, some of which do (by common consent) not use brainwashing and others of which some authors report as using brainwashing. (Langone, 1993)

In their Handbook of Cults and Sects in America, Bromley and Hadden present one possible ideological foundation of brainwashing theories that they state demonstrates the lack of scientific support: they argue that a simplistic perspective (one they see as inherent in the brainwashing metaphor) appeals to those attempting to locate an effective social weapon to use against disfavored groups, and that any relative success of such efforts at social control should not detract from any lack of scientific basis for such opinions.

Philip Zimbardo, Professor Emeritus of Psychology at Stanford University, writes: "Whatever any member of a cult has done, you and I could be recruited or seduced into doing — under the right or wrong conditions. The majority of 'normal, average, intelligent' individuals can be led to engage in immoral, illegal, irrational, aggressive and self destructive actions that are contrary to their values or personality — when manipulated situational conditions exert their power over individual dispositions."(Zimbardo, 1997)

Some religious groups, especially those of Hindu and Buddhist origin, openly state that they seek to improve what they call the "natural" human mind by spiritual exercises. Intense spiritual exercises have an effect on the mind, for example by leading to an altered state of consciousness. These groups also state that they do not [condone the] use [of] coercive techniques to acquire or to retain converts. [citation needed]

On the other hand, several scholars in sociology and psychology have in recent years stated that many scholars of NRMs express a bias to deny any possibility of brainwashing and to disregard actual evidence. (Zablocki 1997, Amitrani 1998, Kent 1998, Beit-Hallahmi 2001)

Psychologist Steven Hassan, author of the book Combatting Cult Mind Control, has suggested that the influence of sincere but misled people can provide a significant factor in the process of thought-reform. Many scholars in the field of new religious movements do not accept Hassan's BITE model for understanding cults.

Brainwashing in fiction

Video media

In video games

  • In the video game Psychonauts, Boyd Cooper, the security-guard at Thorney Towers, undergoes hypnosis and has a second personality (dubbed "The Milkman") implanted into his mind, which certain actions or commands can trigger.
  • In Half-Life 2, the Combine race uses brainwashing on humans to produce soldiers and CP units. They extract organs (brainwashed brains) from humans to create synths.
  • In Quake 4, the Strogg race "brainwashes" the humans by activating the neutrocyte (mind-control chip), thus fully "Stroggifying" them.
  • In a Captain N the Gamemaster where Simon Belmont suffers temporary amnesia, King Hippo and Eggplant Wizard are ordered by Mother Brain to brainwash Simon into becoming an enemy of the N-Team, but not before they start scrubbing Mother Brain's glass casing as they misunderstand the meaning of the word "brainwash".

References

  • Amitrani, Alberto et al.: Blind, or just don't want to see? "Brainwashing", mystification and suspicion, 1998,
  • Amitrani, Alberto et al.: Blind, or just don't want to see? ""Mind Control" in New Religious Movements and the American Psychological Association, 2001, Cultic Studies Review [1]
  • Anthony, Dick. 1990. "Religious Movements and 'Brainwashing' Litigation" in Dick Anthony and Thomas Robbins, In Gods We Trust. New Brunswick, NJ: Transaction. Excerpt
  • APA Amicus curiae, February 11, 1987 [2]
  • APA Motion to withdraw amicus curiae March 27, 1987[3]
  • APA Board of Social and Ethical Responsibility for Psychology, Memorandum on Brainwashing: Final Report of the Task Force, May 11, 1987 [4]
  • Bardin, David, "Mind Control ("Brainwashing") Exists, in Psychological Coercion & Human Rights, April 1994, [5]
  • Benjamin Beith-Hallahmi: Dear Colleagues: Integrity and Suspicion in NRM Research, 2001 http://www.apologeticsindex.org/c59.html
  • David Bromley, "A Tale of Two Theories: Brainwashing and Conversion as Competing Political Narratives" in Benjamin Zablocki and Thomas Robbins (editors), Misunderstanding Cults, 2001, ISBN 0-8020-8188-6
  • Hadden, Jeffrey K., "The Brainwashing Controversy", November 2000
  • Hadden, Jeffery K., and Bromley, David, eds. (1993), The Handbook of Cults and Sects in America. Greenwich, CT: JAI Press, Inc., pp. 75-97
  • Hassan, Steven Releasing The Bonds: Empowering People to Think for Themselves, 2000. ISBN 0-9670688-0-0.
  • Hindery, Roderick, Indoctrination and Self-deception or Free and Critical Thought? Lewiston, N.Y.: E. Mellen Press, 2001. ISBN 0773474072
  • Huxley, Aldous, Brave New World Revisited. Perennial (2000); ISBN 0-06-095551-1
  • Introvigne, Massimo, “Liar, Liar”: Brainwashing, CESNUR and APA, 1998 [6]
  • Kent, Stephen A., Brainwashing in Scientology's Rehabilitation Project Force (RPF)", November 7, 1997
  • Stephen A. Kent and Theresa Krebs: "When Scholars Know Sin", Skeptic Magazine (Vol. 6, No. 3, 1998).
  • Kent, Stephen A.: Brainwashing Programs in The Family/Children of God and Scientology , in Benjamin Zablocki and Thomas Robbins (ed.), Misunderstanding Cults, 2001, ISBN 0-8020-8188-6
  • Langone, Michael D, ed.: Recovery from cults : help for victims of psychological and spiritual abuse. New York: W.W. Norton, 1993. ISBN 0393701646 , ISBN 0-393-31321-2
  • Robert J. Lifton, Thought Reform and the Psychology of Totalism (1961), ISBN 0-8078-4253-2
  • Marks, John , "The Search for the Manchurian Candidate", 1978 [7]
  • Richardson, James T., "Brainwashing Claims and Minority Religions Outside the United States: Cultural Diffusion of a Questionable Concept in the Legal Arena", Brigham Young University Law Review circa 1994
  • Scheflin, Alan W and Opton, Edward M. Jr., The Mind Manipulators. A Non-Fiction Account. New York: Paddington Press, 1978, p. 437. ISBN 0448229773
  • Schein, Edgar H. et al., Coercive persuasion;: A socio-psychological analysis of the "brainwashing" of American civilian prisoners by the Chinese Communists New York: W. W. Norton, 1961
  • Shapiro, K. A. et al, "Grammatical distinctions in the left frontal cortex", J. Cogn. Neurosci. 13, pp. 713-720 (2001). [8]
  • Singer, Margaret "Group Psychodynamics", in Merck's Manual, 1987.
  • Wakefield, Hollida, M.A. and Underwager, Ralph, Ph.D., Coerced or Nonvoluntary Confessions, Institute for Psychological Therapies, 1998
  • West, Louis J., "Persuasive Techniques in Religious Cults, 1989
  • Zablocki, Benjamin: The Blacklisting of a Concept: The Strange History of the Brainwashing Conjecture in the Sociology of Religion. Nova Religio, October 1997, Vol. 1, No. 1: 96-121.
  • Zablocki, Benjamin, Towards a Demystified and Disinterested Scientific Theory of Brainwashing, in Benjamin Zablocki and Thomas Robbins (ed.), Misunderstanding Cults, 2001, ISBN 0-8020-8188-6
  • Zablocki, Benjamin, "Methodological Fallacies in Anthony's Critique of Exit Cost Analysis", ca. 2002,
  • Zimbardo, Philip, What messages are behind today's cults? in Monitor on Psychology, May 1997
  • Zimbardo, Philip, Mind Control: Psychological Reality or Mindless Rhetoric? in Monitor on Psychology, November 2002

Bibliography / Further Reading

  • Anthony, Dick, Brainwashing and Totalitarian Influence. An Exploration of Admissibility Criteria for Testimony in Brainwashing Trials, Ph.D. Diss., Berkeley (California): Graduate Theological Union, 1996, p. 165.
  • Barker, Eileen, The Making of a Moonie: Choice or Brainwashing, Oxford, UK : Blackwell Publishers, 1984 ISBN 0-631-13246-5
  • Committee on Un-American Activities (HUAC), Communist Psychological Warfare (Brainwashing), United States House of Representatives, Washington, D. C., Tuesday, March 13, 1958
  • Hassan, Steven. Releasing The Bonds: Empowering People to Think for Themselves, 2000. ISBN 0-9670688-0-0.
  • Hunter, Edward, Brain-Washing in Red China. The Calculated Destruction of Men’s Minds, New York: The Vanguard Press, 1951; 2nd expanded ed.: New York: The Vanguard Press, 1953
  • Robert J. Lifton, Thought Reform and the Psychology of Totalism (1961), ISBN 0-8078-4253-2
  • Sargant, William, Battle for the Mind: A Physiology of Conversion and Brainwashing, 1996, ISBN 1-883536-06-5
  • Streatfeild, Dominic, Brainwash: The Secret History of Mind Control, 2006, ISBN 0-340-92103-X
  • Taylor, Kathleen, Brainwashing: The Science Of Thought Control, 2005, ISBN 0-19-280496-0
  • Benjamin Zablocki and Thomas Robbins (editors), Misunderstanding Cults, 2001, ISBN 0-8020-8188-6
  • Philip Zimbardo, "Mind control: psychological reality or mindless rhetoric?" Monitor on Psychology, Volume 33, No. 10 November 2002

See also

Footnotes

  1. ^ Compare: Dorland, Newman W. Dorland's Illustrated Medical Dictionary. 29th. edition. Philadelphia, Saunders, 2000.
  2. ^ Dittmann, Melisa, Cults of Hatred: Panelists at a convention session on hatred asked APA to form a task force to investigate mind control among destructive cults., Volume 33, No. 10, November 2002, Melissa Dittmann, pg. 30, American Psychological Association, Monitor, "Available online"
  3. ^ Marks, John. The Search for the "Manchurian Candidate": The CIA and Mind Control. New York: McGraw-Hill, 1980.
  4. ^ Eileen Barker explains the attractions for observers of explaining — using the concept of "brainwashing" — the behavior of those who join new religious movements. See Barker, Eileen: New Religious Movements: A Practical Introduction. London: Her Majesty's Stationery office, 1989.
  5. ^ http://www.reveal.org/library/psych/lifton.html http://www.rickross.com/reference/brainwashing/brainwashing19.html
  6. ^ http://www.druglibrary.org/schaffer/lsd/marks8.htm
  7. ^ Clifford, John W, In the Presence of My Enemies. New York: Norton, 1963.
  8. ^ W Allyn Rickett and Adele Rickett: Prisoners of liberation. New York, Cameron Associates, 1957.
  9. ^ The American Heritage Dictionary of the English Language (Fourth Edition, 2000), for example, records advertising as an example of a type of brainwashing. Online at http://www.bartleby.com/61/1/B0450100.html, retrieved 2007-09-02.
  10. ^ Charlotte Allen, "Brainwashed! Scholars of Cults Accuse Each Other of Bad Faith", Lingua Franca, December 1998. Online at http://www.rickross.com/reference/apologist/apologist29.html - retrieved 2007-03-25
  11. ^ http://www.cesnur.org/testi/molko_brief.htm
  12. ^ http://www.rickross.com/reference/apologist/apologist25.html
  13. ^ APA memo and two enclosures
  14. ^ Case No. 730012-8 Margaret Singer v. American Psychological Association
  15. ^ Brainwashed! Scholars of Cults Accuse Each Other of Bad Faith, Lingua Franca, December 1998.
  16. ^ Use of Forced Deprogramming F.A.C.T.net