Propaganda Persuasion: Difference between revisions
BloomWiki: Propaganda Persuasion |
BloomWiki: Propaganda Persuasion |
||
| Line 89: | Line 89: | ||
== Evaluating == | == Evaluating == | ||
Evaluating a message: (1) '''Source''': Who is telling me this, and what do they want? (2) '''Emotional Trigger''': Is this making me feel "Angry" or "Afraid" (a sign of manipulation)? (3) '''The 'Missing' Info''': What is this story | Evaluating a message: (1) '''Source''': Who is telling me this, and what do they want? (2) '''Emotional Trigger''': Is this making me feel "Angry" or "Afraid" (a sign of manipulation)? (3) '''The 'Missing' Info''': What is this story ''not'' telling me? (4) '''Fact-Checking''': Can I find this same information on three different, unrelated websites? | ||
== Creating == | == Creating == | ||
Revision as of 14:29, 23 April 2026
How to read this page: This article maps the topic from beginner to expert across six levels � Remembering, Understanding, Applying, Analyzing, Evaluating, and Creating. Scan the headings to see the full scope, then read from wherever your knowledge starts to feel uncertain. Learn more about how BloomWiki works ?
Propaganda and Persuasion are the techniques used to influence the attitudes, beliefs, and behaviors of a group of people. While "Persuasion" is often seen as a neutral or positive act (like a debate or a commercial), "Propaganda" is usually associated with the organized, often misleading use of media by governments or groups to promote a specific ideology. In the 21st century, these techniques have moved from "Posters and Radio" to "Micro-targeted AI Ads." By understanding the psychological tools of persuasion—like repetition, authority, and emotional appeal—we can become more "Media Literate" and protect our own critical thinking from manipulation.
Remembering
- Propaganda — Information, especially of a biased or misleading nature, used to promote or publicize a particular political cause or point of view.
- Persuasion — The action or fact of persuading someone or of being persuaded to do or believe something.
- Cognitive Dissonance — The mental discomfort felt when holding two conflicting beliefs; propagandists use this to "force" a shift in thinking.
- Confirmation Bias — The tendency to search for, interpret, and remember information that confirms one's pre-existing beliefs.
- Disinformation — False information which is intended to mislead, especially propaganda issued by a government organization to a rival power.
- Misinformation — False or inaccurate information that is spread regardless of intent to deceive (e.g., a "Rumor").
- Astroturfing — Creating the appearance of a "grassroots" movement when it is actually being funded and organized by a central power.
- Gaslighting — A form of manipulation where someone is made to doubt their own perceptions and memories.
- The Big Lie — A propaganda technique where a lie so "colossal" is told that no one would believe that someone "could have the impudence to distort the truth so infamously."
- Echo Chamber — A situation where people only hear opinions that match their own.
- Appeal to Authority — Using a famous person or a "scientist" to vouch for an idea, regardless of their actual expertise.
- Bandwagon Effect — Persuading someone to do something because "everyone else is doing it."
- Fear Mongering — The action of deliberately arousing public fear or alarm about a particular issue.
- Loaded Language — Words with strong emotional connotations used to influence an audience (e.g., "Freedom Fighter" vs. "Terrorist").
Understanding
Propaganda and persuasion are understood through Emotion over Reason.
1. The "Path of Least Resistance": The human brain is lazy (it wants to save energy). Propagandists use this by:
- Repetition: If you hear a lie 1,000 times, your brain starts to treat it as a "Fact" because it is "Easy" to recall. (The Illusory Truth Effect).
- Simplicity: Complex problems (like the economy) are reduced to a simple "Enemy" or "Slogan."
2. The Six Principles of Persuasion (Robert Cialdini):
- Reciprocity: If I give you something for free, you feel you "owe" me.
- Scarcity: "Limited time offer!" makes you want it more.
- Authority: We follow people in "Uniforms" or with "Titles."
- Commitment: If I get you to agree to a small thing, you are more likely to agree to a big thing.
- Liking: We are persuaded by people we find attractive or similar to us.
- Consensus: We look to others to see how we should act.
3. Black vs. White vs. Gray Propaganda:
- White: The source is known and the information is mostly true (e.g., Government health ads).
- Black: The source is hidden or "Faked" to be an enemy (e.g., a fake "Opposition" group).
- Gray: The source is unknown and the truth is a mix of facts and lies.
The Filter Bubble: In the digital age, propaganda is "Personalized." Instead of one poster for the whole city, the algorithm shows a "Fear" ad to one person and a "Pride" ad to another, based on their private data. This is called Micro-targeting.
Applying
Modeling 'The Echo Chamber' (The Bias Loop): <syntaxhighlight lang="python"> def update_belief(current_belief, information_received):
"""
Shows how Confirmation Bias protects a belief.
"""
# If the info matches my belief, I believe it MORE.
# If it doesn't, I ignore it or call it 'Fake News'.
if information_received['lean'] == current_belief['lean']:
current_belief['strength'] += 0.2
return "ACCEPTED: This confirms what I already knew."
else:
current_belief['strength'] -= 0.05 # Only a small impact
return "REJECTED: This is biased/unreliable information."
- User is 'Pro-Choice'. Gets a 'Pro-Choice' news bit.
my_mind = {'lean': 'Left', 'strength': 0.5} news = {'text': 'Study shows X is good', 'lean': 'Left'}
print(update_belief(my_mind, news))
- This logic is why 'Facts' rarely change people's minds
- once they are part of an echo chamber.
</syntaxhighlight>
- Propaganda Landmarks
- The Office of War Information (WWII) → The birth of modern "Mass Persuasion" through posters (We Can Do It!) and film.
- Edward Bernays (Propaganda) → The nephew of Freud who applied "Psychology" to "Public Relations," convincing women to smoke by calling cigarettes "Torches of Freedom."
- The Cambridge Analytica Scandal → Showing how Facebook data was used to "Profile" and "Manipulate" voters during major elections.
- Deepfakes in Politics → The current frontier where video and audio are faked to make it seem like a leader said something they didn't.
Analyzing
| Feature | Education | Propaganda |
|---|---|---|
| Goal | To teach 'How' to think | To teach 'What' to think |
| Method | Challenges ideas / Encourages questions | Repetition / Discourages doubt |
| View of Truth | Complex and nuanced | Simple and one-sided |
| Outcome | An independent mind | A loyal follower |
The Concept of "Information Fatigue": If you flood someone with 10,000 conflicting stories, they don't know who to trust. They become "Exhausted" and stop caring about the truth altogether. This is a common strategy of modern "Gray Propaganda"—not to make you believe a lie, but to make you stop believing in the truth.
Evaluating
Evaluating a message: (1) Source: Who is telling me this, and what do they want? (2) Emotional Trigger: Is this making me feel "Angry" or "Afraid" (a sign of manipulation)? (3) The 'Missing' Info: What is this story not telling me? (4) Fact-Checking: Can I find this same information on three different, unrelated websites?
Creating
Future Frontiers: (1) AI-Generated Influence: Bots that can talk to you for days, building "Trust" before trying to change your vote. (2) Neuromarketing: Using brain scans to see exactly which words or colors trigger a "Buy" or "Believe" response. (3) Deepfake Forensics: Developing tools that can "prove" a video is real or fake at the speed of light. (4) Global Media Literacy: A movement to teach every child in the world how to "Deconstruct" a message before they believe it.