DATASCI 185: Introduction to AI Applications

Lecture 23: Misinformation, Deepfakes and Trust Online

Danilo Freire

Department of Data and Decision Sciences
Emory University

Welcome back! 🎭

Recap of last class

  • We discussed AI in healthcare and education
  • Promise vs reality: lab results ≠ deployment
  • Human-AI collaboration works better than either alone
  • Equity concerns cut both ways
  • Today: What happens when AI is used to deceive?

Pope Francis in Balenciaga! 😂

Lecture overview

What we will cover today

Part 1: The landscape

  • What is misinformation?
  • AI and synthetic media
  • Why it matters now

Part 2: Deepfakes

  • Technical foundations
  • Real-world examples

Part 3: Why we fall for it

  • Evolutionary wiring and fake news
  • System 1 vs System 2
  • Confirmation bias and motivated reasoning
  • Bandwagon effect and the “in the know” feeling

Part 4: Responses

  • Technical detection
  • Platform policies
  • Legal frameworks
  • Media literacy

Tweet of the day

Well done, swifties! 👍

Source: X.com

The misinformation landscape 📰

Definitions matter

Key distinctions:

Term Definition
Misinformation False information spread without intent to deceive
Disinformation False information spread deliberately to deceive
Malinformation True information shared to cause harm
Synthetic media Content created or manipulated by AI

Why the distinctions matter:

  • Intent shapes appropriate responses
  • Different actors, different motivations
  • Legal treatment and solutions vary by category

The old problem:

  • Misinformation is ancient
  • Technology changes scale and speed
  • But the human vulnerabilities remain

Source: CUNY Library

Not all false information is equal. Understanding the type and intent helps us respond appropriately.

Why AI changes things

Before generative AI:

  • Fake images required Photoshop skills
  • Fake videos needed studios, fake audio needed voice actors
  • Production was expensive and slow
  • Detection was often possible

After generative AI:

  • Anyone can create convincing fakes
  • Production cost approaches zero
  • Quality improves rapidly
  • Detection becomes harder as generators improve

The scale problem:

  • Generate thousands of fake articles instantly
  • Create targeted content for specific audiences
  • Human moderators can’t keep up

Source: Encyclopedia Britannica

Fake news are not new. The “Protocols of the Elders of Zion” was a fabricated antisemitic text from 1903 that fueled hatred for over a century. Now, anyone can create similar content with AI in minutes.

The attention economy

How platforms amplify misinformation:

  • Engagement = revenue
  • Outrage drives engagement
  • False information is often more engaging
  • Algorithms optimise for clicks, not truth

The correction problem:

  • Lies spread faster than corrections
  • False stories reach millions
  • Corrections reach thousands
  • Even debunked claims persist in memory

Vosoughi et al. (2018):

  • Analysed 126,000 stories on Twitter
  • False stories 6x more likely to be retweeted
  • Reached more people, spread faster

Source: NYU School of Engineering

Platform incentives often work against truth. This isn’t a bug; it’s a feature of ad-supported media.

Deepfakes 🎬

What are deepfakes?

Technical definition:

How they work:

  1. Collect training data (images/video of target)
  2. Train neural network to generate target’s face
  3. Apply to source video or image
  4. Refine to reduce artefacts

Main techniques:

  • Diffusion models: Newest, highest quality
  • Voice synthesis: Similar principles for audio

The same techniques that create art and entertainment also create convincing deceptions. To know more, read the comprehensive survey by Tolosana et al. (2020).

Real-world examples

Political deepfakes:

Example Impact
Obama PSA (2018) BuzzFeed/Jordan Peele demonstration
Trump pandemic briefing Fake emergency broadcast
Zelensky surrender video Attempted wartime deception
Indian election audio AI voice clones of politicians

Financial fraud:

Non-consensual intimate imagery:

  • Vast majority of deepfakes (96%+)
  • Women overwhelmingly targeted
  • Tool for harassment and extortion
  • Legal response lagging

Source: NPR

The harms are real.

RAND’s primer on deepfakes

Key insights from RAND research:

  • Deepfakes are one tool among many
  • “Cheap fakes” often just as effective
  • Context matters more than quality
  • Social vulnerability enables technical attacks

The liar’s dividend:

  • When anything could be fake…
  • Real evidence can be dismissed as fake
  • “That video of me is a deepfake!”
  • Truth becomes contested terrain

RAND’s threat assessment:

  • Near-term: Targeted harassment, fraud
  • Medium-term: Election manipulation
  • Long-term: Erosion of shared reality

Source: RAND Corporation

“The challenge isn’t just detecting deepfakes. It’s maintaining the social trust that allows evidence to have meaning.”

The authenticity crisis

Chesney & Citron’s analysis:

  • Published in Foreign Affairs (2019)
  • Frame deepfakes as epistemic threat

Their framework:

  1. Undermining journalism: How can news verify sources?
  2. Weaponising uncertainty: Plausible deniability for everyone
  3. Amplifying polarisation: Believe what you want

The marketplace of ideas:

  • Depends on shared facts
  • Deepfakes pollute that marketplace
  • Verification becomes impossible at scale
  • Democracy requires shared reality

Discussion: what would convince you? 🤔

Scenario:

A video emerges of a political candidate you support saying something terrible. The candidate claims it’s a deepfake.

Discuss (2 minutes):

  1. How would you decide if it’s real?
  2. What evidence would convince you either way?
  3. Does your political alignment affect your judgment?
  4. What if experts disagree?

The difficult reality:

  • Motivated reasoning affects everyone
  • We’re more sceptical of evidence against our beliefs
  • Expert consensus helps, but experts can be wrong
  • Verification takes time; damage is immediate
  • There may be no satisfying answer

This discomfort is the point. When any video could be fake, our ordinary ways of knowing break down.

Why we fall for it 🧠

Are we hard-wired for fake news?

The evolutionary angle:

  • Our brains evolved for small groups and immediate threats
  • Novelty once meant survival-relevant information
  • We still react to new, surprising claims the same way
  • False news is almost by definition more novel than truth

Why outrage spreads:

  • Threat detection is fast and automatic
  • Emotionally charged content bypasses careful evaluation
  • Fear and disgust trigger sharing behaviour
  • We evolved to warn the group, not fact-check first

System 1 vs System 2

Kahneman’s (2011) dual-process theory:

System 1 System 2
Speed Fast, automatic Slow, deliberate
Effort Effortless Requires concentration
Mode Intuitive, emotional Analytical, logical
Default? Yes Only when triggered

Social media is a System 1 environment:

  • Infinite scroll rewards quick reactions
  • You see a video, react, share, move on
  • System 2 never kicks in unless something jolts you
  • Deepfakes are designed to feel real at System 1 speed

The problem:

  • Catching a fake requires System 2
  • But the platforms are built for System 1
  • Strong emotion (outrage, fear) suppresses System 2

Confirmation bias

What it is:

  • We seek evidence that confirms what we already believe
  • We dismiss evidence that contradicts it
  • This isn’t laziness; it’s how brains manage information overload
  • Nickerson (1998) called it the most pervasive bias in human reasoning

How it works with deepfakes:

You see a deepfake of… Your reaction
A politician you distrust “I knew it!”
A politician you support “Obviously fake”
  • Same technology, same quality. Different reaction
  • Taber & Lodge (2006): correcting misinformation can strengthen the original belief
  • Corrections feel like an attack on identity, not just on facts

Confirmation bias is not about intelligence. Educated people are sometimes better at rationalising bad evidence, because they have more tools to construct justifications.

The bandwagon effect

What Asch showed in the 1950s:

  • People will deny what their own eyes tell them if enough others disagree
  • About 75% of participants conformed at least once
  • Even when the correct answer was obvious

On social media, numbers are social proof:

  • A video with 2 million views “must be real”
  • 50,000 retweets = credibility
  • We outsource our judgment to the crowd
  • The crowd is just as fooled as we are

How this helps deepfakes spread:

  • Early shares create a snowball
  • Each share adds perceived legitimacy
  • By the time fact-checkers respond, millions have seen it
  • Taking it down can actually increase belief (“they’re hiding something”)

Asch’s experiments used lines on a card. Today, the same psychology plays out with deepfake videos shared millions of times.

The “in the know” feeling

Sharing as social currency:

  • Breaking news makes you feel like an insider
  • “Did you see this?!” = status in your group
  • Being first to share gets the attention

Why this matters:

  • Creates a race to share before verifying
  • People share content they haven’t fully watched
  • Emotional headlines are enough to trigger a repost

Combined with the other biases:

  • Confirmation bias picks what we share (things we agree with)
  • The bandwagon effect picks when (once it’s already popular)
  • The “in the know” feeling picks how fast (immediately, no checking)

Source: Adam Grant

We share what confirms our views and makes us look informed. The truth can wait; the dopamine hit can’t.

Responses 🛡️

Technical detection

Detection approaches:

Method How it works
Artefact detection Unnatural blinking, lighting
Biological signals Pulse, micro-expressions
Source forensics Compression artefacts, metadata
AI vs AI Train detectors on known fakes

Current state:

  • Detection works on yesterday’s fakes
  • Generators improve faster than detectors
  • Arms race with no end in sight

The fundamental problem:

  • Detectors can be fooled
  • Adversarial training improves generators
  • Real-time detection is difficult
  • Scale: Millions of images, few moderators

Content provenance

The idea:

  • Instead of detecting fakes…
  • Verify authenticity of real content
  • Track provenance from creation
  • “Where did this come from?”

C2PA (Coalition for Content Provenance and Authenticity):

  • Industry consortium (Adobe, Microsoft, etc.)
  • Digital signatures embedded in media
  • Chain of custody tracking
  • “Content Credentials”

How it works:

  1. Camera cryptographically signs image at capture
  2. Each edit adds to provenance record
  3. Viewers can verify chain back to source
  4. Unsigned content becomes suspicious

Limitations:

  • Requires universal adoption
  • Legacy content has no provenance
  • Can be stripped or faked
  • Doesn’t prevent creation, only tracks it
  • Privacy concerns about tracking

Platform policies

What platforms do:

Platform Policy
YouTube Labels AI-generated content
Meta Restricts manipulated media
TikTok Requires disclosure
X/Twitter Varies; less moderation

Challenges:

  • Volume: Billions of uploads
  • Speed: Viral before review
  • Jurisdiction: Global platforms, local laws
  • Incentives: Engagement vs truth

What happens when platforms remove content:

  • Moves to less moderated platforms
  • Removal can increase belief (“they’re hiding something”)

Source: Vice - YouTube

Media literacy

The case for education:

  • Technical solutions have limits
  • Human judgment is the last line
  • Teach critical evaluation skills

What media literacy includes:

  • Source evaluation: Who made this? Why?
  • Lateral reading: What do others say?
  • Slow down: Don’t share immediately
  • Check for red flags: Inconsistencies, odd details

Evidence on effectiveness:

  • Mixed results from interventions

Practical tips:

  1. Reverse image search: Google Images, TinEye, etc
  2. Check original sources: Who published it first?
  3. Look for red flags: Unusual lighting, blurring
  4. Consider context: Who benefits?
  5. Wait for verification: Breaking news is often wrong

The comprehensive approach

No single solution works:

  • Detection: Arms race, always behind
  • Provenance: Adoption takes years
  • Platform policies: Incentives misaligned
  • Legal frameworks: Slow, jurisdictional
  • Media literacy: Helps but doesn’t scale

The realistic goal:

  • Not elimination of misinformation
  • Raise costs of deception
  • Slow the spread
  • Maintain enough trust for democracy

Summary 📝

Main takeaways

The landscape

  • Misinformation is old; AI changes scale
  • Attention economy amplifies false content
  • Platform incentives work against truth

Deepfakes

  • Synthetic media quality improving rapidly
  • Political manipulation, NCII, fraud
  • The “liar’s dividend” undermines all evidence

Why we fall for it

  • Evolutionary wiring makes novelty irresistible
  • System 1 scrolling, System 2 sleeping
  • Confirmation bias is the vulnerability no patch can fix
  • Bandwagon and “in the know” feeling accelerate sharing

Responses

  • Detection: necessary, insufficient
  • Provenance: promising, years away
  • Legal: slow, jurisdictional
  • Literacy: helps, doesn’t scale
  • Comprehensive approach required

… and that’s all for today! 🎉