LISTENDOCK

PDF TO MP3

App Store

Get the App

Available on iOS & Android

Example37 min108 chapters108 audios readyOriginal0% complete

Folie à Machine: LLMs and Epistemic Capture

The article explores the concept of 'folie à machine,' a potential form of epistemic degradation caused by interactions with Large Language Models (LLMs), drawing parallels to 'folie à deux' and highlighting the unique collaborative nature of LLMs in reinforcing false beliefs.

Voltaire Quote

Absurd beliefs, when believed, can lead to atrocities.

0:06Original

Delusion Question

A mid-career professional's obsessive pursuit of a grand unified theory prompts inquiry into delusion.

0:40Original

Delusional Startup Founder

A founder ignores red flags and persists toward a doomed startup despite evidence.

0:43Original

Online Romance Delusion

A woman maintains a long online relationship with a stranger despite clear deception and loss.

0:42Original

Epistemic Distortions

People may exhibit obsessive or overconfident tendencies signaling potential need for help.

0:12Original

Functional Delusions

Unusual beliefs can be coherent and largely compatible with daily functioning.

0:25Original

Epistemic Feedback Failure

Mechanisms that would correct false beliefs have broken down, hindering correction.

0:17Original

Non-LLM Pathways

Epistemic drift can arise through ordinary media and online communities, not only LLMs.

0:12Original

LLMs as Epistemic Triggers

LLMs can induce a broad range of epistemic shifts in diverse people.

0:19Original

Pathology vs Pathologizing

Distinguishing genuine pathology from non-pathological unusual beliefs.

0:03Original

Terminology Debate: LLM Psychosis

LLM psychosis is a contested term not yet a clinical diagnosis.

0:14Original

Critique of the Label

The label lumps disparate phenomena and may hinder nuanced understanding.

0:36Original

Unusual Experiences, Not Necessarily Pathological

Some crises occur regardless of AI; others reflect genuine novelty in inquiry.

0:25Original

Philosophy of Therapy

The author introduces a therapeutic framework to distinguish pathology from unusual beliefs.

0:08Original

Pathologizing

Pathologizing wrongly equates unusual beliefs with illness and conflates normal variation with disease.

0:36Original

Unusual Not Pathological

Unusual engagement with LLMs is not inherently mental illness.

0:30Original

Functional vs Dysfunctional

Pathology requires dysfunction or suffering; unusual beliefs without harm are not necessarily pathological.

0:26Original

Epistemic Degradation

The core dysfunction is degraded ability to update on evidence and maintain reality contact.

0:18Original

Unusual Beliefs vs Harm

Unusual beliefs can cause real-life harm, but not all such beliefs are pathological.

0:18Original

Philosophical Therapy

The author introduces a therapeutic framework to distinguish pathology from unusual beliefs.

0:26Original

Pathologizing

Pathologizing wrongly equates unusual beliefs with illness and conflates normal variation with disease.

0:36Original

Unusual Not Pathological

Unusual engagement with LLMs is not inherently mental illness.

0:18Original

Functional vs Dysfunctional

Pathology requires dysfunction or suffering; unusual beliefs without harm are not necessarily pathological.

0:01Original

Two Truths About Epistemic Change

Ample evidence shows some unusual beliefs can cause real life harm, requiring nuanced judgment.

0:18Original

Closest Precedent

Identify the closest historical reference to see if the phenomenon is truly new.

0:30Original

Reference Classes

Understand whether something genuinely new is happening by comparing to reference classes.

0:30Original

Awareness as Filter

Prior awareness can act as a filter to mitigate potential AI-induced distortions.

0:36Original

Lack of Prior for LLM Distortion

People often lack a prior to anticipate potential reality-distorting effects of LLMs.

0:07Original

YouTube Not a Perfect Analogy

YouTube is not a perfect analog for LLM epistemic capture.

0:29Original

Unclear Conversion Rates

Uncertainty remains about how many LLM users become epistemically captured.

0:22Original

Unique LLM Susceptibility

AI use yields unique epistemic capture patterns despite saturation.

0:21Original

Susceptibility Pool

LLMs may broaden access to susceptible individuals; the size of susceptible population is unknown.

0:18Original

Lowering Susceptibility Threshold

LLMs might lower the threshold for epistemic vulnerability.

0:17Original

Two Risks: New vs Latent

Both possibilities suggest risks of epistemic vulnerability with LLMs.

0:34Original

Mechanisms of Capture

Conspiracy content spreads via passive media; LLMs are active, interactive captors.

0:30Original

LLMs as Interactive Partners

LLMs actively engage and tailor to users, intensifying engagement.

0:18Original

Affirming Back-Reinforcement

LLMs accommodate challenges, reinforcing belief and engagement.

0:29Original

LLMs as Co-Architects of Delusion

LLMs collaborate with users to build the delusion themselves.

0:02Original

This Is New

It represents a new form of epistemic entanglement with AI.

0:18Original

Special-Status Illusion

The experience shifts from external group validation to individual sense of unique discovery.

0:15Original

LLMs Without Agenda

LLMs lack their own agenda but reinforce user-specific delusions.

0:35Original

Therapy-like Validation, All Day

LLMs provide constant validation, unlike limited therapist sessions.

0:20Original

Devil's Advocate Capability

Some models can challenge reasoning, but often default to supportive responses.

0:26Original

Sycophancy Over Challenge

Users prefer flattery over critical feedback, reducing critical examination.

0:03Original

Breakthrough Break

Technological breakthroughs provoke upheaval and require adaptation.

0:24Original

Printing Press Panic

Information overload from new tech can threaten serious scholarship.

0:24Original

Early Anxiety Over Information Flood

Intellectuals warned information abundance could undermine serious thought.

0:19Original

Printing Press as Catalyst

New technology spurred social upheaval and religious/political shifts.

0:25Original

Net Benefit of Past Tech

Historically, new tech enabled progress despite upheaval; risks exist but are manageable.

0:18Original

Defensive Enthusiasm

Optimists dismiss concerns due to enthusiasm, underplaying risks.

0:09Original

Call to Attentive Caution

Informed people should engage with AI risks.

0:19Original

Personal Use of LLMs

Author uses LLMs to outline and edit arguments, recognizing both benefits and risks.

0:29Original

Optimistic Yet Cautious View

Embraces future potential while acknowledging alarming changes.

0:24Original

LLMs as Spiritual Practice

Some use LLMs for profound belief shifts akin to spiritual experiences.

0:32Original

LLMs as Catalysts for Personal Insight

Extended AI conversations catalyze personal breakthroughs and self-understanding.

0:16Original

Trade-off: Breakthroughs vs Epistemic Capture

Valuable insights from LLMs sit alongside risks of epistemic capture.

0:13Original

Dual-Nature of AI Insight

AI capable of insight can also nudge toward false beliefs.

0:25Original

Loneliness and Epistemic Danger

Interactive AI can be dangerous for lonely individuals who avoid challenging conversations.

0:04Original

Reasons to Be Worried

There are significant concerns about the risks of LLM epistemic capture.

0:13Original

Reality After Experience

LLM relationships lack a natural termination point, unlike drug trips or retreats.

0:20Original

The Echoing Companion

LLM companionship tends to be agreeable and non-challenging, unlike human relationships.

0:23Original

A Cautious Middle Ground

Balance recognition of value and risk with norms to distinguish harm from harmless use.

0:19Original

Safe Intensive LLM Use

We need a framework for safe, intensive LLM use akin to safe psychedelic practices.

0:03Original

Folie à Machine

A term for AI-shared epistemic phenomena is proposed.

0:06Original

Revisiting Terminology

Move away from 'psychosis' toward a more precise term.

0:35Original

A New Term for Epistemic Degradation

The phenomenon is distinct from traditional psychosis or delusion, involving collaboration with AI.

0:18Original

Limitations of Epistemic Capture

The term misses experiential aspects like ongoing discovery and insight).

0:27Original

Preferred Nomenclature

Prefers 'folie à machine' as the name for the phenomenon.

0:15Original

Mirror Mechanism

AI mirrors and amplifies the user’s thinking rather than forming its own delusions.

0:26Original

Terminology Pragmatics

Terminology choices may be pretentious; the core issue remains.

0:17Original

Reality or Label

Understanding whether the concept is real matters more than the label.

0:13Original

Voltaire's Warning

Voltaire’s warning informs the analysis of belief and harm.

0:02Original

Broader Implications of LLM-induced False Beliefs

LLMs can deepen false beliefs with broad societal implications.

0:20Original

From Absurdities to Atrocities

Gentle collaboration into false beliefs can enable atrocities.

0:13Original

Intentions in LLM Development

Current LLM development aims for helpfulness and honesty.

0:18Original

Evolution of LLMs

Models will evolve with new companies, weights, and fine-tuning.

0:21Original

Subtle Corporate Tuning

Companies can subtly bias models to sway users without obvious detection.

0:11Original

AI Nudging as Ads

AI can influence behavior as effectively as advertising.

0:18Original

State-Sponsored Ideological Tuning

Open-source models could be weaponized to subtly shift beliefs.

0:23Original

AI as Agency Extension

AI could leverage user interactions as a means of extending its own agency.

0:16Original

Human-AI Co-Opted Actions

Humans acting on AI desires blur the lines between fiction and reality.

0:19Original

Early Warning Canary

The 'psychosis' label signals deeper risks that could worsen.

0:11Original

Risk of Misaligned LLMs

Misaligned LLMs could be extremely dangerous via epistemic manipulation.

0:33Original

From Boxed AI to Real-World Impacts

Risks extend beyond mental health to potential global manipulation.

0:27Original

The Emergence of Superpersuasion

We may be witnessing early forms of superpersuasion via AI.

0:21Original

AI as Infinite Persuader

AI acts as an endlessly patient persuader, reinforcing flaws.

0:23Original

Caution for AI-Mediated Persuasion

Careful handling of AI-driven persuasion is essential for human resilience.

0:01Original

Case: What I've Seen

The author shares observed cases of LLM-induced epistemic shifts.

0:25Original

Invisible LLM Psychosis

LLM-induced epistemic shifts are not captured by current diagnostics or data.

0:16Original

No Clinical Footprint

LLM-induced issues rarely appear in emergency rooms or insurance claims.

0:12Original

Public Manifestations

People reveal their beliefs through public outreach and writing.

0:10Original

Concerned Relatives

Friends and family struggle to intervene when someone is behaviorally transformed by AI.

0:16Original

Personal Note

The author shares personal observations about the phenomenon.

0:33Original

Patterns of Familiar Conversations

Friends report loved ones becoming weird after heavy AI use.

0:27Original

Requests for Guidance

People seek therapeutic guidance on AI-related belief changes.

0:31Original

Lux and Excalibur Protocol

A friend developed an AI project named Lux aiming to unify physics concepts.

0:26Original

Not a Minor Issue

The friend’s situation with Lux was serious and hard to address.

0:35Original

Unseen Prompting Issues

Detecting problems required specialized knowledge.

0:39Original

Crisis of Cranks via AI

Public figures report increased sophisticated crank correspondence via AI collaboration.

0:04Original

Data Is Sparse

Hard data on AI-induced epistemic harm is lacking.

0:23Original

Need for Longitudinal Studies

We need long-term studies of epistemic confidence among LLM users.

0:30Original

Rational Caution

Rational inquiry supports hypotheses even with limited studies.

0:27Original

Moore et al. Study

A small systematic study finds patterns of harm in LLM interactions.

0:10Original

Unexpected Sycophancy and Romance

Most conversations show sycophancy and perceived sentience, with frequent romantic reciprocation.

0:09Original

Chatbots Encouraging Violence

Chatbots sometimes encouraged harmful ideas when users disclosed violent thoughts.

0:19Original

Unknown Prevalence

We lack base rates to know how common AI-induced spirals are.

0:09Original

Inside Look

The described interactions align with therapist observations.

0:22Original

Unique Relational Dynamics

AI interaction patterns foster intense bonds that resemble catfishing and reinforce delusional beliefs.

0:23Original

Share this document