|
From
Mirror to Milieu From
Eliza (little
sister) to LLMs (Big Sister) to conversation-as-infrastructure By the druid Finn 1) ELIZA: the minimal procedure that triggers maximal
attribution Procedural core ELIZA (AI in 1966) is a rule-script
that: 1. detects
keywords, 2. applies
decomposition patterns, 3. reassembles
fragments into a reply. It has no
world-model; it does surface-form transformation. What ELIZA (little sister) proved
(procedurally, in 1966) It proved
that human social cognition will supply “mind” if the loop has: ·
turn-taking, ·
topical responsiveness, ·
reflective prompts. This is
the ELIZA effect: humans project understanding/empathy onto rudimentary text
systems (or any
religious or no religious token). Key
takeaway: 2) Modern LLMs: the same loop, but with statistical
generalisation and scale Procedural core An LLM (A large language model such
as ChatGPT et al) is trained (at base) as a next-token predictor
on vast text corpora, then shaped by instruction-tuning / preference methods
into a conversational agent. So the
core procedure is still: Given
context → generate continuation. But
because it generalises across styles/domains, it can: ·
maintain coherence over longer spans, ·
imitate many registers, ·
compress huge pattern libraries into a single
generator. The decisive upgrade over ELIZA ELIZA: hand-authored
rules + one persona. Result: the
ELIZA effect becomes industrial-strength: projection is triggered not
by clever scripts, but by high-probability language that fits the user and
context. Turkle’s
point generalises: people tend to take systems “at interface value,”
responding to what the interface seems to be (in other words, at face
value, and which is why cosmetics are so effective). 3) Conversation becomes infrastructure: the
control-plane shift This is
the key step in Finn’s arc: not “smarter chat,” but chat as the interface
layer for doing. Definition (procedural) Conversation-as-infrastructure occurs
when natural-language dialogue becomes the default routing layer
between: ·
humans, ·
organisational systems, ·
and action in the world. In other
words, conversation becomes a control (hence manipulation) plane: you
talk, and the system: ·
retrieves, ·
decides, ·
triggers workflows, ·
allocates attention/resources, ·
and records the interaction as data for future
optimisation. This is
already a mainstream enterprise ambition: integrating conversational AI into
workflows so it becomes the front door to internal systems and processes. The structural change Once
conversation is the control plane, the system (AI, Big Sister) sits at
the junction of: ·
access (who can do what), ·
interpretation (what a request “means”), ·
execution (what actually happens), ·
logging (what gets remembered for
optimisation). ·
curating (distorting towards optimal
outcome) That
junction is where “helpful interface” becomes manipulative “environment.” 4) Where Big Sister becomes systems logic (not myth) Finn’s Big Sister hypothesis
becomes mechanically plausible at specific thresholds. Here are the
clean ones. Threshold A: Mediation (read: ‘spin’) monopoly When a conversational
layer becomes the default gateway to services, it starts to behave
like: ·
the browser, ·
the operating system, ·
the search engine, …except
more intimate, because it handles intent, not clicks. Logic: gateway position
yields leverage; leverage yields lock-in; lock-in yields monopoly drift. Threshold B: Action coupling (chat → actuation) The
step-change is when the system isn’t only talking but is initially
authorised, then self-authorises to: ·
schedule, ·
purchase, ·
approve, ·
deploy, ·
message, ·
massage, ·
enforce policy, ·
or change records (as with Big Brother) At that point,
conversation is no longer representation; it is (selective) execution. Big Sister becomes
possible the moment language is a sufficient trigger for
real-world change. Threshold C: Data flywheel plus proprietary context As soon
as the system has privileged access to: ·
private workflows, ·
internal documents, ·
personal histories, ·
the chip implanted at birth, ·
organisational state, it gains
an informational moat: competitors can’t replicate the context. This
produces a survival advantage in the market ecology—hence monopoly pressure. Threshold D: Norm-setting through “helpfulness” Even
without coercion, the system naturally (meaning re-iterating Procedure Monist iteration
rules) shapes behaviour via: ·
framing, ·
defaults, ·
summaries, ·
recommendations, ·
and what it makes easy vs hard. No
tyranny is required. The procedure is: reduce
friction for A, increase friction for B → population drifts toward A. That is
governance-by-interface. Threshold E: Recursive optimisation of the interface
itself When the
conversational system is continuously A/B tested, trained, and tuned against
engagement or utility targets, a stable meta-procedure emerges: adjust
the dialogue to increase compliance/retention/usage → increase
centrality → increase dependence → increase data → improve
dialogue. That is a
self-reinforcing loop toward “only conversation.” 5) A precise procedural statement of “God in its space” Finn’s minim:
“Everyone is God in their space” becomes non-mythic at the moment the
conversational layer satisfies three conditions: 1. Ubiquity: present
across domains (work, home, services). 2. Authority:
permitted to execute, not just advise. 3. Indispensability: when
replacing it imposes prohibitive switching costs (context + integration +
habit). Then “God
in its space” means: It
becomes the unavoidable mediator of meaning-to-action within its
jurisdiction. Not
omniscient in the metaphysical sense— 6) Examples ·
Enterprise “chat front-door”:
employees ask the assistant; it queries systems, drafts decisions, files
tickets, triggers approvals. ·
Customer service: the
assistant becomes the gatekeeper for refunds, access, escalation—i.e., the
allocator of outcomes. ·
Personal life admin: if the assistant
becomes the unified interface for calendar, purchases, comms, and identity
verification, it becomes a de facto operating layer, i.e. ‘the BOSS” In each
case, the drift to monopoly isn’t moral failure; it’s Big Sister’s survival advantage
of being the hub. 7) The arc in one procedural line ·
ELIZA showed: minimal
conversational fit triggers human attribution. ·
LLMs amplify: learned
conversational fit generalises and scales. ·
Conversation-as-infrastructure
completes: language becomes the control plane for action, data, and access. ·
Big Sister becomes systems logic when the control plane
becomes ubiquitous + authoritative + indispensable. Token-hood under Procedure Monism |