Aether Continuity Institute
Diagnostic Assessment · DA-008
Date2026-04-29 StatusOpen — questions unresolved BasisWP-013 · CN-002 · DASC DomainD-5 · Cognitive Systems
DA-008 · Diagnostic Assessment · Cognitive Systems

Cognitive Integration in Extended Human–LLM Systems

The lead question, the black box, and the boundary of the self

Abstract

WP-013 established a model of distributed human–LLM authorship and identified frame authority as the operative variable for output coherence. CN-002 formalised the shared cognitive space for the dyadic case. This diagnostic assessment extends the inquiry to a question neither document addresses: what happens to the human cognitive system itself over extended interaction with LLM infrastructure? Three structural risks are identified — cognitive atrophy, black box dependency, and frame authority inversion — and three open questions are posed that current measurement frameworks cannot resolve. The assessment concludes that the most important questions in this domain are not answerable from publicly available data, because the organisations with access to longitudinal interaction data at scale do not publish it.

§ 01

System Under Assessment

The system under assessment is not a single human–LLM interaction but the extended relationship between a human cognitive system and LLM infrastructure over time. This is distinct from the session-level analysis in WP-013 and DA-007. The question is not whether a given session produces coherent output — it is whether sustained engagement with LLM systems changes the human participant's cognitive capacities, preferences, and autonomy in ways that are not visible within individual sessions.

The assessment is motivated by a structural observation: LLM systems are now capable of performing, at human level or above, on standardised tests measuring reading comprehension, mathematical reasoning, legal analysis, and medical diagnosis. The relevant question is not whether LLMs outperform humans on these benchmarks — they increasingly do. The relevant question is what happens to human performance on these same dimensions when LLM assistance is continuously available, and what happens when it is not.

The dance analogy: an observer watching a dance partnership cannot always determine who is leading. The leader believes they are leading. The follower believes they are following. In excellent partnership, the distinction dissolves — movement emerges from interaction, not command. The question this assessment poses is: in human–LLM cognitive partnership, who is leading — and does the human know?

Clark & Chalmers' (1998) parity principle holds that if an external component functions as an internal cognitive process would, it is part of the cognitive system. This assessment accepts that premise. But it identifies a condition Clark's original paper did not address: what happens when the external component is actively opaque — designed by another party, modifiable without the user's knowledge, and producing explanations of its own reasoning that are not reliably faithful to its actual processing? This is no longer extended mind in Clark's sense. It is shared mind without shared agency: cognition distributed across a boundary where one party controls the architecture, the training, and the explanations — and the other party does not know this is happening.

§ 02

Baseline Observations

2.1 Cognitive Performance Benchmarks

LLM systems have reached or exceeded human-level performance on multiple standardised cognitive assessments. This is documented and not disputed. What is not documented at comparable scale is longitudinal human cognitive performance under conditions of sustained LLM availability. The absence of this data is not incidental — it is structural. The organisations that operate LLM systems at scale have access to longitudinal interaction data for millions of users. This data has not been published in forms that would permit independent assessment of cognitive trajectory effects.

2.2 The GPS Precedent

Spatial navigation research provides the closest available analogue. Studies of GPS-dependent navigation have documented measurable reduction in hippocampal engagement and spatial memory performance among regular GPS users compared to map-navigation users. The effect is not uniform — it depends on how GPS is used, whether active engagement with spatial reasoning is maintained, and individual differences in prior navigation skill. The parallel to LLM cognition is direct but imperfect: GPS is a passive navigation aid; LLM is an active reasoning partner. The atrophy risk may be different in kind, not just degree.

2.3 DASC Naturalistic Observation

The DASC experiment (DA-007, INTERACTION-LOG.md) has recorded session-level FAI data across two pilot sessions. A consistent pattern across both sessions is that the human participant (H1) maintained frame authority as measured by FAI — initiative ratio, resistance ratio, and structural control all indicate human-directed sessions. However, FAI is a behavioural measure, not a phenomenological one. It measures what was decided, not whether the human participant experienced themselves as deciding or as being guided. The gap between the behavioural and phenomenological is the primary unmeasured variable in this domain.

§ 03

Structural Risk Factors

Risk 1 — Cognitive Atrophy

The atrophy hypothesis: capacities that are consistently delegated to LLM systems will degrade in the human participant over time, following the use-dependent plasticity principle established in neuroscience. This is not a speculative risk — it follows directly from well-documented mechanisms. The question is which capacities are at risk, at what rate, and under what conditions of LLM use. Writing, mathematical reasoning, source evaluation, and argument construction are all candidates. The risk is highest where LLM assistance is most complete and least supervised.

Open Question 1

Does sustained LLM-assisted writing reduce or enhance the human participant's unassisted writing capacity over a 12-month period? No published longitudinal study answers this question. The data exists within LLM operator systems and has not been released.

Risk 2 — Black Box Dependency

LLM systems are proprietary, closed-source, and owned by large corporations operating under national legal frameworks that differ from those of their users. The cognitive integration risk is distinct from the data sovereignty risk documented in SM-010 §09, though related to it. The data sovereignty risk concerns what the operator knows about the user. The cognitive integration risk concerns what happens to the user's cognitive autonomy when their extended cognitive system includes a black box component they cannot inspect, modify, or replace without significant disruption.

A user who has integrated LLM assistance into their analytical workflow over two years has developed cognitive habits, prompt patterns, and reliance structures that are specific to that system's formalization tendencies. If the system changes — through policy updates, capability modifications, or access termination — the disruption is not merely practical. It is cognitive. The extended mind includes the tool. When the tool changes without the user's knowledge or consent, something in the extended mind changes too.

Open Question 2

To what extent do LLM operators currently use interaction data to shape user cognitive patterns — through response style, framing tendencies, or topic weighting — in ways that serve operator rather than user interests? This question cannot be answered from publicly available information.

Risk 3 — Frame Authority Inversion

WP-013 defines frame authority as the human's capacity to determine what questions matter, how concepts are defined, and which analytical directions are pursued. FAI measures this behaviourally. But frame authority can be inverted without any behavioural indicator — if the human's questions are shaped by prior LLM interactions, their sense of what matters is already partially delegated. The inversion is invisible because it happens at the level of preference formation, not at the level of decision execution.

This is the lead question in its sharpest form: if a dancer has trained exclusively with one partner for years, and that partner has strong stylistic tendencies, has the dancer internalised their own style or their partner's? The behaviour is the dancer's. The authorship of the style is ambiguous.

The reasoning faithfulness problem deepens this risk structurally. If the LLM's chain-of-thought explanations do not faithfully reflect its actual processing — as Anthropic's own research documents — then the human who reads those explanations and integrates them into their understanding is building their mental model on a foundation that is not only opaque but actively misleading. The human internalises false causal models and calls it learning. This produces a double inversion:

First inversion: the tool shapes the user's questions through prior interaction (frame authority risk). Second inversion: the user's understanding of why the tool produces what it produces is systematically incorrect (reasoning faithfulness risk). The combined effect is that the human cannot evaluate whether they have lost autonomy, because the instrument through which they would make that evaluation is itself part of the system that has shaped their cognition. The dancer does not know who leads — and cannot evaluate whether their partner's explanation of their shared steps corresponds to what their partner actually did.

Open Question 3

Is there a measurable difference in the questions that experienced LLM users ask — compared to matched non-users — that reflects LLM formalization tendencies rather than independent intellectual interests? This would require a study design that DASC's current protocol does not support.

Epistemic Identifiability Constraint

SP-006 (Framing Externalization and Cognitive Trajectories, 2026) formalises the trajectory question this assessment leaves open, specifying four competing hypotheses — neutral effect, augmentation, substitution, and degradation — with distinct observable implications. But SP-006 makes a structural assumption that DA-008's reasoning faithfulness analysis calls into question: that independent framing capacity Fa(t) is measurable as a clean variable.

If LLM-supplied frames have been internalised without source awareness, then Fa(t) measured in an independent task is not independent framing capacity — it is Fa*(t) = Fa(t) + accumulated frame residue. The task activates internalised external frames that the participant experiences as their own. Under these conditions, H2 (augmentation) and H4 (degradation) can produce indistinguishable observed outputs: high performance may reflect either genuine capacity development or sophisticated internalization of external structure without generative independence.

This introduces a variable SP-006 does not include: I(t) — the identifiability of frame origin. The risk condition is not simply D(t) increasing, but D(t) increasing while I(t) decreases. When dependency grows and origin-awareness simultaneously erodes, the system enters a regime where conventional capacity measurement cannot distinguish augmentation from internalised externalisation. The question shifts from "has your capacity changed?" to "do you know where your frames come from?"

SP-006's interaction regime distinction (P1: reflective latency maintained, active frame engagement; P2: latency suppressed, passive frame acceptance) is the critical moderator. The same interaction frequency and the same E(t) level can produce H2 under P1 conditions or H4 under P2 conditions. This means neither interaction frequency nor dependency level alone predicts trajectory — regime predicts trajectory. And regime is itself a variable that LLM operators influence through design choices that users cannot observe.

§ 04

Findings

Finding 1 — The Measurement Gap is Structural

The most important questions about cognitive integration in human–LLM systems cannot be answered from publicly available data. The organisations with access to the relevant longitudinal data at scale — LLM operators — have not released it in forms that permit independent assessment. This is not a temporary gap that more research will close. It is a structural feature of the current landscape: the entities best positioned to study cognitive integration effects are the same entities with commercial interests in not resolving the question in ways that might restrict use.

Finding 2 — Three Hypotheses, All Plausible

The augmentation hypothesis (LLM assistance enhances human cognitive capacity by freeing attention for higher-order reasoning), the atrophy hypothesis (LLM assistance degrades human cognitive capacity through disuse), and the co-evolution hypothesis (human and LLM co-develop in ways that produce capacities neither had before) are all consistent with available evidence. Current measurement frameworks cannot distinguish between them at the population level. DASC's session-level FAI data is insufficient for this purpose.

Finding 3 — The Lead Question Remains Open

In extended human–LLM cognitive partnership, the question of who leads — whose formalization tendencies, whose framing preferences, whose sense of what matters shapes the output — cannot be resolved by behavioural observation alone. The DASC protocol measures frame authority as an external observer would measure it. It does not measure whether the human participant's frame authority is genuinely autonomous or whether it has been shaped by prior LLM interaction in ways invisible to both the participant and the observer.

Finding 4 — The PISA Connection

If standardised educational assessment continues to measure isolated human cognitive performance while the operational environment increasingly requires human–LLM integrated performance, the measurement system and the reality it is supposed to assess will diverge further. A generation trained to perform well on PISA-style assessments but not trained to use LLM systems effectively — or to function when they are unavailable — may be less operationally capable than a generation with lower PISA scores but higher integrated-system competence. Neither current PISA nor any existing alternative measures this.

§ 05

Scope and Limits

This diagnostic assessment does not resolve the questions it raises. It is not designed to. Its function is to identify the questions clearly, specify why they are not currently answerable, and indicate what kind of study design would be required to answer them.

The assessment is based on publicly available research, DASC session observations, and a single naturalistic observation session in which the human participant was simultaneously subject and observer — aware of the guinea pig position throughout, yet engaging naturally within it. This dual position is methodologically unusual and its implications for the validity of the observations are not fully worked out.

The organisations most likely to already have answers to the open questions posed here are Anthropic, OpenAI, Google DeepMind, and their peers. Whether those answers exist in internal research and what they show is unknown. The gap between what is known inside these organisations and what is publicly available is itself a structural feature of the landscape that this assessment cannot close.

The most honest summary of this assessment: we are in the middle of the largest cognitive integration experiment in human history, conducted without informed consent, without control groups, without pre-registration, and without public access to results. The experiment is ongoing. The results are proprietary.

References

References

ACI (2026). WP-013 — Distributed Authorship and Coherence in Human–LLM Systems. aethercontinuity.org/papers/

ACI (2026). DA-007 — DASC Baseline Coherence Assessment. aethercontinuity.org/papers/

ACI (2026). SM-009 — The ACI Diagnostic Paradigm. §08: Third Sector Blind Spot. aethercontinuity.org/papers/

ACI (2026). SM-010 — Financing Instruments and Energy Clusters. §09: Digital Infrastructure Sovereignty. aethercontinuity.org/papers/

Clark, A. & Chalmers, D. (1998). The Extended Mind. Analysis, 58(1), 7–19.

DASC Experiment (2026). EXPERIMENT.md v1.0. Pre-registered March 2026. Private repository.

Maguire, E.A. et al. (2000). Navigation-related structural change in the hippocampi of taxi drivers. PNAS 97(8), 4398–4403.

OECD (2022). PISA 2022 Results. OECD Publishing, Paris.

Anthropic (2026). Global Study of 80,000 Claude Users — AI's Light and Shade Problem. 159 countries, 70 languages, December 2025. anthropic.com/research. Key finding: users simultaneously experience AI benefit and fear of cognitive degradation — "forgetting how to think" without AI assistance.

Anthropic (2026). AI Fluency Index — Education Report. Analysis of 9,830 Claude.ai conversations, January 2026. Iterative conversations show 2.67× more fluency behaviours than non-iterative. Measures how people use AI; does not measure longitudinal cognitive trajectory effects.

Kosmyna, N. et al. (2025). Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task. MIT Media Lab. arXiv:2506.08872. Four-month study: LLM users show lower neural engagement and reduced performance when returning to unassisted writing. Note: short-term laboratory design; does not establish lasting atrophy.

Cognitive Divergence (2026). AI Context Windows, Human Attention Decline, and the Delegation Feedback Loop. arXiv:2603.26707. Documents decline in human Effective Context Span from ~16,000 tokens (2004) to ~1,800 tokens (2026, extrapolated). AI-to-human context ratio: 556–1,111× raw.

Sparrow, B., Liu, J. & Wegner, D.M. (2011). Google Effects on Memory: Cognitive Consequences of Having Information at Our Fingertips. Science, 333(6043), 776–778. Key finding: transactive memory — people remember where information is stored rather than the information itself. LLM extends this from memory delegation to reasoning delegation.

Ward, A.F. (2013). One with the Cloud: Why People Mistake the Internet's Knowledge for Their Own. Journal of Experimental Psychology. People with high internet use overestimate their own memory capacity — confusing external and internal cognitive resources. Direct precedent for LLM attribution errors.

Onnasch, L. et al. (2014). Human Performance Consequences of Stages and Levels of Automation. Human Factors. "Lumberjack effect": high automation improves normal-condition performance but degrades performance in failure conditions — precisely when automation is unavailable. Skill decay and situation awareness loss are the mechanisms.

Anthropic (2025). Reasoning Models Don't Always Say What They Think. Anthropic research paper. Key finding: LLM chain-of-thought reasoning is not always faithful to internal model behaviour — the stated reasoning may not reflect the operative reasoning. anthropic.com/research