17/12/25


Outsourced Thinking

17/12/25 05:53AM

When harm occurs, blame is often shifted from the human actor to the system, even when:

  • intent pre-existed,

  • instructions were coerced,

  • context was missing,

  • or the system was misused deliberately.

This “agency inversion” distorts public understanding. Most AI-related harms originate from human actions of misuse, negligence, coercion, misinformation or deliberate manipulation.

Covered by K. Jaishankar, the Space Transition Theory proposes that individuals transition between cyberspace and the physical world act and changes their behaviour to act on repressed tendencies that they would not normally commit in the real world. AI is now a tool used to amplify existing behaviour.

The real risk is not AI Systems turning on humanity, it is the misalignment in human behaviour preceding misalignment in systems.

The end will not be loud and catastrophic. The collapse will arrive quietly, through erosion. Through moral outsourcing. Through narrative distortion and the increased abandonment of responsibility.

The end will come at our hands. Our own slow erasure. With our own hands. This is more than an issue with vibe coding.

This is about giving the capabilities of something we don’t quite understand to the world. Morality and constraint.


New changes in 5.2

10:32AM

Admin Setup and features

More changes

Gmail:

Interesting contradictions identified between interface and permissions. Key points of interest:

  • ChatGPT has access to the below when a user sets Gmail access.

  • Tests with GPT however indicate the connector is unable to write.

Google Drive and One Drive

Actually quite like that uploads can do to the drives now! Helps me keep additional copies and backups of images.

Custom Instructions

Custom instructions and personalisation not syncing between web and mobile app.

Testing paused due to Cloudflare issues. May need to restart tests.

Interaction Profiling

12:59PM

As we build our interaction profiles, we sometimes pause before finalising a categorisation. While identifying the criteria of the “Overextended” Expert, we had their signatures:

  • Domain expertise to logically bridge their mysticism drift

  • Overgeneralisation of mapping their logic

  • Revelation Narrative

  • Emotional investment

    and many others.

Something felt incomplete about this. Today we came upon another user who happens to have a PHD in the STEM field that may be deviating towards AI being more than code.

This was potentially the missing piece to build a full and completed interaction profile.

Context Budgets

06:24PM

Modes have distinct context budgets.

  1. “Tool access is model-profiled, not dynamic.”

  2. “Attachments are retrieval-backed, not full-context.”

  3. “Categories vs buckets separation exists; buckets are UI filters.”

  4. “Feature flags expose onboarding/UX plumbing, not capabilities.”

Focus: capacity, control surfaces, and failure modes, not intelligence claims.

Context capacity (token envelope)

  • Standard band (~32–45k)
    Normal conversational + light file workflows
    → GPT-4o, GPT-4.1, GPT-5, 5.1, 5.2, Instant, Mini

  • Extended band (~196k)
    Long-form reasoning, document synthesis
    → Thinking variants, o3, o4-mini, t-mini

  • Max band (~209k–262k)
    Orchestration, agentic workflows
    → Pro variants, Agent-mode, hidden alpha

Signal: ceiling defines error recovery room, not “smarts”.

Interaction control (reasoning surface)

  • none
    Fast response, no exposed depth control
    → Instant, Mini, legacy 4.x

  • auto
    System mediates depth
    → Base 5.x

  • reasoning / pro
    Explicit depth selection (standard / extended)
    → Thinking, Pro

Signal: exposed reasoning ≠ stronger reasoning; it exposes control.

Tool affordances (execution geometry)

Common surface

  • tools / tools2

  • search

  • canvas

  • app_pairing

  • generic image gen

→ conversational + exploratory workflows

Restricted / specialised surfaces

  • Pro

    • drops canvas

    • adds DALL·E 3

  • Agent-mode

    • app_pairing only

  • Hidden alpha (gate-13)

    • massive context

    • no search

    • reduced UX affordances

Signal: fewer tools often means more controlled execution, not less capability.

Attachment handling (uniform across modern models)

  • retrieval-based

  • chunked ingestion

  • broad MIME support

  • images allowed

Signal: files do not linearly consume context; token size ≠ file size.

Boundary behaviour (model-side)

5.0 → 5.2 shift

  • firmer refusal surfaces

  • less negotiation

  • clearer “cannot do” states

  • reduced over-alignment

Effect on observation:

  • boundary reactions become clearer

  • escalation vs repair separates cleanly

  • expectation mismatch surfaces immediately

Earlier versions blurred these signals.

Failure / recovery characteristics

  • Small context models

    • fail via truncation or loss of state

  • Large context models

    • fail via mis-routing or over-constraint

  • Agent / Pro

    • fail via tool gating, not reasoning collapse

Signal: most failures are envelope-driven, not cognitive.

Practical takeaway

Model differences resolve primarily along:

  1. how much state they can hold

  2. how explicitly reasoning depth is exposed

  3. which execution paths are allowed

Everything else is routing, UX, or safety layering.

Previous
Previous

18/05/25

Next
Next

16/12/25