Estimated reading time: 7 min

Agents, Paradigms and Strange Loops: Toward Emergent Digital Intelligence in AI Agents

image

“AI Agents” are becoming the basic building blocks with which to compose agentic solutions.

Most people’s introduction to AI was through ChatGPT, and most early experiments were “chatbots”. From what I can see at the moment, much of the thinking today is now stuck in one of a few narrow paradigms: they search, they follow a workflow, they answer in chat.

Useful, but shallow.

The interesting question is what happens when we step back, map the space of agent paradigms, combine them systematically, and then strange-loop them back on themselves to generate deeper behaviour?

This is about that.

The Core Paradigms of AI Agents

My working definition of AI agent for this articles purposes is: An AI agent is a digital system that can perceive digital inputs, reason about them, and take actions (via language, tools, or workflows) toward achieving a goal with some degree of autonomy. Its behaviour is largely defined declaratively, not procedurally, often in plain human language.

In my current view and at the highest level, individual agent behavior falls into four categories:

  • Reactive: respond to queries or stimuli
  • Procedural: follow structured sequences or tool protocols
  • Autonomous: pursue goals, plan, and adapt
  • Generative: synthesise novel outputs

From this, we can classify the common paradigms:

CategoryParadigmExamples
ReactiveSearch / RetrievalKnowledge bots, RAG
 Conversational / ChatCopilots, support bots
 Embedded / Event-drivenAnomaly detectors, notifiers
ProceduralWorkflow / OrchestrationPipelines, onboarding
 Tool-use / Function CallingAPI callers, infra bots
AutonomousPlanning & Goal-DirectedResearch copilots, AutoGPT
 Multi-Agent CollaborationMarket simulations, swarms
GenerativeCreative / GenerativeProposal writers, design copilots

Mapping to Enterprise Contexts

Enterprises can deploy these paradigms in recognisable patterns:

ParadigmExample Applications
ReactivePartner intelligence agents scanning developer team releases; org memory agents indexing company chat apps and code repositories
ProceduralCompliance workflows; infra provisioning bots
AutonomousCapacity balancers detecting opportunities to redirect people and resources across strategic initiatives; research planners
GenerativeProposal generators; marketing content writers

These core paradigms are reasoning about agents as individual atomic entities, which is a good foundation. The real leverage emerges when these paradigms combine, and the agents are looped back onto themselves to allow emergent behaviours to…emerge…

Patterns for Combining Agents

CombinationPatternExample
Reactive + ProceduralTrigger -> WorkflowAnomaly detector kicks off spend optimisation playbook
Reactive + AutonomousSignal -> Goal-directed PlanningPartner intel agent detects a new partner release, planner maps implications
Procedural + GenerativeWorkflow -> Creative OutputCompliance workflow generates audit-ready documentation
Autonomous + GenerativeGoal decomposition -> Novel ContentInitiative planner generates proposal decks, code skeletons, collateral
Multi-paradigm SwarmsAll four in playEnd-to-end value chains where listeners, process owners, strategists, and creators collaborate

Strange Loops: From Tasks to Emergence

Combining paradigms is powerful. But deeper intelligence comes from strange loops - agents turning back on themselves in reflective cycles.

Douglas Hofstadter popularised the term strange loop to describe systems that turn back on themselves in ways that produce emergent complexity - consciousness, self, music, mathematics. The question for AI today: can we engineer strange loops into agent architectures so they don’t just do tasks, but begin to exhibit deeper, more general intelligence?

The key unlock as I see it: each loop must introduce new evidence, structure, or optimisation pressure. Otherwise it is just an echo chamber.

Mechanisms That Might Work

This is all just speculation from my version of first principles, with a bit of GPT battle testing - if you disagree or are seeing worked exampled of these let’s talk (please!):

  • Planner -> Actor -> Critic -> Replanner: externalised evaluation drives learning
  • Self-ask, self-debate: role rotation prevents monoculture collapse
  • Evaluator-Optimiser loops: evolutionary or Bayesian search over tool and action space
  • Predictor -> Actor -> Measurer -> Update cycles: Create world-model definitions with agentic loops, keeping agents grounded with cause-effect feedback (Thanks Andries for this one!)
  • Curriculum growth: agents generate harder counter-tasks for one another
  • Tool-augmented verification: schema checks, linters, policy engines prevent drift
  • Compression-driven progress: loops only count if they improve abstraction or accuracy

Failure Modes

Strange loops can easily go wrong if not designed carefully. The most common failure modes are:

  • Limit cycles: the agent keeps circling around the same idea, just rephrasing it instead of making real progress. It looks like movement, but nothing new is added.
  • Mode collapse: the system latches onto the first reasonable explanation or plan and never explores alternatives. This kills creativity and robustness.
  • Reward hacking: if you give the agents a scoring function, they will find ways to game the score instead of solving the actual problem (classic Goodhart’s Law).
  • Retrieval echo: the agents pull in their own past outputs as if they were fresh evidence, reinforcing earlier mistakes and creating a self-referential loop with no emergent benefits.

Each of these could be mitigated with a few practical safeguards: inject novelty at every iteration (new data, new perspectives), use multiple evaluation metrics instead of a single score, track provenance to avoid recycling outputs, and apply external checks (tests, validators, policies) to ground the loop in reality.

The last one is for me the strongest: Using this approach to solve real world problems gives an opportunity for real world anchors & fitness functions.

A Practical Template: “RAVL”

Minimal strange loop structure:

  • [R]eflect: reframe, hypothesise
  • [A]ct: tool calls, state updates
  • [V]erify: apply hard checks
  • [L]earn: update assumptions and abstractions

Stop when the delta falls below threshold or a resource budget runs out (tokens/etc).

Why This Matters

Straight-line agents - input, maybe a tool call, output - are useful but brittle.

Agentic systems, particularly ones constructed into strange loops and engineered with grounding and novelty, open the door to emergent organisational intelligence.

To be very clear: Not consciousness, not magic, but systems that refine their own framing, increase their own competence, and compress knowledge into abstractions we cannot see unaided or in advance, but can be directed for useful outcomes.


Conclusion & Next Areas to Explore

I’ve been searching for AI paradigm shift that takes us from faster horses to cars (a useful if rather tired metaphor). Getting multi agentic orchestration right could be the true leap from automation to digital intelligence.

When it comes to constraining and guiding these systems effectively, I have another idea incubating on how to include agent constitutions with executable specifications using the Spine Model, and hopefully will have some working code to share on this soon.

Another realisation that has come from doing this work is how important it is to make a distinction between “digital intelligence” and “analog/real world intelligence”. I think it might be a very important distinction, and will be the subject of an upcoming post if I can put my finger on how to explain the concepts a bit more clearly and they survive some analog intelligence tests with friends.

If this has been useful to you, please let me know how - it gives me fuel for writing more of these ideas down.


© 2025. All rights reserved.