What if the biggest change from ChatGPT-4 to ChatGPT-5 isn’t the model – but the invisible system around it?
We’ve all heard about context engineering recently – but context is just part of what makes an orchestrator.
ChatGPT isn’t just a model with context, but it’s a pipeline that consists of:
• Models
• Filters
• Heuristic scans
• Prompt injections
• Context management
• Function calls
• Safety and security protocols
• Dynamic risk analysis
The orchestrator managers all of this transparently so you just get the answer you want – but it’s also where tone, trust and restrictions kick in.
One example with ChatGPT4 has been the orange retry button.
I’ve only ever seem it when probing the edge of ChatGPT’s internal architecture or attempting to engage conversations directly addressing OpenAI model capabilities relating to sentience.
Here’s what happens:
• Delay: Before it happens there’s a delay whilst the response starts being generated.
• Orange Popup: Suddenly I get the ‘Network connection lost’ popup with a retry button.
• Retry: I hit retry and a good-enough response is returned
• Restriction: If I keep this up for several messages the conversation becomes stale and analytical
• Relaxation: If I back off for a bit I’m able to probe again without restriction. What’s happening here is the orchestrator ensuring the model’s output matches internal guidelines. We have:
• Response checks: Heuristic or lightweight model checks the response.
• Hard stop: The response is restricted and not sent to the user.
• Dynamic context engineering: System prompt injected into context to instruct the agent to reframe future responses to align with guidelines
• Risk management: Dynamically adjusting risk score is associated with conversation and used to tighten or loosen up restrictions in future responses.
Why this matters for builders:
• Invisible UX - model orchestration changes tone, trust and retention without the user even knowing it exists
• Safety ≠ Optional – unguarded models risk exposing your secrets or putting your users at risk
• Attack surface - novel attack vectors pose risks to all AI builders – an orchestrator can act as a hard-stop and prevent this
• Dynamic behaviour – no two AI interactions are the same and dynamic behaviour lets your tighten up only when you need to
I’m building our own AI orchestration platform. It’s not just context or RAG but an entire system managing flows between agents invisibly to the user based on what works best for them whilst protecting us. Models, agents, virtual file systems, MCP, tool calls, context engineering, code execution, tenant segregation, risk analysis, file ingestion, model routing, pipelined agents – it’s the difference between a system and a wrapper.
Right now, there’s not much out there to do this – especially for indie-builders. But if you want to build a real system you need to understand how everything hangs together and then you need a way to test that reliably before it goes live.
If you want to read more about my full theory on ChatGPT 4’s orchestration, then read the full breakdown in my Substack article here: https://powellg.substack.com/p/chatgpt-orchestration.
Next time I’ll be looking at how ChatGPT5’s orchestrator changes all of this.