ChatGPT request assembly architecture

Author-mapped schematic of request assembly and context selection: gateway/policy enforcement, request assembly (selection/ordering/truncation), LLM inference, tool execution, retrieval/caching inputs, memory inputs, and observability outputs with risk checkpoints R1–R8.
ChatGPT request assembly architecture — author-mapped reference model (not vendor internal placement).

What this diagram is

This is an author-mapped reference diagram that models how a chat request can be assembled and executed across:

Provenance note: This is not a vendor-published internal architecture diagram.
According to the author, the mapping was derived from external observation (including network-level signals and repeatable behavior). This page intentionally does not include a step-by-step methodology; treat the diagram as a reference model for analysis and review.

How to read it

1) The primary request path (numbered boxes)

  1. User Input
  2. Gateway / Policy — enforcement (allowlist), redaction, auth binding
  3. Request Assembly / Context Selectorpack + order + truncate
  4. LLM Inference
  5. Answer (with a feedback signal)

2) Planes (color legend)

3) Context Inputs Hub

The Context Inputs (Hub) is the conceptual merge point shown at the bottom: it represents that multiple inputs can be merged before request assembly.

Components (as labeled in the diagram)

Gateway / Policy

A control point for:

Request Assembly / Context Selector

A packaging step that:

Execution / Tools (purple)

Retrieval / Caching (blue)

Memory plane (green)

Streaming / Observability (orange)

Risk checkpoints (R1–R8)

The red boxes mark risk checkpoints used for review. They are intended as analysis anchors (threat-model checkpoints), not as claims about any specific vendor implementation details.

Intended use

Use this page as a reference map for:

References (feature-level; not internal placement)